From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 06E24C4332F for ; Fri, 18 Nov 2022 04:29:52 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 483736B0071; Thu, 17 Nov 2022 23:29:52 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id 434006B0072; Thu, 17 Nov 2022 23:29:52 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 2D4208E0001; Thu, 17 Nov 2022 23:29:52 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0016.hostedemail.com [216.40.44.16]) by kanga.kvack.org (Postfix) with ESMTP id 1D0996B0071 for ; Thu, 17 Nov 2022 23:29:52 -0500 (EST) Received: from smtpin08.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay08.hostedemail.com (Postfix) with ESMTP id D9DAD141419 for ; Fri, 18 Nov 2022 04:29:51 +0000 (UTC) X-FDA: 80145284982.08.FDDA051 Received: from mail-qt1-f178.google.com (mail-qt1-f178.google.com [209.85.160.178]) by imf28.hostedemail.com (Postfix) with ESMTP id 49F9DC0006 for ; Fri, 18 Nov 2022 04:29:50 +0000 (UTC) Received: by mail-qt1-f178.google.com with SMTP id z6so2486783qtv.5 for ; Thu, 17 Nov 2022 20:29:49 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=cmpxchg-org.20210112.gappssmtp.com; s=20210112; h=in-reply-to:content-disposition:mime-version:references:message-id :subject:cc:to:from:date:from:to:cc:subject:date:message-id:reply-to; bh=HTToeUbo/I6A+yeg0A23kiy55DPhxnbXYNCgTS9YFP0=; b=8JZgZuBCQUrGdUyof1lFOwNJklwkGMmn66bvebQZMWX5RXNDsZpVLzJPTg3iTTNgqk Q+hnZ7xO0x4Yo1xfExuxn4E7tP0g2cliTJ5pnvXrI9F01f3YmiCytP32+JJkwrb4cAx2 znzoo8p7F50mV1yTB0wlO2BGQRl1vL1GLa7ETW3LqxzgrasqVztR8E3/lyJWRBQKz7Ha B47vjyRqpYpbZFxACxBeLgRp3l1ujGALASGNGwIfbvAbWJauBVOIeDej+yd2g9DjUNhC axW436LvqgfEW8lrAB8VxYzolxxF/uShR+/jz870LfOv5GPtv9GO9uh4SFVpCo00GMy/ aF2g== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=in-reply-to:content-disposition:mime-version:references:message-id :subject:cc:to:from:date:x-gm-message-state:from:to:cc:subject:date :message-id:reply-to; bh=HTToeUbo/I6A+yeg0A23kiy55DPhxnbXYNCgTS9YFP0=; b=D2n+mizSPhhB9ncX7eJd2H3heeDGzILKYpdPjl1fz3AIOnq99V/Gmgl+yDTbJtaiCQ LF2FxsqfLje9H6PArJ2iikOKaabCsz28I+t9riMCdFwmZ7BEFKHkZVyxKYu6uDd4hiQc q9nI2U4zt/Pcxf+4Hi84JvDRW6DM4uDcW/s9CcNYBsigALsvuArE4uzN8dEvLXz8xjGu y1PJEYSXnDP2g0NriaoFiFOBEnT0xeFL6HRBItGoWIYfTUBlLT+iLvaegi+u8PidSmwH O6RKgSLZgOcEckd045c4PeB8O90OEJZ7fo+aNbKC1NHAZnNfnjbIIytQCCKr+92y43ix 5F4g== X-Gm-Message-State: ANoB5ploUJ7XXJ4cfsy4Ob+5hiO00B3qwTn7Kbc65jMC050ywWqVLBZv vhCNzYsfM/Jg1s91ICuhRL1EkQ== X-Google-Smtp-Source: AA0mqf6kdwsd0Ac7KvFF9V8mmXyJXvqRDkVkXzItreJ6LO1CPyXHdHfhhuqVkFqtaPuSAc1Ia/owlQ== X-Received: by 2002:a05:622a:4110:b0:3a5:4502:3d67 with SMTP id cc16-20020a05622a411000b003a545023d67mr5234952qtb.270.1668745789362; Thu, 17 Nov 2022 20:29:49 -0800 (PST) Received: from localhost ([2620:10d:c091:480::1:bc4]) by smtp.gmail.com with ESMTPSA id br35-20020a05620a462300b006cfc01b4461sm1767979qkb.118.2022.11.17.20.29.47 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Thu, 17 Nov 2022 20:29:48 -0800 (PST) Date: Thu, 17 Nov 2022 23:30:11 -0500 From: Johannes Weiner To: Minchan Kim Cc: Nhat Pham , akpm@linux-foundation.org, linux-mm@kvack.org, linux-kernel@vger.kernel.org, ngupta@vflare.org, senozhatsky@chromium.org, sjenning@redhat.com, ddstreet@ieee.org, vitaly.wool@konsulko.com Subject: Re: [PATCH v4 3/5] zsmalloc: Add a LRU to zs_pool to keep track of zspages in LRU order Message-ID: References: <20221117163839.230900-1-nphamcs@gmail.com> <20221117163839.230900-4-nphamcs@gmail.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1668745791; a=rsa-sha256; cv=none; b=XQtu86wO18Z9UySU6rsbTdo4sELnQS5f4AoEBRL89BMUPMDDVikKFWYlaUEU7czFWmmdN+ MrJBMmEraE607gutg3fhrxRo5EdlJEq/aAYXH1UYDJcLxDqm7OMKou8r2dJd4bmyb51AdE 9Z0zqd4a42NW6ps1GywiGw4SFtI/CUo= ARC-Authentication-Results: i=1; imf28.hostedemail.com; dkim=pass header.d=cmpxchg-org.20210112.gappssmtp.com header.s=20210112 header.b=8JZgZuBC; dmarc=pass (policy=none) header.from=cmpxchg.org; spf=pass (imf28.hostedemail.com: domain of hannes@cmpxchg.org designates 209.85.160.178 as permitted sender) smtp.mailfrom=hannes@cmpxchg.org ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1668745791; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=HTToeUbo/I6A+yeg0A23kiy55DPhxnbXYNCgTS9YFP0=; b=bbFDER4mQAJT9HNAvNgCu8Pnhszr9q69XBAQa+XK/oU0POTYOi+q7Yk8OvOug9Splej2wd DKjT39DBm6Bg+eJ6T9KJkQXWAobKbUtrQRQ80tm7TJnQfYJdZ9A5ymjS0AuKftZ32+zawa 6sd7qSuEOOG8Malrk/YhZn35RPq1/DY= X-Rspam-User: X-Rspamd-Server: rspam08 X-Rspamd-Queue-Id: 49F9DC0006 Authentication-Results: imf28.hostedemail.com; dkim=pass header.d=cmpxchg-org.20210112.gappssmtp.com header.s=20210112 header.b=8JZgZuBC; dmarc=pass (policy=none) header.from=cmpxchg.org; spf=pass (imf28.hostedemail.com: domain of hannes@cmpxchg.org designates 209.85.160.178 as permitted sender) smtp.mailfrom=hannes@cmpxchg.org X-Stat-Signature: yh8xk5yjooe5eni4treouri13qwf5nkw X-HE-Tag: 1668745790-301450 X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: On Thu, Nov 17, 2022 at 02:15:11PM -0800, Minchan Kim wrote: > On Thu, Nov 17, 2022 at 08:38:37AM -0800, Nhat Pham wrote: > > This helps determines the coldest zspages as candidates for writeback. > > > > Signed-off-by: Nhat Pham > > --- > > mm/zsmalloc.c | 46 ++++++++++++++++++++++++++++++++++++++++++++++ > > 1 file changed, 46 insertions(+) > > > > diff --git a/mm/zsmalloc.c b/mm/zsmalloc.c > > index 326faa751f0a..2557b55ec767 100644 > > --- a/mm/zsmalloc.c > > +++ b/mm/zsmalloc.c > > @@ -239,6 +239,11 @@ struct zs_pool { > > /* Compact classes */ > > struct shrinker shrinker; > > > > +#ifdef CONFIG_ZPOOL > > + /* List tracking the zspages in LRU order by most recently added object */ > > + struct list_head lru; > > +#endif > > + > > #ifdef CONFIG_ZSMALLOC_STAT > > struct dentry *stat_dentry; > > #endif > > @@ -260,6 +265,12 @@ struct zspage { > > unsigned int freeobj; > > struct page *first_page; > > struct list_head list; /* fullness list */ > > + > > +#ifdef CONFIG_ZPOOL > > + /* links the zspage to the lru list in the pool */ > > + struct list_head lru; > > +#endif > > + > > struct zs_pool *pool; > > #ifdef CONFIG_COMPACTION > > rwlock_t lock; > > @@ -352,6 +363,18 @@ static void cache_free_zspage(struct zs_pool *pool, struct zspage *zspage) > > kmem_cache_free(pool->zspage_cachep, zspage); > > } > > > > +#ifdef CONFIG_ZPOOL > > +/* Moves the zspage to the front of the zspool's LRU */ > > +static void move_to_front(struct zs_pool *pool, struct zspage *zspage) > > +{ > > + assert_spin_locked(&pool->lock); > > + > > + if (!list_empty(&zspage->lru)) > > + list_del(&zspage->lru); > > + list_add(&zspage->lru, &pool->lru); > > +} > > +#endif > > + > > /* pool->lock(which owns the handle) synchronizes races */ > > static void record_obj(unsigned long handle, unsigned long obj) > > { > > @@ -953,6 +976,9 @@ static void free_zspage(struct zs_pool *pool, struct size_class *class, > > } > > > > remove_zspage(class, zspage, ZS_EMPTY); > > +#ifdef CONFIG_ZPOOL > > + list_del(&zspage->lru); > > +#endif > > __free_zspage(pool, class, zspage); > > } > > > > @@ -998,6 +1024,10 @@ static void init_zspage(struct size_class *class, struct zspage *zspage) > > off %= PAGE_SIZE; > > } > > > > +#ifdef CONFIG_ZPOOL > > + INIT_LIST_HEAD(&zspage->lru); > > +#endif > > + > > set_freeobj(zspage, 0); > > } > > > > @@ -1418,6 +1448,11 @@ unsigned long zs_malloc(struct zs_pool *pool, size_t size, gfp_t gfp) > > fix_fullness_group(class, zspage); > > record_obj(handle, obj); > > class_stat_inc(class, OBJ_USED, 1); > > + > > +#ifdef CONFIG_ZPOOL > > + /* Move the zspage to front of pool's LRU */ > > + move_to_front(pool, zspage); > > +#endif > > spin_unlock(&pool->lock); > > > > return handle; > > @@ -1444,6 +1479,10 @@ unsigned long zs_malloc(struct zs_pool *pool, size_t size, gfp_t gfp) > > > > /* We completely set up zspage so mark them as movable */ > > SetZsPageMovable(pool, zspage); > > +#ifdef CONFIG_ZPOOL > > + /* Move the zspage to front of pool's LRU */ > > + move_to_front(pool, zspage); > > +#endif > > spin_unlock(&pool->lock); > > Why do we move the zspage in the alloc instead of accessor? > > Isn't zs_map_object better place since it's clear semantic > that user start to access the object? Remember that this is used for swap, and these entries aren't accessed on an ongoing basis while in the pool. An access means swapin. So functionally this is fine. On cleaner choices, I would actually agree with you that map would be more appropriate. But all I can do is repeat replies from previous questions: We're not reinventing the wheel here. zbud and z3fold do it this way, and this follows their precedent. We've talked about wanting to generalize the LRU, and that's a heck of a lot easier if we have copies of *one* implementation that we can deduplicate - instead of having to merge multiple implementations with arbitrary differences. So as you review these patches, please compare what they do to zbud and z3fold. Those existing zpool implementations have been a common answer in response to review questions.