* [PATCH 1/2] mm: swap_cgroup: allocate swap_cgroup map using vcalloc()
@ 2024-11-15 19:02 Roman Gushchin
2024-11-15 19:02 ` [PATCH 2/2] mm: swap_cgroup: get rid of __lookup_swap_cgroup() Roman Gushchin
2024-11-15 20:17 ` [PATCH 1/2] mm: swap_cgroup: allocate swap_cgroup map using vcalloc() Shakeel Butt
0 siblings, 2 replies; 5+ messages in thread
From: Roman Gushchin @ 2024-11-15 19:02 UTC (permalink / raw)
To: Andrew Morton
Cc: linux-kernel, linux-mm, Shakeel Butt, Michal Hocko,
Johannes Weiner, Roman Gushchin
Currently swap_cgroup's map is constructed as a vmalloc()'s-based
array of pointers to individual struct pages. This brings an
unnecessary complexity into the code.
This patch turns the swap_cgroup's map into a single space
allocated by vcalloc().
Signed-off-by: Roman Gushchin <roman.gushchin@linux.dev>
---
mm/swap_cgroup.c | 83 ++++++++----------------------------------------
1 file changed, 13 insertions(+), 70 deletions(-)
diff --git a/mm/swap_cgroup.c b/mm/swap_cgroup.c
index da1278f0563b..18de498c84a4 100644
--- a/mm/swap_cgroup.c
+++ b/mm/swap_cgroup.c
@@ -6,17 +6,18 @@
#include <linux/swapops.h> /* depends on mm.h include */
static DEFINE_MUTEX(swap_cgroup_mutex);
+
+struct swap_cgroup {
+ unsigned short id;
+};
+
struct swap_cgroup_ctrl {
- struct page **map;
- unsigned long length;
+ struct swap_cgroup *map;
spinlock_t lock;
};
static struct swap_cgroup_ctrl swap_cgroup_ctrl[MAX_SWAPFILES];
-struct swap_cgroup {
- unsigned short id;
-};
#define SC_PER_PAGE (PAGE_SIZE/sizeof(struct swap_cgroup))
/*
@@ -33,44 +34,10 @@ struct swap_cgroup {
* TODO: we can push these buffers out to HIGHMEM.
*/
-/*
- * allocate buffer for swap_cgroup.
- */
-static int swap_cgroup_prepare(int type)
-{
- struct page *page;
- struct swap_cgroup_ctrl *ctrl;
- unsigned long idx, max;
-
- ctrl = &swap_cgroup_ctrl[type];
-
- for (idx = 0; idx < ctrl->length; idx++) {
- page = alloc_page(GFP_KERNEL | __GFP_ZERO);
- if (!page)
- goto not_enough_page;
- ctrl->map[idx] = page;
-
- if (!(idx % SWAP_CLUSTER_MAX))
- cond_resched();
- }
- return 0;
-not_enough_page:
- max = idx;
- for (idx = 0; idx < max; idx++)
- __free_page(ctrl->map[idx]);
-
- return -ENOMEM;
-}
-
static struct swap_cgroup *__lookup_swap_cgroup(struct swap_cgroup_ctrl *ctrl,
pgoff_t offset)
{
- struct page *mappage;
- struct swap_cgroup *sc;
-
- mappage = ctrl->map[offset / SC_PER_PAGE];
- sc = page_address(mappage);
- return sc + offset % SC_PER_PAGE;
+ return &ctrl->map[offset];
}
static struct swap_cgroup *lookup_swap_cgroup(swp_entry_t ent,
@@ -168,32 +135,20 @@ unsigned short lookup_swap_cgroup_id(swp_entry_t ent)
int swap_cgroup_swapon(int type, unsigned long max_pages)
{
- void *array;
- unsigned long length;
+ struct swap_cgroup *map;
struct swap_cgroup_ctrl *ctrl;
if (mem_cgroup_disabled())
return 0;
- length = DIV_ROUND_UP(max_pages, SC_PER_PAGE);
-
- array = vcalloc(length, sizeof(void *));
- if (!array)
+ map = vcalloc(max_pages, sizeof(struct swap_cgroup));
+ if (!map)
goto nomem;
ctrl = &swap_cgroup_ctrl[type];
mutex_lock(&swap_cgroup_mutex);
- ctrl->length = length;
- ctrl->map = array;
+ ctrl->map = map;
spin_lock_init(&ctrl->lock);
- if (swap_cgroup_prepare(type)) {
- /* memory shortage */
- ctrl->map = NULL;
- ctrl->length = 0;
- mutex_unlock(&swap_cgroup_mutex);
- vfree(array);
- goto nomem;
- }
mutex_unlock(&swap_cgroup_mutex);
return 0;
@@ -205,8 +160,7 @@ int swap_cgroup_swapon(int type, unsigned long max_pages)
void swap_cgroup_swapoff(int type)
{
- struct page **map;
- unsigned long i, length;
+ struct swap_cgroup *map;
struct swap_cgroup_ctrl *ctrl;
if (mem_cgroup_disabled())
@@ -215,19 +169,8 @@ void swap_cgroup_swapoff(int type)
mutex_lock(&swap_cgroup_mutex);
ctrl = &swap_cgroup_ctrl[type];
map = ctrl->map;
- length = ctrl->length;
ctrl->map = NULL;
- ctrl->length = 0;
mutex_unlock(&swap_cgroup_mutex);
- if (map) {
- for (i = 0; i < length; i++) {
- struct page *page = map[i];
- if (page)
- __free_page(page);
- if (!(i % SWAP_CLUSTER_MAX))
- cond_resched();
- }
- vfree(map);
- }
+ kvfree(map);
}
--
2.47.0.338.g60cca15819-goog
^ permalink raw reply [flat|nested] 5+ messages in thread
* [PATCH 2/2] mm: swap_cgroup: get rid of __lookup_swap_cgroup()
2024-11-15 19:02 [PATCH 1/2] mm: swap_cgroup: allocate swap_cgroup map using vcalloc() Roman Gushchin
@ 2024-11-15 19:02 ` Roman Gushchin
2024-11-15 20:19 ` Shakeel Butt
2024-11-15 20:17 ` [PATCH 1/2] mm: swap_cgroup: allocate swap_cgroup map using vcalloc() Shakeel Butt
1 sibling, 1 reply; 5+ messages in thread
From: Roman Gushchin @ 2024-11-15 19:02 UTC (permalink / raw)
To: Andrew Morton
Cc: linux-kernel, linux-mm, Shakeel Butt, Michal Hocko,
Johannes Weiner, Roman Gushchin
Because swap_cgroup map is now virtually contiguous,
swap_cgroup_record() can be simplified, which eliminates
a need to use __lookup_swap_cgroup().
Now as __lookup_swap_cgroup() is really trivial and is used only once,
it can be inlined.
Signed-off-by: Roman Gushchin <roman.gushchin@linux.dev>
---
mm/swap_cgroup.c | 18 ++----------------
1 file changed, 2 insertions(+), 16 deletions(-)
diff --git a/mm/swap_cgroup.c b/mm/swap_cgroup.c
index 18de498c84a4..0db907308c94 100644
--- a/mm/swap_cgroup.c
+++ b/mm/swap_cgroup.c
@@ -33,13 +33,6 @@ static struct swap_cgroup_ctrl swap_cgroup_ctrl[MAX_SWAPFILES];
*
* TODO: we can push these buffers out to HIGHMEM.
*/
-
-static struct swap_cgroup *__lookup_swap_cgroup(struct swap_cgroup_ctrl *ctrl,
- pgoff_t offset)
-{
- return &ctrl->map[offset];
-}
-
static struct swap_cgroup *lookup_swap_cgroup(swp_entry_t ent,
struct swap_cgroup_ctrl **ctrlp)
{
@@ -49,7 +42,7 @@ static struct swap_cgroup *lookup_swap_cgroup(swp_entry_t ent,
ctrl = &swap_cgroup_ctrl[swp_type(ent)];
if (ctrlp)
*ctrlp = ctrl;
- return __lookup_swap_cgroup(ctrl, offset);
+ return &ctrl->map[offset];
}
/**
@@ -104,16 +97,9 @@ unsigned short swap_cgroup_record(swp_entry_t ent, unsigned short id,
spin_lock_irqsave(&ctrl->lock, flags);
old = sc->id;
- for (;;) {
+ for (; offset < end; offset++, sc++) {
VM_BUG_ON(sc->id != old);
sc->id = id;
- offset++;
- if (offset == end)
- break;
- if (offset % SC_PER_PAGE)
- sc++;
- else
- sc = __lookup_swap_cgroup(ctrl, offset);
}
spin_unlock_irqrestore(&ctrl->lock, flags);
--
2.47.0.338.g60cca15819-goog
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [PATCH 1/2] mm: swap_cgroup: allocate swap_cgroup map using vcalloc()
2024-11-15 19:02 [PATCH 1/2] mm: swap_cgroup: allocate swap_cgroup map using vcalloc() Roman Gushchin
2024-11-15 19:02 ` [PATCH 2/2] mm: swap_cgroup: get rid of __lookup_swap_cgroup() Roman Gushchin
@ 2024-11-15 20:17 ` Shakeel Butt
2024-11-15 22:46 ` Roman Gushchin
1 sibling, 1 reply; 5+ messages in thread
From: Shakeel Butt @ 2024-11-15 20:17 UTC (permalink / raw)
To: Roman Gushchin
Cc: Andrew Morton, linux-kernel, linux-mm, Michal Hocko, Johannes Weiner
On Fri, Nov 15, 2024 at 07:02:28PM +0000, Roman Gushchin wrote:
> Currently swap_cgroup's map is constructed as a vmalloc()'s-based
> array of pointers to individual struct pages. This brings an
> unnecessary complexity into the code.
>
> This patch turns the swap_cgroup's map into a single space
> allocated by vcalloc().
>
> Signed-off-by: Roman Gushchin <roman.gushchin@linux.dev>
Acked-by: Shakeel Butt <shakeel.butt@linux.dev>
[...]
> @@ -215,19 +169,8 @@ void swap_cgroup_swapoff(int type)
> mutex_lock(&swap_cgroup_mutex);
> ctrl = &swap_cgroup_ctrl[type];
> map = ctrl->map;
> - length = ctrl->length;
> ctrl->map = NULL;
> - ctrl->length = 0;
> mutex_unlock(&swap_cgroup_mutex);
>
> - if (map) {
> - for (i = 0; i < length; i++) {
> - struct page *page = map[i];
> - if (page)
> - __free_page(page);
> - if (!(i % SWAP_CLUSTER_MAX))
> - cond_resched();
> - }
> - vfree(map);
> - }
> + kvfree(map);
Any reason to use kvfree() instead of just vfree()?
> }
> --
> 2.47.0.338.g60cca15819-goog
>
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [PATCH 2/2] mm: swap_cgroup: get rid of __lookup_swap_cgroup()
2024-11-15 19:02 ` [PATCH 2/2] mm: swap_cgroup: get rid of __lookup_swap_cgroup() Roman Gushchin
@ 2024-11-15 20:19 ` Shakeel Butt
0 siblings, 0 replies; 5+ messages in thread
From: Shakeel Butt @ 2024-11-15 20:19 UTC (permalink / raw)
To: Roman Gushchin
Cc: Andrew Morton, linux-kernel, linux-mm, Michal Hocko, Johannes Weiner
On Fri, Nov 15, 2024 at 07:02:29PM +0000, Roman Gushchin wrote:
> Because swap_cgroup map is now virtually contiguous,
> swap_cgroup_record() can be simplified, which eliminates
> a need to use __lookup_swap_cgroup().
>
> Now as __lookup_swap_cgroup() is really trivial and is used only once,
> it can be inlined.
>
> Signed-off-by: Roman Gushchin <roman.gushchin@linux.dev>
Acked-by: Shakeel Butt <shakeel.butt@linux.dev>
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [PATCH 1/2] mm: swap_cgroup: allocate swap_cgroup map using vcalloc()
2024-11-15 20:17 ` [PATCH 1/2] mm: swap_cgroup: allocate swap_cgroup map using vcalloc() Shakeel Butt
@ 2024-11-15 22:46 ` Roman Gushchin
0 siblings, 0 replies; 5+ messages in thread
From: Roman Gushchin @ 2024-11-15 22:46 UTC (permalink / raw)
To: Shakeel Butt
Cc: Andrew Morton, linux-kernel, linux-mm, Michal Hocko, Johannes Weiner
On Fri, Nov 15, 2024 at 12:17:43PM -0800, Shakeel Butt wrote:
> On Fri, Nov 15, 2024 at 07:02:28PM +0000, Roman Gushchin wrote:
> > Currently swap_cgroup's map is constructed as a vmalloc()'s-based
> > array of pointers to individual struct pages. This brings an
> > unnecessary complexity into the code.
> >
> > This patch turns the swap_cgroup's map into a single space
> > allocated by vcalloc().
> >
> > Signed-off-by: Roman Gushchin <roman.gushchin@linux.dev>
>
> Acked-by: Shakeel Butt <shakeel.butt@linux.dev>
Thank you for reviewing the patchset!
>
> [...]
> > @@ -215,19 +169,8 @@ void swap_cgroup_swapoff(int type)
> > mutex_lock(&swap_cgroup_mutex);
> > ctrl = &swap_cgroup_ctrl[type];
> > map = ctrl->map;
> > - length = ctrl->length;
> > ctrl->map = NULL;
> > - ctrl->length = 0;
> > mutex_unlock(&swap_cgroup_mutex);
> >
> > - if (map) {
> > - for (i = 0; i < length; i++) {
> > - struct page *page = map[i];
> > - if (page)
> > - __free_page(page);
> > - if (!(i % SWAP_CLUSTER_MAX))
> > - cond_resched();
> > - }
> > - vfree(map);
> > - }
> > + kvfree(map);
>
> Any reason to use kvfree() instead of just vfree()?
No, you're right, vfree() is a better choice here.
Andrew, do you prefer a v2 or it's easier to master a fix-up?
Thank you!
^ permalink raw reply [flat|nested] 5+ messages in thread
end of thread, other threads:[~2024-11-15 22:46 UTC | newest]
Thread overview: 5+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2024-11-15 19:02 [PATCH 1/2] mm: swap_cgroup: allocate swap_cgroup map using vcalloc() Roman Gushchin
2024-11-15 19:02 ` [PATCH 2/2] mm: swap_cgroup: get rid of __lookup_swap_cgroup() Roman Gushchin
2024-11-15 20:19 ` Shakeel Butt
2024-11-15 20:17 ` [PATCH 1/2] mm: swap_cgroup: allocate swap_cgroup map using vcalloc() Shakeel Butt
2024-11-15 22:46 ` Roman Gushchin
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox