* [PATCH] /dev/zero: make private mapping full anonymous mapping
@ 2025-01-13 22:30 Yang Shi
2025-01-14 12:05 ` Lorenzo Stoakes
` (2 more replies)
0 siblings, 3 replies; 35+ messages in thread
From: Yang Shi @ 2025-01-13 22:30 UTC (permalink / raw)
To: arnd, gregkh, Liam.Howlett, lorenzo.stoakes, vbabka, jannh,
willy, liushixin2, akpm
Cc: yang, linux-mm, linux-kernel
When creating private mapping for /dev/zero, the driver makes it an
anonymous mapping by calling set_vma_anonymous(). But it just sets
vm_ops to NULL, vm_file is still valid and vm_pgoff is also file offset.
This is a special case and the VMA doesn't look like either anonymous VMA
or file VMA. It confused other kernel subsystem, for example, khugepaged [1].
It seems pointless to keep such special case. Making private /dev/zero
mapping a full anonymous mapping doesn't change the semantic of
/dev/zero either.
The user visible effect is the mapping entry shown in /proc/<PID>/smaps
and /proc/<PID>/maps.
Before the change:
ffffb7190000-ffffb7590000 rw-p 00001000 00:06 8 /dev/zero
After the change:
ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
[1]: https://lore.kernel.org/linux-mm/20250111034511.2223353-1-liushixin2@huawei.com/
Signed-off-by: Yang Shi <yang@os.amperecomputing.com>
---
drivers/char/mem.c | 4 ++++
1 file changed, 4 insertions(+)
diff --git a/drivers/char/mem.c b/drivers/char/mem.c
index 169eed162a7f..dae113f7fc1b 100644
--- a/drivers/char/mem.c
+++ b/drivers/char/mem.c
@@ -527,6 +527,10 @@ static int mmap_zero(struct file *file, struct vm_area_struct *vma)
if (vma->vm_flags & VM_SHARED)
return shmem_zero_setup(vma);
vma_set_anonymous(vma);
+ fput(vma->vm_file);
+ vma->vm_file = NULL;
+ vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT;
+
return 0;
}
--
2.47.0
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-13 22:30 [PATCH] /dev/zero: make private mapping full anonymous mapping Yang Shi
@ 2025-01-14 12:05 ` Lorenzo Stoakes
2025-01-14 16:53 ` Yang Shi
2025-01-14 13:01 ` David Hildenbrand
2025-01-28 3:14 ` kernel test robot
2 siblings, 1 reply; 35+ messages in thread
From: Lorenzo Stoakes @ 2025-01-14 12:05 UTC (permalink / raw)
To: Yang Shi
Cc: arnd, gregkh, Liam.Howlett, vbabka, jannh, willy, liushixin2,
akpm, linux-mm, linux-kernel
+ Willy for the fs/weirdness elements of this.
On Mon, Jan 13, 2025 at 02:30:33PM -0800, Yang Shi wrote:
> When creating private mapping for /dev/zero, the driver makes it an
> anonymous mapping by calling set_vma_anonymous(). But it just sets
> vm_ops to NULL, vm_file is still valid and vm_pgoff is also file offset.
Hm yikes.
>
> This is a special case and the VMA doesn't look like either anonymous VMA
> or file VMA. It confused other kernel subsystem, for example, khugepaged [1].
>
> It seems pointless to keep such special case. Making private /dev/zero
> mapping a full anonymous mapping doesn't change the semantic of
> /dev/zero either.
My concern is that ostensibly there _is_ a file right? Are we certain that by
not setting this we are not breaking something somewhere else?
Are we not creating a sort of other type of 'non-such-beast' here?
I mean already setting it anon and setting vm_file non-NULL is really strange.
>
> The user visible effect is the mapping entry shown in /proc/<PID>/smaps
> and /proc/<PID>/maps.
>
> Before the change:
> ffffb7190000-ffffb7590000 rw-p 00001000 00:06 8 /dev/zero
>
> After the change:
> ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
>
Yeah this seems like it might break somebody to be honest, it's really
really really strange to map a file then for it not to be mapped.
But it's possibly EVEN WEIRDER to map a file and for it to seem mapped as a
file but for it to be marked anonymous.
God what a mess.
> [1]: https://lore.kernel.org/linux-mm/20250111034511.2223353-1-liushixin2@huawei.com/
I kind of hate that we have to mitigate like this for a case that should
never ever happen so I'm inclined towards your solution but a lot more
inclined towards us totally rethinking this.
Do we _have_ to make this anonymous?? Why can't we just reference the zero
page as if it were in the page cache (Willy - feel free to correct naive
misapprehension here).
>
> Signed-off-by: Yang Shi <yang@os.amperecomputing.com>
> ---
> drivers/char/mem.c | 4 ++++
> 1 file changed, 4 insertions(+)
>
> diff --git a/drivers/char/mem.c b/drivers/char/mem.c
> index 169eed162a7f..dae113f7fc1b 100644
> --- a/drivers/char/mem.c
> +++ b/drivers/char/mem.c
> @@ -527,6 +527,10 @@ static int mmap_zero(struct file *file, struct vm_area_struct *vma)
> if (vma->vm_flags & VM_SHARED)
> return shmem_zero_setup(vma);
> vma_set_anonymous(vma);
> + fput(vma->vm_file);
> + vma->vm_file = NULL;
> + vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT;
Hmm, this might have been mremap()'d _potentially_ though? And then now
this will be wrong? But then we'd have no way of tracking it correctly...
I've not checked the function but do we mark this as a special mapping of
some kind?
> +
> return 0;
> }
>
> --
> 2.47.0
>
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-13 22:30 [PATCH] /dev/zero: make private mapping full anonymous mapping Yang Shi
2025-01-14 12:05 ` Lorenzo Stoakes
@ 2025-01-14 13:01 ` David Hildenbrand
2025-01-14 14:52 ` Lorenzo Stoakes
2025-01-28 3:14 ` kernel test robot
2 siblings, 1 reply; 35+ messages in thread
From: David Hildenbrand @ 2025-01-14 13:01 UTC (permalink / raw)
To: Yang Shi, arnd, gregkh, Liam.Howlett, lorenzo.stoakes, vbabka,
jannh, willy, liushixin2, akpm
Cc: linux-mm, linux-kernel
On 13.01.25 23:30, Yang Shi wrote:
> When creating private mapping for /dev/zero, the driver makes it an
> anonymous mapping by calling set_vma_anonymous(). But it just sets
> vm_ops to NULL, vm_file is still valid and vm_pgoff is also file offset.
>
> This is a special case and the VMA doesn't look like either anonymous VMA
> or file VMA. It confused other kernel subsystem, for example, khugepaged [1].
>
> It seems pointless to keep such special case. Making private
/dev/zero> mapping a full anonymous mapping doesn't change the semantic of
> /dev/zero either.
>
> The user visible effect is the mapping entry shown in /proc/<PID>/smaps
> and /proc/<PID>/maps.
>
> Before the change:
> ffffb7190000-ffffb7590000 rw-p 00001000 00:06 8 /dev/zero
>
> After the change:
> ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
>
Hm, not sure about this. It's actually quite consistent to have that
output in smaps the way it is. You mapped a file at an offset, and it
behaves like an anonymous mapping apart from that.
Not sure if the buggy khugepaged thing is a good indicator to warrant
this change.
--
Cheers,
David / dhildenb
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 13:01 ` David Hildenbrand
@ 2025-01-14 14:52 ` Lorenzo Stoakes
2025-01-14 15:06 ` David Hildenbrand
0 siblings, 1 reply; 35+ messages in thread
From: Lorenzo Stoakes @ 2025-01-14 14:52 UTC (permalink / raw)
To: David Hildenbrand
Cc: Yang Shi, arnd, gregkh, Liam.Howlett, vbabka, jannh, willy,
liushixin2, akpm, linux-mm, linux-kernel
On Tue, Jan 14, 2025 at 02:01:32PM +0100, David Hildenbrand wrote:
> On 13.01.25 23:30, Yang Shi wrote:
> > When creating private mapping for /dev/zero, the driver makes it an
> > anonymous mapping by calling set_vma_anonymous(). But it just sets
> > vm_ops to NULL, vm_file is still valid and vm_pgoff is also file offset.
> >
> > This is a special case and the VMA doesn't look like either anonymous VMA
> > or file VMA. It confused other kernel subsystem, for example, khugepaged [1].
> >
> > It seems pointless to keep such special case. Making private /dev/zero>
> mapping a full anonymous mapping doesn't change the semantic of
> > /dev/zero either.
> >
> > The user visible effect is the mapping entry shown in /proc/<PID>/smaps
> > and /proc/<PID>/maps.
> >
> > Before the change:
> > ffffb7190000-ffffb7590000 rw-p 00001000 00:06 8 /dev/zero
> >
> > After the change:
> > ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
> >
>
> Hm, not sure about this. It's actually quite consistent to have that output
> in smaps the way it is. You mapped a file at an offset, and it behaves like
> an anonymous mapping apart from that.
>
> Not sure if the buggy khugepaged thing is a good indicator to warrant this
> change.
Yeah, this is a user-facing fundamental change that hides information and
defies expectation so I mean - it's a no go really isn't it?
I'd rather we _not_ make this anon though, because isn't life confusing
enough David? I thought it was bad enough with 'anon, file and lol shmem'
but 'lol lol also /dev/zero' is enough to make me want to frolick in the
fields...
>
> --
> Cheers,
>
> David / dhildenb
>
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 14:52 ` Lorenzo Stoakes
@ 2025-01-14 15:06 ` David Hildenbrand
2025-01-14 17:01 ` Yang Shi
2025-01-14 17:02 ` David Hildenbrand
0 siblings, 2 replies; 35+ messages in thread
From: David Hildenbrand @ 2025-01-14 15:06 UTC (permalink / raw)
To: Lorenzo Stoakes
Cc: Yang Shi, arnd, gregkh, Liam.Howlett, vbabka, jannh, willy,
liushixin2, akpm, linux-mm, linux-kernel
On 14.01.25 15:52, Lorenzo Stoakes wrote:
> On Tue, Jan 14, 2025 at 02:01:32PM +0100, David Hildenbrand wrote:
>> On 13.01.25 23:30, Yang Shi wrote:
>>> When creating private mapping for /dev/zero, the driver makes it an
>>> anonymous mapping by calling set_vma_anonymous(). But it just sets
>>> vm_ops to NULL, vm_file is still valid and vm_pgoff is also file offset.
>>>
>>> This is a special case and the VMA doesn't look like either anonymous VMA
>>> or file VMA. It confused other kernel subsystem, for example, khugepaged [1].
>>>
>>> It seems pointless to keep such special case. Making private /dev/zero>
>> mapping a full anonymous mapping doesn't change the semantic of
>>> /dev/zero either.
>>>
>>> The user visible effect is the mapping entry shown in /proc/<PID>/smaps
>>> and /proc/<PID>/maps.
>>>
>>> Before the change:
>>> ffffb7190000-ffffb7590000 rw-p 00001000 00:06 8 /dev/zero
>>>
>>> After the change:
>>> ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
>>>
>>
>> Hm, not sure about this. It's actually quite consistent to have that output
>> in smaps the way it is. You mapped a file at an offset, and it behaves like
>> an anonymous mapping apart from that.
>>
>> Not sure if the buggy khugepaged thing is a good indicator to warrant this
>> change.
>
> Yeah, this is a user-facing fundamental change that hides information and
> defies expectation so I mean - it's a no go really isn't it?
>
> I'd rather we _not_ make this anon though, because isn't life confusing
> enough David? I thought it was bad enough with 'anon, file and lol shmem'
> but 'lol lol also /dev/zero' is enough to make me want to frolick in the
> fields...
I recall there are users that rely on this memory to get the shared
zeropage on reads etc (in comparison to shmem!), so I better not ...
mess with this *at all* :)
--
Cheers,
David / dhildenb
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 12:05 ` Lorenzo Stoakes
@ 2025-01-14 16:53 ` Yang Shi
2025-01-14 18:14 ` Lorenzo Stoakes
0 siblings, 1 reply; 35+ messages in thread
From: Yang Shi @ 2025-01-14 16:53 UTC (permalink / raw)
To: Lorenzo Stoakes
Cc: arnd, gregkh, Liam.Howlett, vbabka, jannh, willy, liushixin2,
akpm, linux-mm, linux-kernel
On 1/14/25 4:05 AM, Lorenzo Stoakes wrote:
> + Willy for the fs/weirdness elements of this.
>
> On Mon, Jan 13, 2025 at 02:30:33PM -0800, Yang Shi wrote:
>> When creating private mapping for /dev/zero, the driver makes it an
>> anonymous mapping by calling set_vma_anonymous(). But it just sets
>> vm_ops to NULL, vm_file is still valid and vm_pgoff is also file offset.
> Hm yikes.
>
>> This is a special case and the VMA doesn't look like either anonymous VMA
>> or file VMA. It confused other kernel subsystem, for example, khugepaged [1].
>>
>> It seems pointless to keep such special case. Making private /dev/zero
>> mapping a full anonymous mapping doesn't change the semantic of
>> /dev/zero either.
> My concern is that ostensibly there _is_ a file right? Are we certain that by
> not setting this we are not breaking something somewhere else?
>
> Are we not creating a sort of other type of 'non-such-beast' here?
But the file is /dev/zero. I don't see this could break the semantic of
/dev/zero. The shared mapping of /dev/zero is not affected by this
change, kernel already treated private mapping of /dev/zero as anonymous
mapping, but with some weird settings in VMA. When reading the mapping,
it returns 0 with zero page, when writing the mapping, a new anonymous
folio is allocated.
>
> I mean already setting it anon and setting vm_file non-NULL is really strange.
>
>> The user visible effect is the mapping entry shown in /proc/<PID>/smaps
>> and /proc/<PID>/maps.
>>
>> Before the change:
>> ffffb7190000-ffffb7590000 rw-p 00001000 00:06 8 /dev/zero
>>
>> After the change:
>> ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
>>
> Yeah this seems like it might break somebody to be honest, it's really
> really really strange to map a file then for it not to be mapped.
Yes, it is possible if someone really care whether the anonymous-like
mapping is mapped by /dev/zero or just created by malloc(). But I don't
know who really do...
>
> But it's possibly EVEN WEIRDER to map a file and for it to seem mapped as a
> file but for it to be marked anonymous.
>
> God what a mess.
>
>> [1]: https://lore.kernel.org/linux-mm/20250111034511.2223353-1-liushixin2@huawei.com/
> I kind of hate that we have to mitigate like this for a case that should
> never ever happen so I'm inclined towards your solution but a lot more
> inclined towards us totally rethinking this.
>
> Do we _have_ to make this anonymous?? Why can't we just reference the zero
> page as if it were in the page cache (Willy - feel free to correct naive
> misapprehension here).
TBH, I don't see why page cache has to be involved. When reading, 0 is
returned by zero page. When writing a CoW is triggered if page cache is
involved, but the content of the page cache should be just 0, so we copy
0 to the new folio then write to it. It doesn't make too much sense. I
think this is why private /dev/zero mapping is treated as anonymous
mapping in the first place.
>
>> Signed-off-by: Yang Shi <yang@os.amperecomputing.com>
>> ---
>> drivers/char/mem.c | 4 ++++
>> 1 file changed, 4 insertions(+)
>>
>> diff --git a/drivers/char/mem.c b/drivers/char/mem.c
>> index 169eed162a7f..dae113f7fc1b 100644
>> --- a/drivers/char/mem.c
>> +++ b/drivers/char/mem.c
>> @@ -527,6 +527,10 @@ static int mmap_zero(struct file *file, struct vm_area_struct *vma)
>> if (vma->vm_flags & VM_SHARED)
>> return shmem_zero_setup(vma);
>> vma_set_anonymous(vma);
>> + fput(vma->vm_file);
>> + vma->vm_file = NULL;
>> + vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT;
> Hmm, this might have been mremap()'d _potentially_ though? And then now
> this will be wrong? But then we'd have no way of tracking it correctly...
I'm not quite familiar with the subtle details and corner cases of
meremap(). But mmap_zero() should be called by mmap(), so the VMA has
not been visible to user yet at this point IIUC. How come mremap() could
move it?
>
> I've not checked the function but do we mark this as a special mapping of
> some kind?
>
>> +
>> return 0;
>> }
>>
>> --
>> 2.47.0
>>
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 15:06 ` David Hildenbrand
@ 2025-01-14 17:01 ` Yang Shi
2025-01-14 17:23 ` David Hildenbrand
2025-01-14 17:02 ` David Hildenbrand
1 sibling, 1 reply; 35+ messages in thread
From: Yang Shi @ 2025-01-14 17:01 UTC (permalink / raw)
To: David Hildenbrand, Lorenzo Stoakes
Cc: arnd, gregkh, Liam.Howlett, vbabka, jannh, willy, liushixin2,
akpm, linux-mm, linux-kernel
On 1/14/25 7:06 AM, David Hildenbrand wrote:
> On 14.01.25 15:52, Lorenzo Stoakes wrote:
>> On Tue, Jan 14, 2025 at 02:01:32PM +0100, David Hildenbrand wrote:
>>> On 13.01.25 23:30, Yang Shi wrote:
>>>> When creating private mapping for /dev/zero, the driver makes it an
>>>> anonymous mapping by calling set_vma_anonymous(). But it just sets
>>>> vm_ops to NULL, vm_file is still valid and vm_pgoff is also file
>>>> offset.
>>>>
>>>> This is a special case and the VMA doesn't look like either
>>>> anonymous VMA
>>>> or file VMA. It confused other kernel subsystem, for example,
>>>> khugepaged [1].
>>>>
>>>> It seems pointless to keep such special case. Making private
>>>> /dev/zero>
>>> mapping a full anonymous mapping doesn't change the semantic of
>>>> /dev/zero either.
>>>>
>>>> The user visible effect is the mapping entry shown in
>>>> /proc/<PID>/smaps
>>>> and /proc/<PID>/maps.
>>>>
>>>> Before the change:
>>>> ffffb7190000-ffffb7590000 rw-p 00001000 00:06
>>>> 8 /dev/zero
>>>>
>>>> After the change:
>>>> ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
>>>>
>>>
>>> Hm, not sure about this. It's actually quite consistent to have that
>>> output
>>> in smaps the way it is. You mapped a file at an offset, and it
>>> behaves like
>>> an anonymous mapping apart from that.
>>>
>>> Not sure if the buggy khugepaged thing is a good indicator to
>>> warrant this
>>> change.
I admit this may be a concern, but I doubt who really care about it...
>>
>> Yeah, this is a user-facing fundamental change that hides information
>> and
>> defies expectation so I mean - it's a no go really isn't it?
>>
>> I'd rather we _not_ make this anon though, because isn't life confusing
>> enough David? I thought it was bad enough with 'anon, file and lol
>> shmem'
>> but 'lol lol also /dev/zero' is enough to make me want to frolick in the
>> fields...
>
> I recall there are users that rely on this memory to get the shared
> zeropage on reads etc (in comparison to shmem!), so I better not ...
> mess with this *at all* :)
The behavior won't be changed.
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 15:06 ` David Hildenbrand
2025-01-14 17:01 ` Yang Shi
@ 2025-01-14 17:02 ` David Hildenbrand
2025-01-14 17:20 ` Yang Shi
1 sibling, 1 reply; 35+ messages in thread
From: David Hildenbrand @ 2025-01-14 17:02 UTC (permalink / raw)
To: Lorenzo Stoakes
Cc: Yang Shi, arnd, gregkh, Liam.Howlett, vbabka, jannh, willy,
liushixin2, akpm, linux-mm, linux-kernel
On 14.01.25 16:06, David Hildenbrand wrote:
> On 14.01.25 15:52, Lorenzo Stoakes wrote:
>> On Tue, Jan 14, 2025 at 02:01:32PM +0100, David Hildenbrand wrote:
>>> On 13.01.25 23:30, Yang Shi wrote:
>>>> When creating private mapping for /dev/zero, the driver makes it an
>>>> anonymous mapping by calling set_vma_anonymous(). But it just sets
>>>> vm_ops to NULL, vm_file is still valid and vm_pgoff is also file offset.
>>>>
>>>> This is a special case and the VMA doesn't look like either anonymous VMA
>>>> or file VMA. It confused other kernel subsystem, for example, khugepaged [1].
>>>>
>>>> It seems pointless to keep such special case. Making private /dev/zero>
>>> mapping a full anonymous mapping doesn't change the semantic of
>>>> /dev/zero either.
>>>>
>>>> The user visible effect is the mapping entry shown in /proc/<PID>/smaps
>>>> and /proc/<PID>/maps.
>>>>
>>>> Before the change:
>>>> ffffb7190000-ffffb7590000 rw-p 00001000 00:06 8 /dev/zero
>>>>
>>>> After the change:
>>>> ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
>>>>
>>>
>>> Hm, not sure about this. It's actually quite consistent to have that output
>>> in smaps the way it is. You mapped a file at an offset, and it behaves like
>>> an anonymous mapping apart from that.
>>>
>>> Not sure if the buggy khugepaged thing is a good indicator to warrant this
>>> change.
>>
>> Yeah, this is a user-facing fundamental change that hides information and
>> defies expectation so I mean - it's a no go really isn't it?
>>
>> I'd rather we _not_ make this anon though, because isn't life confusing
>> enough David? I thought it was bad enough with 'anon, file and lol shmem'
>> but 'lol lol also /dev/zero' is enough to make me want to frolick in the
>> fields...
>
> I recall there are users that rely on this memory to get the shared
> zeropage on reads etc (in comparison to shmem!), so I better not ...
> mess with this *at all* :)
Heh, and I recall reading something about odd behavior of /dev/zero and
some interesting history [1].
"
Unlike /dev/null, /dev/zero may be used as a source, not only as a sink
for data. All write operations to /dev/zero succeed with no other
effects. However, /dev/null is more commonly used for this purpose.
When /dev/zero is memory-mapped, e.g., with mmap, to the virtual address
space, it is equivalent to using anonymous memory; i.e. memory not
connected to any file.
"
"equivalent to using anonymous memory" is interesting.
Also, /dev/zero was there before MAP_ANONYMOUS was invented according to
[1], which is quite interesting.
... so this is anonymous memory as "real" as it can get :)
[1] https://en.wikipedia.org/wiki//dev/zero
--
Cheers,
David / dhildenb
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 17:02 ` David Hildenbrand
@ 2025-01-14 17:20 ` Yang Shi
2025-01-14 17:24 ` David Hildenbrand
0 siblings, 1 reply; 35+ messages in thread
From: Yang Shi @ 2025-01-14 17:20 UTC (permalink / raw)
To: David Hildenbrand, Lorenzo Stoakes
Cc: arnd, gregkh, Liam.Howlett, vbabka, jannh, willy, liushixin2,
akpm, linux-mm, linux-kernel
On 1/14/25 9:02 AM, David Hildenbrand wrote:
> On 14.01.25 16:06, David Hildenbrand wrote:
>> On 14.01.25 15:52, Lorenzo Stoakes wrote:
>>> On Tue, Jan 14, 2025 at 02:01:32PM +0100, David Hildenbrand wrote:
>>>> On 13.01.25 23:30, Yang Shi wrote:
>>>>> When creating private mapping for /dev/zero, the driver makes it an
>>>>> anonymous mapping by calling set_vma_anonymous(). But it just sets
>>>>> vm_ops to NULL, vm_file is still valid and vm_pgoff is also file
>>>>> offset.
>>>>>
>>>>> This is a special case and the VMA doesn't look like either
>>>>> anonymous VMA
>>>>> or file VMA. It confused other kernel subsystem, for example,
>>>>> khugepaged [1].
>>>>>
>>>>> It seems pointless to keep such special case. Making private
>>>>> /dev/zero>
>>>> mapping a full anonymous mapping doesn't change the semantic of
>>>>> /dev/zero either.
>>>>>
>>>>> The user visible effect is the mapping entry shown in
>>>>> /proc/<PID>/smaps
>>>>> and /proc/<PID>/maps.
>>>>>
>>>>> Before the change:
>>>>> ffffb7190000-ffffb7590000 rw-p 00001000 00:06
>>>>> 8 /dev/zero
>>>>>
>>>>> After the change:
>>>>> ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
>>>>>
>>>>
>>>> Hm, not sure about this. It's actually quite consistent to have
>>>> that output
>>>> in smaps the way it is. You mapped a file at an offset, and it
>>>> behaves like
>>>> an anonymous mapping apart from that.
>>>>
>>>> Not sure if the buggy khugepaged thing is a good indicator to
>>>> warrant this
>>>> change.
>>>
>>> Yeah, this is a user-facing fundamental change that hides
>>> information and
>>> defies expectation so I mean - it's a no go really isn't it?
>>>
>>> I'd rather we _not_ make this anon though, because isn't life confusing
>>> enough David? I thought it was bad enough with 'anon, file and lol
>>> shmem'
>>> but 'lol lol also /dev/zero' is enough to make me want to frolick in
>>> the
>>> fields...
>>
>> I recall there are users that rely on this memory to get the shared
>> zeropage on reads etc (in comparison to shmem!), so I better not ...
>> mess with this *at all* :)
>
> Heh, and I recall reading something about odd behavior of /dev/zero
> and some interesting history [1].
>
> "
> Unlike /dev/null, /dev/zero may be used as a source, not only as a
> sink for data. All write operations to /dev/zero succeed with no other
> effects. However, /dev/null is more commonly used for this purpose.
>
> When /dev/zero is memory-mapped, e.g., with mmap, to the virtual
> address space, it is equivalent to using anonymous memory; i.e. memory
> not connected to any file.
> "
>
> "equivalent to using anonymous memory" is interesting.
For private mapping. Shared mapping is equivalent to shmem.
>
>
> Also, /dev/zero was there before MAP_ANONYMOUS was invented according
> to [1], which is quite interesting.
Interesting... Didn't know this before.
>
> ... so this is anonymous memory as "real" as it can get :)
Let's make /dev/zero as real as anonymous memory :)
>
>
> [1] https://en.wikipedia.org/wiki//dev/zero
>
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 17:01 ` Yang Shi
@ 2025-01-14 17:23 ` David Hildenbrand
2025-01-14 17:38 ` Yang Shi
0 siblings, 1 reply; 35+ messages in thread
From: David Hildenbrand @ 2025-01-14 17:23 UTC (permalink / raw)
To: Yang Shi, Lorenzo Stoakes
Cc: arnd, gregkh, Liam.Howlett, vbabka, jannh, willy, liushixin2,
akpm, linux-mm, linux-kernel
On 14.01.25 18:01, Yang Shi wrote:
>
>
>
> On 1/14/25 7:06 AM, David Hildenbrand wrote:
>> On 14.01.25 15:52, Lorenzo Stoakes wrote:
>>> On Tue, Jan 14, 2025 at 02:01:32PM +0100, David Hildenbrand wrote:
>>>> On 13.01.25 23:30, Yang Shi wrote:
>>>>> When creating private mapping for /dev/zero, the driver makes it an
>>>>> anonymous mapping by calling set_vma_anonymous(). But it just sets
>>>>> vm_ops to NULL, vm_file is still valid and vm_pgoff is also file
>>>>> offset.
>>>>>
>>>>> This is a special case and the VMA doesn't look like either
>>>>> anonymous VMA
>>>>> or file VMA. It confused other kernel subsystem, for example,
>>>>> khugepaged [1].
>>>>>
>>>>> It seems pointless to keep such special case. Making private
>>>>> /dev/zero>
>>>> mapping a full anonymous mapping doesn't change the semantic of
>>>>> /dev/zero either.
>>>>>
>>>>> The user visible effect is the mapping entry shown in
>>>>> /proc/<PID>/smaps
>>>>> and /proc/<PID>/maps.
>>>>>
>>>>> Before the change:
>>>>> ffffb7190000-ffffb7590000 rw-p 00001000 00:06
>>>>> 8 /dev/zero
>>>>>
>>>>> After the change:
>>>>> ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
>>>>>
>>>>
>>>> Hm, not sure about this. It's actually quite consistent to have that
>>>> output
>>>> in smaps the way it is. You mapped a file at an offset, and it
>>>> behaves like
>>>> an anonymous mapping apart from that.
>>>>
>>>> Not sure if the buggy khugepaged thing is a good indicator to
>>>> warrant this
>>>> change.
>
> I admit this may be a concern, but I doubt who really care about it...
>
There is an example in the man page [1] about /proc/self/map_files/.
I assume that will also change here.
It's always hard to tell who that could affect, but I'm not convinced
this is worth it to find it out :)
>>>
>>> Yeah, this is a user-facing fundamental change that hides information
>>> and
>>> defies expectation so I mean - it's a no go really isn't it?
>>>
>>> I'd rather we _not_ make this anon though, because isn't life confusing
>>> enough David? I thought it was bad enough with 'anon, file and lol
>>> shmem'
>>> but 'lol lol also /dev/zero' is enough to make me want to frolick in the
>>> fields...
>>
>> I recall there are users that rely on this memory to get the shared
>> zeropage on reads etc (in comparison to shmem!), so I better not ...
>> mess with this *at all* :)
>
> The behavior won't be changed.
Yes, I know. And that's good ;)
[1] https://man7.org/linux/man-pages/man5/proc_pid_map_files.5.html
--
Cheers,
David / dhildenb
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 17:20 ` Yang Shi
@ 2025-01-14 17:24 ` David Hildenbrand
0 siblings, 0 replies; 35+ messages in thread
From: David Hildenbrand @ 2025-01-14 17:24 UTC (permalink / raw)
To: Yang Shi, Lorenzo Stoakes
Cc: arnd, gregkh, Liam.Howlett, vbabka, jannh, willy, liushixin2,
akpm, linux-mm, linux-kernel
On 14.01.25 18:20, Yang Shi wrote:
>
>
>
> On 1/14/25 9:02 AM, David Hildenbrand wrote:
>> On 14.01.25 16:06, David Hildenbrand wrote:
>>> On 14.01.25 15:52, Lorenzo Stoakes wrote:
>>>> On Tue, Jan 14, 2025 at 02:01:32PM +0100, David Hildenbrand wrote:
>>>>> On 13.01.25 23:30, Yang Shi wrote:
>>>>>> When creating private mapping for /dev/zero, the driver makes it an
>>>>>> anonymous mapping by calling set_vma_anonymous(). But it just sets
>>>>>> vm_ops to NULL, vm_file is still valid and vm_pgoff is also file
>>>>>> offset.
>>>>>>
>>>>>> This is a special case and the VMA doesn't look like either
>>>>>> anonymous VMA
>>>>>> or file VMA. It confused other kernel subsystem, for example,
>>>>>> khugepaged [1].
>>>>>>
>>>>>> It seems pointless to keep such special case. Making private
>>>>>> /dev/zero>
>>>>> mapping a full anonymous mapping doesn't change the semantic of
>>>>>> /dev/zero either.
>>>>>>
>>>>>> The user visible effect is the mapping entry shown in
>>>>>> /proc/<PID>/smaps
>>>>>> and /proc/<PID>/maps.
>>>>>>
>>>>>> Before the change:
>>>>>> ffffb7190000-ffffb7590000 rw-p 00001000 00:06
>>>>>> 8 /dev/zero
>>>>>>
>>>>>> After the change:
>>>>>> ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
>>>>>>
>>>>>
>>>>> Hm, not sure about this. It's actually quite consistent to have
>>>>> that output
>>>>> in smaps the way it is. You mapped a file at an offset, and it
>>>>> behaves like
>>>>> an anonymous mapping apart from that.
>>>>>
>>>>> Not sure if the buggy khugepaged thing is a good indicator to
>>>>> warrant this
>>>>> change.
>>>>
>>>> Yeah, this is a user-facing fundamental change that hides
>>>> information and
>>>> defies expectation so I mean - it's a no go really isn't it?
>>>>
>>>> I'd rather we _not_ make this anon though, because isn't life confusing
>>>> enough David? I thought it was bad enough with 'anon, file and lol
>>>> shmem'
>>>> but 'lol lol also /dev/zero' is enough to make me want to frolick in
>>>> the
>>>> fields...
>>>
>>> I recall there are users that rely on this memory to get the shared
>>> zeropage on reads etc (in comparison to shmem!), so I better not ...
>>> mess with this *at all* :)
>>
>> Heh, and I recall reading something about odd behavior of /dev/zero
>> and some interesting history [1].
>>
>> "
>> Unlike /dev/null, /dev/zero may be used as a source, not only as a
>> sink for data. All write operations to /dev/zero succeed with no other
>> effects. However, /dev/null is more commonly used for this purpose.
>>
>> When /dev/zero is memory-mapped, e.g., with mmap, to the virtual
>> address space, it is equivalent to using anonymous memory; i.e. memory
>> not connected to any file.
>> "
>>
>> "equivalent to using anonymous memory" is interesting.
>
> For private mapping. Shared mapping is equivalent to shmem.
"shared anonymous memory", yes.
--
Cheers,
David / dhildenb
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 17:23 ` David Hildenbrand
@ 2025-01-14 17:38 ` Yang Shi
2025-01-14 17:46 ` David Hildenbrand
0 siblings, 1 reply; 35+ messages in thread
From: Yang Shi @ 2025-01-14 17:38 UTC (permalink / raw)
To: David Hildenbrand, Lorenzo Stoakes
Cc: arnd, gregkh, Liam.Howlett, vbabka, jannh, willy, liushixin2,
akpm, linux-mm, linux-kernel
On 1/14/25 9:23 AM, David Hildenbrand wrote:
> On 14.01.25 18:01, Yang Shi wrote:
>>
>>
>>
>> On 1/14/25 7:06 AM, David Hildenbrand wrote:
>>> On 14.01.25 15:52, Lorenzo Stoakes wrote:
>>>> On Tue, Jan 14, 2025 at 02:01:32PM +0100, David Hildenbrand wrote:
>>>>> On 13.01.25 23:30, Yang Shi wrote:
>>>>>> When creating private mapping for /dev/zero, the driver makes it an
>>>>>> anonymous mapping by calling set_vma_anonymous(). But it just sets
>>>>>> vm_ops to NULL, vm_file is still valid and vm_pgoff is also file
>>>>>> offset.
>>>>>>
>>>>>> This is a special case and the VMA doesn't look like either
>>>>>> anonymous VMA
>>>>>> or file VMA. It confused other kernel subsystem, for example,
>>>>>> khugepaged [1].
>>>>>>
>>>>>> It seems pointless to keep such special case. Making private
>>>>>> /dev/zero>
>>>>> mapping a full anonymous mapping doesn't change the semantic of
>>>>>> /dev/zero either.
>>>>>>
>>>>>> The user visible effect is the mapping entry shown in
>>>>>> /proc/<PID>/smaps
>>>>>> and /proc/<PID>/maps.
>>>>>>
>>>>>> Before the change:
>>>>>> ffffb7190000-ffffb7590000 rw-p 00001000 00:06
>>>>>> 8 /dev/zero
>>>>>>
>>>>>> After the change:
>>>>>> ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
>>>>>>
>>>>>
>>>>> Hm, not sure about this. It's actually quite consistent to have that
>>>>> output
>>>>> in smaps the way it is. You mapped a file at an offset, and it
>>>>> behaves like
>>>>> an anonymous mapping apart from that.
>>>>>
>>>>> Not sure if the buggy khugepaged thing is a good indicator to
>>>>> warrant this
>>>>> change.
>>
>> I admit this may be a concern, but I doubt who really care about it...
>>
>
> There is an example in the man page [1] about /proc/self/map_files/.
>
> I assume that will also change here.
IIUC, that example is specific to "anonymous shared memory" created by
shared mapping of /dev/zero.
>
> It's always hard to tell who that could affect, but I'm not convinced
> this is worth it to find it out :)
>
>>>>
>>>> Yeah, this is a user-facing fundamental change that hides information
>>>> and
>>>> defies expectation so I mean - it's a no go really isn't it?
>>>>
>>>> I'd rather we _not_ make this anon though, because isn't life
>>>> confusing
>>>> enough David? I thought it was bad enough with 'anon, file and lol
>>>> shmem'
>>>> but 'lol lol also /dev/zero' is enough to make me want to frolick
>>>> in the
>>>> fields...
>>>
>>> I recall there are users that rely on this memory to get the shared
>>> zeropage on reads etc (in comparison to shmem!), so I better not ...
>>> mess with this *at all* :)
>>
>> The behavior won't be changed.
>
> Yes, I know. And that's good ;)
>
>
> [1] https://man7.org/linux/man-pages/man5/proc_pid_map_files.5.html
>
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 17:38 ` Yang Shi
@ 2025-01-14 17:46 ` David Hildenbrand
2025-01-14 18:05 ` Yang Shi
0 siblings, 1 reply; 35+ messages in thread
From: David Hildenbrand @ 2025-01-14 17:46 UTC (permalink / raw)
To: Yang Shi, Lorenzo Stoakes
Cc: arnd, gregkh, Liam.Howlett, vbabka, jannh, willy, liushixin2,
akpm, linux-mm, linux-kernel
On 14.01.25 18:38, Yang Shi wrote:
>
>
>
> On 1/14/25 9:23 AM, David Hildenbrand wrote:
>> On 14.01.25 18:01, Yang Shi wrote:
>>>
>>>
>>>
>>> On 1/14/25 7:06 AM, David Hildenbrand wrote:
>>>> On 14.01.25 15:52, Lorenzo Stoakes wrote:
>>>>> On Tue, Jan 14, 2025 at 02:01:32PM +0100, David Hildenbrand wrote:
>>>>>> On 13.01.25 23:30, Yang Shi wrote:
>>>>>>> When creating private mapping for /dev/zero, the driver makes it an
>>>>>>> anonymous mapping by calling set_vma_anonymous(). But it just sets
>>>>>>> vm_ops to NULL, vm_file is still valid and vm_pgoff is also file
>>>>>>> offset.
>>>>>>>
>>>>>>> This is a special case and the VMA doesn't look like either
>>>>>>> anonymous VMA
>>>>>>> or file VMA. It confused other kernel subsystem, for example,
>>>>>>> khugepaged [1].
>>>>>>>
>>>>>>> It seems pointless to keep such special case. Making private
>>>>>>> /dev/zero>
>>>>>> mapping a full anonymous mapping doesn't change the semantic of
>>>>>>> /dev/zero either.
>>>>>>>
>>>>>>> The user visible effect is the mapping entry shown in
>>>>>>> /proc/<PID>/smaps
>>>>>>> and /proc/<PID>/maps.
>>>>>>>
>>>>>>> Before the change:
>>>>>>> ffffb7190000-ffffb7590000 rw-p 00001000 00:06
>>>>>>> 8 /dev/zero
>>>>>>>
>>>>>>> After the change:
>>>>>>> ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
>>>>>>>
>>>>>>
>>>>>> Hm, not sure about this. It's actually quite consistent to have that
>>>>>> output
>>>>>> in smaps the way it is. You mapped a file at an offset, and it
>>>>>> behaves like
>>>>>> an anonymous mapping apart from that.
>>>>>>
>>>>>> Not sure if the buggy khugepaged thing is a good indicator to
>>>>>> warrant this
>>>>>> change.
>>>
>>> I admit this may be a concern, but I doubt who really care about it...
>>>
>>
>> There is an example in the man page [1] about /proc/self/map_files/.
>>
>> I assume that will also change here.
>
> IIUC, that example is specific to "anonymous shared memory" created by
> shared mapping of /dev/zero.
Note that MAP_PRIVATE of /dev/zero will also make it appear in the same
way right now (I just tried).
The example is about MAP_FILE in general, not just MAP_SHARED IIUC.
--
Cheers,
David / dhildenb
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 17:46 ` David Hildenbrand
@ 2025-01-14 18:05 ` Yang Shi
0 siblings, 0 replies; 35+ messages in thread
From: Yang Shi @ 2025-01-14 18:05 UTC (permalink / raw)
To: David Hildenbrand, Lorenzo Stoakes
Cc: arnd, gregkh, Liam.Howlett, vbabka, jannh, willy, liushixin2,
akpm, linux-mm, linux-kernel
On 1/14/25 9:46 AM, David Hildenbrand wrote:
> On 14.01.25 18:38, Yang Shi wrote:
>>
>>
>>
>> On 1/14/25 9:23 AM, David Hildenbrand wrote:
>>> On 14.01.25 18:01, Yang Shi wrote:
>>>>
>>>>
>>>>
>>>> On 1/14/25 7:06 AM, David Hildenbrand wrote:
>>>>> On 14.01.25 15:52, Lorenzo Stoakes wrote:
>>>>>> On Tue, Jan 14, 2025 at 02:01:32PM +0100, David Hildenbrand wrote:
>>>>>>> On 13.01.25 23:30, Yang Shi wrote:
>>>>>>>> When creating private mapping for /dev/zero, the driver makes
>>>>>>>> it an
>>>>>>>> anonymous mapping by calling set_vma_anonymous(). But it just sets
>>>>>>>> vm_ops to NULL, vm_file is still valid and vm_pgoff is also file
>>>>>>>> offset.
>>>>>>>>
>>>>>>>> This is a special case and the VMA doesn't look like either
>>>>>>>> anonymous VMA
>>>>>>>> or file VMA. It confused other kernel subsystem, for example,
>>>>>>>> khugepaged [1].
>>>>>>>>
>>>>>>>> It seems pointless to keep such special case. Making private
>>>>>>>> /dev/zero>
>>>>>>> mapping a full anonymous mapping doesn't change the semantic of
>>>>>>>> /dev/zero either.
>>>>>>>>
>>>>>>>> The user visible effect is the mapping entry shown in
>>>>>>>> /proc/<PID>/smaps
>>>>>>>> and /proc/<PID>/maps.
>>>>>>>>
>>>>>>>> Before the change:
>>>>>>>> ffffb7190000-ffffb7590000 rw-p 00001000 00:06
>>>>>>>> 8 /dev/zero
>>>>>>>>
>>>>>>>> After the change:
>>>>>>>> ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
>>>>>>>>
>>>>>>>
>>>>>>> Hm, not sure about this. It's actually quite consistent to have
>>>>>>> that
>>>>>>> output
>>>>>>> in smaps the way it is. You mapped a file at an offset, and it
>>>>>>> behaves like
>>>>>>> an anonymous mapping apart from that.
>>>>>>>
>>>>>>> Not sure if the buggy khugepaged thing is a good indicator to
>>>>>>> warrant this
>>>>>>> change.
>>>>
>>>> I admit this may be a concern, but I doubt who really care about it...
>>>>
>>>
>>> There is an example in the man page [1] about /proc/self/map_files/.
>>>
>>> I assume that will also change here.
>>
>> IIUC, that example is specific to "anonymous shared memory" created by
>> shared mapping of /dev/zero.
>
> Note that MAP_PRIVATE of /dev/zero will also make it appear in the
> same way right now (I just tried).
Yes, I will add this in the commit log as another user visible change.
>
> The example is about MAP_FILE in general, not just MAP_SHARED IIUC.
MAP_FILE is actually ignored on Linux per
https://man7.org/linux/man-pages/man2/mmap.2.html. It also says
"(regions created with the MAP_ANON | MAP_SHARED flags)". Anyway it
looks like this man page may be a little bit outdated. We can clean it
up later.
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 16:53 ` Yang Shi
@ 2025-01-14 18:14 ` Lorenzo Stoakes
2025-01-14 18:19 ` Lorenzo Stoakes
` (2 more replies)
0 siblings, 3 replies; 35+ messages in thread
From: Lorenzo Stoakes @ 2025-01-14 18:14 UTC (permalink / raw)
To: Yang Shi
Cc: arnd, gregkh, Liam.Howlett, vbabka, jannh, willy, liushixin2,
akpm, linux-mm, linux-kernel
This is getting into realms of discussion so to risk sounding rude - to be
clear - NACK.
The user-visible change to /proc/$pid/[s]maps kills this patch dead. This
is regardless of any other discussed issue.
But more importantly, I hadn't realise mmap_zero() was on the .mmap()
callback (sorry my mistake) - you're simply not permitted to change
vm_pgoff and vm_file fields here, the mapping logic doesn't expect it, and
it's broken.
To me the alternative would be to have a custom fault handler that hands
back the zero page, but I"m not sure that's workable, you'd have to install
a special mapping etc. and huge pages are weird and...
I do appreciate you raising this especially as I was blissfully unaware,
but I don't see how this patch can possibly work, sorry :(
On Tue, Jan 14, 2025 at 08:53:01AM -0800, Yang Shi wrote:
>
>
>
> On 1/14/25 4:05 AM, Lorenzo Stoakes wrote:
> > + Willy for the fs/weirdness elements of this.
> >
> > On Mon, Jan 13, 2025 at 02:30:33PM -0800, Yang Shi wrote:
> > > When creating private mapping for /dev/zero, the driver makes it an
> > > anonymous mapping by calling set_vma_anonymous(). But it just sets
> > > vm_ops to NULL, vm_file is still valid and vm_pgoff is also file offset.
> > Hm yikes.
> >
> > > This is a special case and the VMA doesn't look like either anonymous VMA
> > > or file VMA. It confused other kernel subsystem, for example, khugepaged [1].
> > >
> > > It seems pointless to keep such special case. Making private /dev/zero
> > > mapping a full anonymous mapping doesn't change the semantic of
> > > /dev/zero either.
> > My concern is that ostensibly there _is_ a file right? Are we certain that by
> > not setting this we are not breaking something somewhere else?
> >
> > Are we not creating a sort of other type of 'non-such-beast' here?
>
> But the file is /dev/zero. I don't see this could break the semantic of
> /dev/zero. The shared mapping of /dev/zero is not affected by this change,
> kernel already treated private mapping of /dev/zero as anonymous mapping,
> but with some weird settings in VMA. When reading the mapping, it returns 0
> with zero page, when writing the mapping, a new anonymous folio is
> allocated.
You're creating a new concept of an anon but not anon but also now with
anon vm_pgoff and missing vm_file even though it does reference a file
and... yeah.
This is not usual :)
>
> >
> > I mean already setting it anon and setting vm_file non-NULL is really strange.
> >
> > > The user visible effect is the mapping entry shown in /proc/<PID>/smaps
> > > and /proc/<PID>/maps.
> > >
> > > Before the change:
> > > ffffb7190000-ffffb7590000 rw-p 00001000 00:06 8 /dev/zero
> > >
> > > After the change:
> > > ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
> > >
> > Yeah this seems like it might break somebody to be honest, it's really
> > really really strange to map a file then for it not to be mapped.
>
> Yes, it is possible if someone really care whether the anonymous-like
> mapping is mapped by /dev/zero or just created by malloc(). But I don't know
> who really do...
>
> >
> > But it's possibly EVEN WEIRDER to map a file and for it to seem mapped as a
> > file but for it to be marked anonymous.
> >
> > God what a mess.
> >
> > > [1]: https://lore.kernel.org/linux-mm/20250111034511.2223353-1-liushixin2@huawei.com/
> > I kind of hate that we have to mitigate like this for a case that should
> > never ever happen so I'm inclined towards your solution but a lot more
> > inclined towards us totally rethinking this.
> >
> > Do we _have_ to make this anonymous?? Why can't we just reference the zero
> > page as if it were in the page cache (Willy - feel free to correct naive
> > misapprehension here).
>
> TBH, I don't see why page cache has to be involved. When reading, 0 is
> returned by zero page. When writing a CoW is triggered if page cache is
> involved, but the content of the page cache should be just 0, so we copy 0
> to the new folio then write to it. It doesn't make too much sense. I think
> this is why private /dev/zero mapping is treated as anonymous mapping in the
> first place.
I'm obviously not suggesting allocating a bunch of extra folios, I was
thinking there would be some means of handing back the actual zero
page. But I am not sure this is workable.
>
> >
> > > Signed-off-by: Yang Shi <yang@os.amperecomputing.com>
> > > ---
> > > drivers/char/mem.c | 4 ++++
> > > 1 file changed, 4 insertions(+)
> > >
> > > diff --git a/drivers/char/mem.c b/drivers/char/mem.c
> > > index 169eed162a7f..dae113f7fc1b 100644
> > > --- a/drivers/char/mem.c
> > > +++ b/drivers/char/mem.c
> > > @@ -527,6 +527,10 @@ static int mmap_zero(struct file *file, struct vm_area_struct *vma)
> > > if (vma->vm_flags & VM_SHARED)
> > > return shmem_zero_setup(vma);
> > > vma_set_anonymous(vma);
> > > + fput(vma->vm_file);
> > > + vma->vm_file = NULL;
> > > + vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT;
This is just not permitted. We maintain mmap state which contains the file
and pgoff state which gets threaded through the mapping operation, and
simply do not expect you to change these fields.
In future we will assert on this or preferably, restrict users to only
changing VMA flags, the private field and vm_ops.
> > Hmm, this might have been mremap()'d _potentially_ though? And then now
> > this will be wrong? But then we'd have no way of tracking it correctly...
>
> I'm not quite familiar with the subtle details and corner cases of
> meremap(). But mmap_zero() should be called by mmap(), so the VMA has not
> been visible to user yet at this point IIUC. How come mremap() could move
> it?
Ah OK, in that case fine on that front.
But you are not permitted to touch this field (we need to enforce this...)
>
> >
> > I've not checked the function but do we mark this as a special mapping of
> > some kind?
> >
> > > +
> > > return 0;
> > > }
> > >
> > > --
> > > 2.47.0
> > >
>
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 18:14 ` Lorenzo Stoakes
@ 2025-01-14 18:19 ` Lorenzo Stoakes
2025-01-14 18:21 ` Lorenzo Stoakes
2025-01-14 18:22 ` Matthew Wilcox
2025-01-14 18:32 ` Jann Horn
2025-01-14 19:03 ` Yang Shi
2 siblings, 2 replies; 35+ messages in thread
From: Lorenzo Stoakes @ 2025-01-14 18:19 UTC (permalink / raw)
To: Yang Shi
Cc: arnd, gregkh, Liam.Howlett, vbabka, jannh, willy, liushixin2,
akpm, linux-mm, linux-kernel
On Tue, Jan 14, 2025 at 06:14:57PM +0000, Lorenzo Stoakes wrote:
> This is getting into realms of discussion so to risk sounding rude - to be
> clear - NACK.
>
> The user-visible change to /proc/$pid/[s]maps kills this patch dead. This
> is regardless of any other discussed issue.
>
> But more importantly, I hadn't realise mmap_zero() was on the .mmap()
> callback (sorry my mistake) - you're simply not permitted to change
> vm_pgoff and vm_file fields here, the mapping logic doesn't expect it, and
> it's broken.
I see shmem_zero_page() does change vma->vm_page, this is broken... ugh. I
will audit this code (and have a look through _all_ mmap() callbacks I
guess). Duly added to TODO. But definitely can't have _another_ case of
doing this.
>
> To me the alternative would be to have a custom fault handler that hands
> back the zero page, but I"m not sure that's workable, you'd have to install
> a special mapping etc. and huge pages are weird and...
>
> I do appreciate you raising this especially as I was blissfully unaware,
> but I don't see how this patch can possibly work, sorry :(
>
> On Tue, Jan 14, 2025 at 08:53:01AM -0800, Yang Shi wrote:
> >
> >
> >
> > On 1/14/25 4:05 AM, Lorenzo Stoakes wrote:
> > > + Willy for the fs/weirdness elements of this.
> > >
> > > On Mon, Jan 13, 2025 at 02:30:33PM -0800, Yang Shi wrote:
> > > > When creating private mapping for /dev/zero, the driver makes it an
> > > > anonymous mapping by calling set_vma_anonymous(). But it just sets
> > > > vm_ops to NULL, vm_file is still valid and vm_pgoff is also file offset.
> > > Hm yikes.
> > >
> > > > This is a special case and the VMA doesn't look like either anonymous VMA
> > > > or file VMA. It confused other kernel subsystem, for example, khugepaged [1].
> > > >
> > > > It seems pointless to keep such special case. Making private /dev/zero
> > > > mapping a full anonymous mapping doesn't change the semantic of
> > > > /dev/zero either.
> > > My concern is that ostensibly there _is_ a file right? Are we certain that by
> > > not setting this we are not breaking something somewhere else?
> > >
> > > Are we not creating a sort of other type of 'non-such-beast' here?
> >
> > But the file is /dev/zero. I don't see this could break the semantic of
> > /dev/zero. The shared mapping of /dev/zero is not affected by this change,
> > kernel already treated private mapping of /dev/zero as anonymous mapping,
> > but with some weird settings in VMA. When reading the mapping, it returns 0
> > with zero page, when writing the mapping, a new anonymous folio is
> > allocated.
>
> You're creating a new concept of an anon but not anon but also now with
> anon vm_pgoff and missing vm_file even though it does reference a file
> and... yeah.
>
> This is not usual :)
>
> >
> > >
> > > I mean already setting it anon and setting vm_file non-NULL is really strange.
> > >
> > > > The user visible effect is the mapping entry shown in /proc/<PID>/smaps
> > > > and /proc/<PID>/maps.
> > > >
> > > > Before the change:
> > > > ffffb7190000-ffffb7590000 rw-p 00001000 00:06 8 /dev/zero
> > > >
> > > > After the change:
> > > > ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
> > > >
> > > Yeah this seems like it might break somebody to be honest, it's really
> > > really really strange to map a file then for it not to be mapped.
> >
> > Yes, it is possible if someone really care whether the anonymous-like
> > mapping is mapped by /dev/zero or just created by malloc(). But I don't know
> > who really do...
> >
> > >
> > > But it's possibly EVEN WEIRDER to map a file and for it to seem mapped as a
> > > file but for it to be marked anonymous.
> > >
> > > God what a mess.
> > >
> > > > [1]: https://lore.kernel.org/linux-mm/20250111034511.2223353-1-liushixin2@huawei.com/
> > > I kind of hate that we have to mitigate like this for a case that should
> > > never ever happen so I'm inclined towards your solution but a lot more
> > > inclined towards us totally rethinking this.
> > >
> > > Do we _have_ to make this anonymous?? Why can't we just reference the zero
> > > page as if it were in the page cache (Willy - feel free to correct naive
> > > misapprehension here).
> >
> > TBH, I don't see why page cache has to be involved. When reading, 0 is
> > returned by zero page. When writing a CoW is triggered if page cache is
> > involved, but the content of the page cache should be just 0, so we copy 0
> > to the new folio then write to it. It doesn't make too much sense. I think
> > this is why private /dev/zero mapping is treated as anonymous mapping in the
> > first place.
>
> I'm obviously not suggesting allocating a bunch of extra folios, I was
> thinking there would be some means of handing back the actual zero
> page. But I am not sure this is workable.
>
> >
> > >
> > > > Signed-off-by: Yang Shi <yang@os.amperecomputing.com>
> > > > ---
> > > > drivers/char/mem.c | 4 ++++
> > > > 1 file changed, 4 insertions(+)
> > > >
> > > > diff --git a/drivers/char/mem.c b/drivers/char/mem.c
> > > > index 169eed162a7f..dae113f7fc1b 100644
> > > > --- a/drivers/char/mem.c
> > > > +++ b/drivers/char/mem.c
> > > > @@ -527,6 +527,10 @@ static int mmap_zero(struct file *file, struct vm_area_struct *vma)
> > > > if (vma->vm_flags & VM_SHARED)
> > > > return shmem_zero_setup(vma);
> > > > vma_set_anonymous(vma);
> > > > + fput(vma->vm_file);
> > > > + vma->vm_file = NULL;
> > > > + vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT;
>
> This is just not permitted. We maintain mmap state which contains the file
> and pgoff state which gets threaded through the mapping operation, and
> simply do not expect you to change these fields.
>
> In future we will assert on this or preferably, restrict users to only
> changing VMA flags, the private field and vm_ops.
>
> > > Hmm, this might have been mremap()'d _potentially_ though? And then now
> > > this will be wrong? But then we'd have no way of tracking it correctly...
> >
> > I'm not quite familiar with the subtle details and corner cases of
> > meremap(). But mmap_zero() should be called by mmap(), so the VMA has not
> > been visible to user yet at this point IIUC. How come mremap() could move
> > it?
>
> Ah OK, in that case fine on that front.
>
> But you are not permitted to touch this field (we need to enforce this...)
>
> >
> > >
> > > I've not checked the function but do we mark this as a special mapping of
> > > some kind?
> > >
> > > > +
> > > > return 0;
> > > > }
> > > >
> > > > --
> > > > 2.47.0
> > > >
> >
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 18:19 ` Lorenzo Stoakes
@ 2025-01-14 18:21 ` Lorenzo Stoakes
2025-01-14 18:22 ` Matthew Wilcox
1 sibling, 0 replies; 35+ messages in thread
From: Lorenzo Stoakes @ 2025-01-14 18:21 UTC (permalink / raw)
To: Yang Shi
Cc: arnd, gregkh, Liam.Howlett, vbabka, jannh, willy, liushixin2,
akpm, linux-mm, linux-kernel
On Tue, Jan 14, 2025 at 06:19:32PM +0000, Lorenzo Stoakes wrote:
> On Tue, Jan 14, 2025 at 06:14:57PM +0000, Lorenzo Stoakes wrote:
> > This is getting into realms of discussion so to risk sounding rude - to be
> > clear - NACK.
> >
> > The user-visible change to /proc/$pid/[s]maps kills this patch dead. This
> > is regardless of any other discussed issue.
> >
> > But more importantly, I hadn't realise mmap_zero() was on the .mmap()
> > callback (sorry my mistake) - you're simply not permitted to change
> > vm_pgoff and vm_file fields here, the mapping logic doesn't expect it, and
> > it's broken.
>
> I see shmem_zero_page() does change vma->vm_page, this is broken... ugh. I
> will audit this code (and have a look through _all_ mmap() callbacks I
> guess). Duly added to TODO. But definitely can't have _another_ case of
> doing this.
* vma->vm_file... it is late here :)
>
> >
> > To me the alternative would be to have a custom fault handler that hands
> > back the zero page, but I"m not sure that's workable, you'd have to install
> > a special mapping etc. and huge pages are weird and...
> >
> > I do appreciate you raising this especially as I was blissfully unaware,
> > but I don't see how this patch can possibly work, sorry :(
> >
> > On Tue, Jan 14, 2025 at 08:53:01AM -0800, Yang Shi wrote:
> > >
> > >
> > >
> > > On 1/14/25 4:05 AM, Lorenzo Stoakes wrote:
> > > > + Willy for the fs/weirdness elements of this.
> > > >
> > > > On Mon, Jan 13, 2025 at 02:30:33PM -0800, Yang Shi wrote:
> > > > > When creating private mapping for /dev/zero, the driver makes it an
> > > > > anonymous mapping by calling set_vma_anonymous(). But it just sets
> > > > > vm_ops to NULL, vm_file is still valid and vm_pgoff is also file offset.
> > > > Hm yikes.
> > > >
> > > > > This is a special case and the VMA doesn't look like either anonymous VMA
> > > > > or file VMA. It confused other kernel subsystem, for example, khugepaged [1].
> > > > >
> > > > > It seems pointless to keep such special case. Making private /dev/zero
> > > > > mapping a full anonymous mapping doesn't change the semantic of
> > > > > /dev/zero either.
> > > > My concern is that ostensibly there _is_ a file right? Are we certain that by
> > > > not setting this we are not breaking something somewhere else?
> > > >
> > > > Are we not creating a sort of other type of 'non-such-beast' here?
> > >
> > > But the file is /dev/zero. I don't see this could break the semantic of
> > > /dev/zero. The shared mapping of /dev/zero is not affected by this change,
> > > kernel already treated private mapping of /dev/zero as anonymous mapping,
> > > but with some weird settings in VMA. When reading the mapping, it returns 0
> > > with zero page, when writing the mapping, a new anonymous folio is
> > > allocated.
> >
> > You're creating a new concept of an anon but not anon but also now with
> > anon vm_pgoff and missing vm_file even though it does reference a file
> > and... yeah.
> >
> > This is not usual :)
> >
> > >
> > > >
> > > > I mean already setting it anon and setting vm_file non-NULL is really strange.
> > > >
> > > > > The user visible effect is the mapping entry shown in /proc/<PID>/smaps
> > > > > and /proc/<PID>/maps.
> > > > >
> > > > > Before the change:
> > > > > ffffb7190000-ffffb7590000 rw-p 00001000 00:06 8 /dev/zero
> > > > >
> > > > > After the change:
> > > > > ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
> > > > >
> > > > Yeah this seems like it might break somebody to be honest, it's really
> > > > really really strange to map a file then for it not to be mapped.
> > >
> > > Yes, it is possible if someone really care whether the anonymous-like
> > > mapping is mapped by /dev/zero or just created by malloc(). But I don't know
> > > who really do...
> > >
> > > >
> > > > But it's possibly EVEN WEIRDER to map a file and for it to seem mapped as a
> > > > file but for it to be marked anonymous.
> > > >
> > > > God what a mess.
> > > >
> > > > > [1]: https://lore.kernel.org/linux-mm/20250111034511.2223353-1-liushixin2@huawei.com/
> > > > I kind of hate that we have to mitigate like this for a case that should
> > > > never ever happen so I'm inclined towards your solution but a lot more
> > > > inclined towards us totally rethinking this.
> > > >
> > > > Do we _have_ to make this anonymous?? Why can't we just reference the zero
> > > > page as if it were in the page cache (Willy - feel free to correct naive
> > > > misapprehension here).
> > >
> > > TBH, I don't see why page cache has to be involved. When reading, 0 is
> > > returned by zero page. When writing a CoW is triggered if page cache is
> > > involved, but the content of the page cache should be just 0, so we copy 0
> > > to the new folio then write to it. It doesn't make too much sense. I think
> > > this is why private /dev/zero mapping is treated as anonymous mapping in the
> > > first place.
> >
> > I'm obviously not suggesting allocating a bunch of extra folios, I was
> > thinking there would be some means of handing back the actual zero
> > page. But I am not sure this is workable.
> >
> > >
> > > >
> > > > > Signed-off-by: Yang Shi <yang@os.amperecomputing.com>
> > > > > ---
> > > > > drivers/char/mem.c | 4 ++++
> > > > > 1 file changed, 4 insertions(+)
> > > > >
> > > > > diff --git a/drivers/char/mem.c b/drivers/char/mem.c
> > > > > index 169eed162a7f..dae113f7fc1b 100644
> > > > > --- a/drivers/char/mem.c
> > > > > +++ b/drivers/char/mem.c
> > > > > @@ -527,6 +527,10 @@ static int mmap_zero(struct file *file, struct vm_area_struct *vma)
> > > > > if (vma->vm_flags & VM_SHARED)
> > > > > return shmem_zero_setup(vma);
> > > > > vma_set_anonymous(vma);
> > > > > + fput(vma->vm_file);
> > > > > + vma->vm_file = NULL;
> > > > > + vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT;
> >
> > This is just not permitted. We maintain mmap state which contains the file
> > and pgoff state which gets threaded through the mapping operation, and
> > simply do not expect you to change these fields.
> >
> > In future we will assert on this or preferably, restrict users to only
> > changing VMA flags, the private field and vm_ops.
> >
> > > > Hmm, this might have been mremap()'d _potentially_ though? And then now
> > > > this will be wrong? But then we'd have no way of tracking it correctly...
> > >
> > > I'm not quite familiar with the subtle details and corner cases of
> > > meremap(). But mmap_zero() should be called by mmap(), so the VMA has not
> > > been visible to user yet at this point IIUC. How come mremap() could move
> > > it?
> >
> > Ah OK, in that case fine on that front.
> >
> > But you are not permitted to touch this field (we need to enforce this...)
> >
> > >
> > > >
> > > > I've not checked the function but do we mark this as a special mapping of
> > > > some kind?
> > > >
> > > > > +
> > > > > return 0;
> > > > > }
> > > > >
> > > > > --
> > > > > 2.47.0
> > > > >
> > >
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 18:19 ` Lorenzo Stoakes
2025-01-14 18:21 ` Lorenzo Stoakes
@ 2025-01-14 18:22 ` Matthew Wilcox
2025-01-14 18:26 ` Lorenzo Stoakes
1 sibling, 1 reply; 35+ messages in thread
From: Matthew Wilcox @ 2025-01-14 18:22 UTC (permalink / raw)
To: Lorenzo Stoakes
Cc: Yang Shi, arnd, gregkh, Liam.Howlett, vbabka, jannh, liushixin2,
akpm, linux-mm, linux-kernel
On Tue, Jan 14, 2025 at 06:19:32PM +0000, Lorenzo Stoakes wrote:
> I see shmem_zero_page() does change vma->vm_page, this is broken... ugh. I
I think you mean shmem_zero_setup() and vma->vm_file, right?
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 18:22 ` Matthew Wilcox
@ 2025-01-14 18:26 ` Lorenzo Stoakes
0 siblings, 0 replies; 35+ messages in thread
From: Lorenzo Stoakes @ 2025-01-14 18:26 UTC (permalink / raw)
To: Matthew Wilcox
Cc: Yang Shi, arnd, gregkh, Liam.Howlett, vbabka, jannh, liushixin2,
akpm, linux-mm, linux-kernel
On Tue, Jan 14, 2025 at 06:22:14PM +0000, Matthew Wilcox wrote:
> On Tue, Jan 14, 2025 at 06:19:32PM +0000, Lorenzo Stoakes wrote:
> > I see shmem_zero_page() does change vma->vm_page, this is broken... ugh. I
>
> I think you mean shmem_zero_setup() and vma->vm_file, right?
Yes, correct. Sorry it's late here and it's showing haha!
The reason I am concerned about this is because we thread mmap state
through the operation which has a separate file pointer which this makes
into a potential UAF.
Will audit all this and for any other problematic .mmap() callback
behaviour.
My view is ideally this should be a callback with a const pointer to the
VMA (or some other mechanism, perhaps) which accepts a change in
_permitted_ fields only.
The 'anything could happen and anybody could manipulate any field of the
VMA' in this callback is highly problematic.
But we definitely shouldn't be adding a _new_ case here.
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 18:14 ` Lorenzo Stoakes
2025-01-14 18:19 ` Lorenzo Stoakes
@ 2025-01-14 18:32 ` Jann Horn
2025-01-14 18:38 ` Lorenzo Stoakes
2025-01-14 19:03 ` Yang Shi
2 siblings, 1 reply; 35+ messages in thread
From: Jann Horn @ 2025-01-14 18:32 UTC (permalink / raw)
To: Lorenzo Stoakes
Cc: Yang Shi, arnd, gregkh, Liam.Howlett, vbabka, willy, liushixin2,
akpm, linux-mm, linux-kernel
On Tue, Jan 14, 2025 at 7:15 PM Lorenzo Stoakes
<lorenzo.stoakes@oracle.com> wrote:
> On Tue, Jan 14, 2025 at 08:53:01AM -0800, Yang Shi wrote:
> > On 1/14/25 4:05 AM, Lorenzo Stoakes wrote:
> > > On Mon, Jan 13, 2025 at 02:30:33PM -0800, Yang Shi wrote:
> > > > + fput(vma->vm_file);
> > > > + vma->vm_file = NULL;
> > > > + vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT;
>
> This is just not permitted. We maintain mmap state which contains the file
> and pgoff state which gets threaded through the mapping operation, and
> simply do not expect you to change these fields.
>
> In future we will assert on this or preferably, restrict users to only
> changing VMA flags, the private field and vm_ops.
>
> > > Hmm, this might have been mremap()'d _potentially_ though? And then now
> > > this will be wrong? But then we'd have no way of tracking it correctly...
> >
> > I'm not quite familiar with the subtle details and corner cases of
> > meremap(). But mmap_zero() should be called by mmap(), so the VMA has not
> > been visible to user yet at this point IIUC. How come mremap() could move
> > it?
>
> Ah OK, in that case fine on that front.
>
> But you are not permitted to touch this field (we need to enforce this...)
Sidenote: I think the GPU DRM subsystem relies on changing pgoff in
some of their mmap handlers; maybe talk to them about this if you
haven't already. See for example drm_gem_prime_mmap() and
dma_buf_mmap().
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 18:32 ` Jann Horn
@ 2025-01-14 18:38 ` Lorenzo Stoakes
0 siblings, 0 replies; 35+ messages in thread
From: Lorenzo Stoakes @ 2025-01-14 18:38 UTC (permalink / raw)
To: Jann Horn
Cc: Yang Shi, arnd, gregkh, Liam.Howlett, vbabka, willy, liushixin2,
akpm, linux-mm, linux-kernel
On Tue, Jan 14, 2025 at 07:32:51PM +0100, Jann Horn wrote:
> On Tue, Jan 14, 2025 at 7:15 PM Lorenzo Stoakes
> <lorenzo.stoakes@oracle.com> wrote:
> > On Tue, Jan 14, 2025 at 08:53:01AM -0800, Yang Shi wrote:
> > > On 1/14/25 4:05 AM, Lorenzo Stoakes wrote:
> > > > On Mon, Jan 13, 2025 at 02:30:33PM -0800, Yang Shi wrote:
> > > > > + fput(vma->vm_file);
> > > > > + vma->vm_file = NULL;
> > > > > + vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT;
> >
> > This is just not permitted. We maintain mmap state which contains the file
> > and pgoff state which gets threaded through the mapping operation, and
> > simply do not expect you to change these fields.
> >
> > In future we will assert on this or preferably, restrict users to only
> > changing VMA flags, the private field and vm_ops.
> >
> > > > Hmm, this might have been mremap()'d _potentially_ though? And then now
> > > > this will be wrong? But then we'd have no way of tracking it correctly...
> > >
> > > I'm not quite familiar with the subtle details and corner cases of
> > > meremap(). But mmap_zero() should be called by mmap(), so the VMA has not
> > > been visible to user yet at this point IIUC. How come mremap() could move
> > > it?
> >
> > Ah OK, in that case fine on that front.
> >
> > But you are not permitted to touch this field (we need to enforce this...)
>
> Sidenote: I think the GPU DRM subsystem relies on changing pgoff in
> some of their mmap handlers; maybe talk to them about this if you
> haven't already. See for example drm_gem_prime_mmap() and
> dma_buf_mmap().
Thanks Jann , I feel like I've opened up a can of worms with this :) I will
note these as things to prioritise in the audit.
It might be worth both auditing and then actually doing the change to
restrict what can be done here too.
The problem is it requires changing a trillion callers, but hey I'm
Mr. Churn after all... ;)
Sorry Yang - I realise this is a pain and not at all obvious. Something we
in mm need to sort out (by which I mean _me_ :) your contribution and ideas
here are very valued!
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 18:14 ` Lorenzo Stoakes
2025-01-14 18:19 ` Lorenzo Stoakes
2025-01-14 18:32 ` Jann Horn
@ 2025-01-14 19:03 ` Yang Shi
2025-01-14 19:13 ` Lorenzo Stoakes
2 siblings, 1 reply; 35+ messages in thread
From: Yang Shi @ 2025-01-14 19:03 UTC (permalink / raw)
To: Lorenzo Stoakes
Cc: arnd, gregkh, Liam.Howlett, vbabka, jannh, willy, liushixin2,
akpm, linux-mm, linux-kernel
On 1/14/25 10:14 AM, Lorenzo Stoakes wrote:
> This is getting into realms of discussion so to risk sounding rude - to be
> clear - NACK.
>
> The user-visible change to /proc/$pid/[s]maps kills this patch dead. This
> is regardless of any other discussed issue.
I admit this is a concern, but I don't think this is really that bad to
kill this patch. May this change result in userspace regression? Maybe,
likely happens to some debugging and monitoring scripts, typically we
don't worry them that much. Of course, I can't completely guarantee no
regression for real life applications, it should just be unlikely IMHO.
>
> But more importantly, I hadn't realise mmap_zero() was on the .mmap()
> callback (sorry my mistake) - you're simply not permitted to change
> vm_pgoff and vm_file fields here, the mapping logic doesn't expect it, and
> it's broken.
>
> To me the alternative would be to have a custom fault handler that hands
> back the zero page, but I"m not sure that's workable, you'd have to install
> a special mapping etc. and huge pages are weird and...
TBH, I don't think we need to make fault handler more complicated, it is
just handled as anonymous fault handler.
I understand your concern about changing those vma filed outside core
mm. An alternative is to move such change to vma.c. For example:
diff --git a/mm/vma.c b/mm/vma.c
index bb2119e5a0d0..2a7ea9901f57 100644
--- a/mm/vma.c
+++ b/mm/vma.c
@@ -2358,6 +2358,12 @@ static int __mmap_new_vma(struct mmap_state *map,
struct vm_area_struct **vmap)
else
vma_set_anonymous(vma);
+ if (vma_is_anonymous(vma) && vma->vm_file) {
+ fput(vma->vm_file);
+ vma->vm_file = NULL;
+ vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT;
+ }
+
if (error)
goto free_iter_vma;
>
> I do appreciate you raising this especially as I was blissfully unaware,
> but I don't see how this patch can possibly work, sorry :(
>
> On Tue, Jan 14, 2025 at 08:53:01AM -0800, Yang Shi wrote:
>>
>>
>> On 1/14/25 4:05 AM, Lorenzo Stoakes wrote:
>>> + Willy for the fs/weirdness elements of this.
>>>
>>> On Mon, Jan 13, 2025 at 02:30:33PM -0800, Yang Shi wrote:
>>>> When creating private mapping for /dev/zero, the driver makes it an
>>>> anonymous mapping by calling set_vma_anonymous(). But it just sets
>>>> vm_ops to NULL, vm_file is still valid and vm_pgoff is also file offset.
>>> Hm yikes.
>>>
>>>> This is a special case and the VMA doesn't look like either anonymous VMA
>>>> or file VMA. It confused other kernel subsystem, for example, khugepaged [1].
>>>>
>>>> It seems pointless to keep such special case. Making private /dev/zero
>>>> mapping a full anonymous mapping doesn't change the semantic of
>>>> /dev/zero either.
>>> My concern is that ostensibly there _is_ a file right? Are we certain that by
>>> not setting this we are not breaking something somewhere else?
>>>
>>> Are we not creating a sort of other type of 'non-such-beast' here?
>> But the file is /dev/zero. I don't see this could break the semantic of
>> /dev/zero. The shared mapping of /dev/zero is not affected by this change,
>> kernel already treated private mapping of /dev/zero as anonymous mapping,
>> but with some weird settings in VMA. When reading the mapping, it returns 0
>> with zero page, when writing the mapping, a new anonymous folio is
>> allocated.
> You're creating a new concept of an anon but not anon but also now with
> anon vm_pgoff and missing vm_file even though it does reference a file
> and... yeah.
>
> This is not usual :)
It does reference a file, but the file is /dev/zero... And if kernel
already treated it as anonymous mapping, it sounds like the file may not
matter that much, so why not make it as a real anonymous mapping? Then
we end up having anonymous VMA and file VMA only instead of anonymous
VMA, file VMA and hybrid special VMA. So we have less thing to worry
about. If VMA is anonymous VMA, it is guaranteed vm_file is NULL, vm_ops
is NULL and vm_pgoff is linear pgoff. But it is not true now.
>
>>> I mean already setting it anon and setting vm_file non-NULL is really strange.
>>>
>>>> The user visible effect is the mapping entry shown in /proc/<PID>/smaps
>>>> and /proc/<PID>/maps.
>>>>
>>>> Before the change:
>>>> ffffb7190000-ffffb7590000 rw-p 00001000 00:06 8 /dev/zero
>>>>
>>>> After the change:
>>>> ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
>>>>
>>> Yeah this seems like it might break somebody to be honest, it's really
>>> really really strange to map a file then for it not to be mapped.
>> Yes, it is possible if someone really care whether the anonymous-like
>> mapping is mapped by /dev/zero or just created by malloc(). But I don't know
>> who really do...
>>
>>> But it's possibly EVEN WEIRDER to map a file and for it to seem mapped as a
>>> file but for it to be marked anonymous.
>>>
>>> God what a mess.
>>>
>>>> [1]: https://lore.kernel.org/linux-mm/20250111034511.2223353-1-liushixin2@huawei.com/
>>> I kind of hate that we have to mitigate like this for a case that should
>>> never ever happen so I'm inclined towards your solution but a lot more
>>> inclined towards us totally rethinking this.
>>>
>>> Do we _have_ to make this anonymous?? Why can't we just reference the zero
>>> page as if it were in the page cache (Willy - feel free to correct naive
>>> misapprehension here).
>> TBH, I don't see why page cache has to be involved. When reading, 0 is
>> returned by zero page. When writing a CoW is triggered if page cache is
>> involved, but the content of the page cache should be just 0, so we copy 0
>> to the new folio then write to it. It doesn't make too much sense. I think
>> this is why private /dev/zero mapping is treated as anonymous mapping in the
>> first place.
> I'm obviously not suggesting allocating a bunch of extra folios, I was
> thinking there would be some means of handing back the actual zero
> page. But I am not sure this is workable.
As I mentioned above, even handing back zero page should be not needed.
>
>>>> Signed-off-by: Yang Shi <yang@os.amperecomputing.com>
>>>> ---
>>>> drivers/char/mem.c | 4 ++++
>>>> 1 file changed, 4 insertions(+)
>>>>
>>>> diff --git a/drivers/char/mem.c b/drivers/char/mem.c
>>>> index 169eed162a7f..dae113f7fc1b 100644
>>>> --- a/drivers/char/mem.c
>>>> +++ b/drivers/char/mem.c
>>>> @@ -527,6 +527,10 @@ static int mmap_zero(struct file *file, struct vm_area_struct *vma)
>>>> if (vma->vm_flags & VM_SHARED)
>>>> return shmem_zero_setup(vma);
>>>> vma_set_anonymous(vma);
>>>> + fput(vma->vm_file);
>>>> + vma->vm_file = NULL;
>>>> + vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT;
> This is just not permitted. We maintain mmap state which contains the file
> and pgoff state which gets threaded through the mapping operation, and
> simply do not expect you to change these fields.
>
> In future we will assert on this or preferably, restrict users to only
> changing VMA flags, the private field and vm_ops.
Sure, hardening the VMA initialization code and making less surprising
corner case is definitely helpful.
>
>>> Hmm, this might have been mremap()'d _potentially_ though? And then now
>>> this will be wrong? But then we'd have no way of tracking it correctly...
>> I'm not quite familiar with the subtle details and corner cases of
>> meremap(). But mmap_zero() should be called by mmap(), so the VMA has not
>> been visible to user yet at this point IIUC. How come mremap() could move
>> it?
> Ah OK, in that case fine on that front.
>
> But you are not permitted to touch this field (we need to enforce this...)
>
>>> I've not checked the function but do we mark this as a special mapping of
>>> some kind?
>>>
>>>> +
>>>> return 0;
>>>> }
>>>>
>>>> --
>>>> 2.47.0
>>>>
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 19:03 ` Yang Shi
@ 2025-01-14 19:13 ` Lorenzo Stoakes
2025-01-14 21:24 ` Yang Shi
0 siblings, 1 reply; 35+ messages in thread
From: Lorenzo Stoakes @ 2025-01-14 19:13 UTC (permalink / raw)
To: Yang Shi
Cc: arnd, gregkh, Liam.Howlett, vbabka, jannh, willy, liushixin2,
akpm, linux-mm, linux-kernel
On Tue, Jan 14, 2025 at 11:03:48AM -0800, Yang Shi wrote:
>
>
>
> On 1/14/25 10:14 AM, Lorenzo Stoakes wrote:
> > This is getting into realms of discussion so to risk sounding rude - to be
> > clear - NACK.
> >
> > The user-visible change to /proc/$pid/[s]maps kills this patch dead. This
> > is regardless of any other discussed issue.
>
> I admit this is a concern, but I don't think this is really that bad to kill
> this patch. May this change result in userspace regression? Maybe, likely
> happens to some debugging and monitoring scripts, typically we don't worry
> them that much. Of course, I can't completely guarantee no regression for
> real life applications, it should just be unlikely IMHO.
Yeah, I don't think we can accept this unfortunately.
This patch is SUPER important though even if rejected, because you've made
me realise we really need to audit all of these mmap handlers... so it's
all super appreciated regardless :)
>
> >
> > But more importantly, I hadn't realise mmap_zero() was on the .mmap()
> > callback (sorry my mistake) - you're simply not permitted to change
> > vm_pgoff and vm_file fields here, the mapping logic doesn't expect it, and
> > it's broken.
> >
> > To me the alternative would be to have a custom fault handler that hands
> > back the zero page, but I"m not sure that's workable, you'd have to install
> > a special mapping etc. and huge pages are weird and...
>
> TBH, I don't think we need to make fault handler more complicated, it is
> just handled as anonymous fault handler.
>
> I understand your concern about changing those vma filed outside core mm. An
> alternative is to move such change to vma.c. For example:
>
> diff --git a/mm/vma.c b/mm/vma.c
> index bb2119e5a0d0..2a7ea9901f57 100644
> --- a/mm/vma.c
> +++ b/mm/vma.c
> @@ -2358,6 +2358,12 @@ static int __mmap_new_vma(struct mmap_state *map,
> struct vm_area_struct **vmap)
> else
> vma_set_anonymous(vma);
>
> + if (vma_is_anonymous(vma) && vma->vm_file) {
> + fput(vma->vm_file);
> + vma->vm_file = NULL;
> + vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT;
> + }
> +
OK that's more interesting. Though the user-facing thing remains...
It's possiible we could detect that the underlying thing is a zero page and
manually print out /dev/zero, but can somebody create a zero page file
elsewhere? In which case they might find this confusing.
It's actually a nice idea to have this _explicitly_ covered off as we could
then also add a comment explaining 'hey there's this weird type of VMA' and
have it in a place where it's actually obvious to mm folk anyway.
But this maps thing is just a killer. Somebody somewhere will be
confused. And it is not for us to judge whether that's silly or not...
> if (error)
> goto free_iter_vma;
>
>
> >
> > I do appreciate you raising this especially as I was blissfully unaware,
> > but I don't see how this patch can possibly work, sorry :(
> >
> > On Tue, Jan 14, 2025 at 08:53:01AM -0800, Yang Shi wrote:
> > >
> > >
> > > On 1/14/25 4:05 AM, Lorenzo Stoakes wrote:
> > > > + Willy for the fs/weirdness elements of this.
> > > >
> > > > On Mon, Jan 13, 2025 at 02:30:33PM -0800, Yang Shi wrote:
> > > > > When creating private mapping for /dev/zero, the driver makes it an
> > > > > anonymous mapping by calling set_vma_anonymous(). But it just sets
> > > > > vm_ops to NULL, vm_file is still valid and vm_pgoff is also file offset.
> > > > Hm yikes.
> > > >
> > > > > This is a special case and the VMA doesn't look like either anonymous VMA
> > > > > or file VMA. It confused other kernel subsystem, for example, khugepaged [1].
> > > > >
> > > > > It seems pointless to keep such special case. Making private /dev/zero
> > > > > mapping a full anonymous mapping doesn't change the semantic of
> > > > > /dev/zero either.
> > > > My concern is that ostensibly there _is_ a file right? Are we certain that by
> > > > not setting this we are not breaking something somewhere else?
> > > >
> > > > Are we not creating a sort of other type of 'non-such-beast' here?
> > > But the file is /dev/zero. I don't see this could break the semantic of
> > > /dev/zero. The shared mapping of /dev/zero is not affected by this change,
> > > kernel already treated private mapping of /dev/zero as anonymous mapping,
> > > but with some weird settings in VMA. When reading the mapping, it returns 0
> > > with zero page, when writing the mapping, a new anonymous folio is
> > > allocated.
> > You're creating a new concept of an anon but not anon but also now with
> > anon vm_pgoff and missing vm_file even though it does reference a file
> > and... yeah.
> >
> > This is not usual :)
>
> It does reference a file, but the file is /dev/zero... And if kernel already
> treated it as anonymous mapping, it sounds like the file may not matter that
> much, so why not make it as a real anonymous mapping? Then we end up having
> anonymous VMA and file VMA only instead of anonymous VMA, file VMA and
> hybrid special VMA. So we have less thing to worry about. If VMA is
> anonymous VMA, it is guaranteed vm_file is NULL, vm_ops is NULL and vm_pgoff
> is linear pgoff. But it is not true now.
It's about user confusion for me really.
>
> >
> > > > I mean already setting it anon and setting vm_file non-NULL is really strange.
> > > >
> > > > > The user visible effect is the mapping entry shown in /proc/<PID>/smaps
> > > > > and /proc/<PID>/maps.
> > > > >
> > > > > Before the change:
> > > > > ffffb7190000-ffffb7590000 rw-p 00001000 00:06 8 /dev/zero
> > > > >
> > > > > After the change:
> > > > > ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
> > > > >
> > > > Yeah this seems like it might break somebody to be honest, it's really
> > > > really really strange to map a file then for it not to be mapped.
> > > Yes, it is possible if someone really care whether the anonymous-like
> > > mapping is mapped by /dev/zero or just created by malloc(). But I don't know
> > > who really do...
> > >
> > > > But it's possibly EVEN WEIRDER to map a file and for it to seem mapped as a
> > > > file but for it to be marked anonymous.
> > > >
> > > > God what a mess.
> > > >
> > > > > [1]: https://lore.kernel.org/linux-mm/20250111034511.2223353-1-liushixin2@huawei.com/
> > > > I kind of hate that we have to mitigate like this for a case that should
> > > > never ever happen so I'm inclined towards your solution but a lot more
> > > > inclined towards us totally rethinking this.
> > > >
> > > > Do we _have_ to make this anonymous?? Why can't we just reference the zero
> > > > page as if it were in the page cache (Willy - feel free to correct naive
> > > > misapprehension here).
> > > TBH, I don't see why page cache has to be involved. When reading, 0 is
> > > returned by zero page. When writing a CoW is triggered if page cache is
> > > involved, but the content of the page cache should be just 0, so we copy 0
> > > to the new folio then write to it. It doesn't make too much sense. I think
> > > this is why private /dev/zero mapping is treated as anonymous mapping in the
> > > first place.
> > I'm obviously not suggesting allocating a bunch of extra folios, I was
> > thinking there would be some means of handing back the actual zero
> > page. But I am not sure this is workable.
>
> As I mentioned above, even handing back zero page should be not needed.
Ack.
>
> >
> > > > > Signed-off-by: Yang Shi <yang@os.amperecomputing.com>
> > > > > ---
> > > > > drivers/char/mem.c | 4 ++++
> > > > > 1 file changed, 4 insertions(+)
> > > > >
> > > > > diff --git a/drivers/char/mem.c b/drivers/char/mem.c
> > > > > index 169eed162a7f..dae113f7fc1b 100644
> > > > > --- a/drivers/char/mem.c
> > > > > +++ b/drivers/char/mem.c
> > > > > @@ -527,6 +527,10 @@ static int mmap_zero(struct file *file, struct vm_area_struct *vma)
> > > > > if (vma->vm_flags & VM_SHARED)
> > > > > return shmem_zero_setup(vma);
> > > > > vma_set_anonymous(vma);
> > > > > + fput(vma->vm_file);
> > > > > + vma->vm_file = NULL;
> > > > > + vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT;
> > This is just not permitted. We maintain mmap state which contains the file
> > and pgoff state which gets threaded through the mapping operation, and
> > simply do not expect you to change these fields.
> >
> > In future we will assert on this or preferably, restrict users to only
> > changing VMA flags, the private field and vm_ops.
>
> Sure, hardening the VMA initialization code and making less surprising
> corner case is definitely helpful.
Yes and I've opened a can of worms and the worms have jumped out and on to
my face and were not worms but in fact an alien facehugger :P
In other words, I am going to be looking into this very seriously and
auditing this whole thing... yay for making work for myself... :>)
>
> >
> > > > Hmm, this might have been mremap()'d _potentially_ though? And then now
> > > > this will be wrong? But then we'd have no way of tracking it correctly...
> > > I'm not quite familiar with the subtle details and corner cases of
> > > meremap(). But mmap_zero() should be called by mmap(), so the VMA has not
> > > been visible to user yet at this point IIUC. How come mremap() could move
> > > it?
> > Ah OK, in that case fine on that front.
> >
> > But you are not permitted to touch this field (we need to enforce this...)
> >
> > > > I've not checked the function but do we mark this as a special mapping of
> > > > some kind?
> > > >
> > > > > +
> > > > > return 0;
> > > > > }
> > > > >
> > > > > --
> > > > > 2.47.0
> > > > >
>
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 19:13 ` Lorenzo Stoakes
@ 2025-01-14 21:24 ` Yang Shi
2025-01-15 12:10 ` Lorenzo Stoakes
0 siblings, 1 reply; 35+ messages in thread
From: Yang Shi @ 2025-01-14 21:24 UTC (permalink / raw)
To: Lorenzo Stoakes
Cc: arnd, gregkh, Liam.Howlett, vbabka, jannh, willy, liushixin2,
akpm, linux-mm, linux-kernel
On 1/14/25 11:13 AM, Lorenzo Stoakes wrote:
> On Tue, Jan 14, 2025 at 11:03:48AM -0800, Yang Shi wrote:
>>
>>
>> On 1/14/25 10:14 AM, Lorenzo Stoakes wrote:
>>> This is getting into realms of discussion so to risk sounding rude - to be
>>> clear - NACK.
>>>
>>> The user-visible change to /proc/$pid/[s]maps kills this patch dead. This
>>> is regardless of any other discussed issue.
>> I admit this is a concern, but I don't think this is really that bad to kill
>> this patch. May this change result in userspace regression? Maybe, likely
>> happens to some debugging and monitoring scripts, typically we don't worry
>> them that much. Of course, I can't completely guarantee no regression for
>> real life applications, it should just be unlikely IMHO.
> Yeah, I don't think we can accept this unfortunately.
>
> This patch is SUPER important though even if rejected, because you've made
> me realise we really need to audit all of these mmap handlers... so it's
> all super appreciated regardless :)
:-)
>
>>> But more importantly, I hadn't realise mmap_zero() was on the .mmap()
>>> callback (sorry my mistake) - you're simply not permitted to change
>>> vm_pgoff and vm_file fields here, the mapping logic doesn't expect it, and
>>> it's broken.
>>>
>>> To me the alternative would be to have a custom fault handler that hands
>>> back the zero page, but I"m not sure that's workable, you'd have to install
>>> a special mapping etc. and huge pages are weird and...
>> TBH, I don't think we need to make fault handler more complicated, it is
>> just handled as anonymous fault handler.
>>
>> I understand your concern about changing those vma filed outside core mm. An
>> alternative is to move such change to vma.c. For example:
>>
>> diff --git a/mm/vma.c b/mm/vma.c
>> index bb2119e5a0d0..2a7ea9901f57 100644
>> --- a/mm/vma.c
>> +++ b/mm/vma.c
>> @@ -2358,6 +2358,12 @@ static int __mmap_new_vma(struct mmap_state *map,
>> struct vm_area_struct **vmap)
>> else
>> vma_set_anonymous(vma);
>>
>> + if (vma_is_anonymous(vma) && vma->vm_file) {
>> + fput(vma->vm_file);
>> + vma->vm_file = NULL;
>> + vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT;
>> + }
>> +
> OK that's more interesting. Though the user-facing thing remains...
>
> It's possiible we could detect that the underlying thing is a zero page and
> manually print out /dev/zero, but can somebody create a zero page file
> elsewhere? In which case they might find this confusing.
I'm not sure about file mapping. However reading an anonymous mapping
will instantiate zero page. It should not be marked as /dev/zero mapping.
>
> It's actually a nice idea to have this _explicitly_ covered off as we could
> then also add a comment explaining 'hey there's this weird type of VMA' and
> have it in a place where it's actually obvious to mm folk anyway.
>
> But this maps thing is just a killer. Somebody somewhere will be
> confused. And it is not for us to judge whether that's silly or not...
I just thought of named anonymous VMA may help. We can give the private
/dev/zero mapping a name, for example, just "/dev/zero". However,
"[anon:/dev/zero]" will show up in smaps/maps. We can't keep the device
numbers and inode number either, but it seems it can tell the user this
mapping comes from /dev/zero, and it also explicitly tells us it is
specially treated by kernel. Hopefully setting anon_name is permitted.
>
>> if (error)
>> goto free_iter_vma;
>>
>>
>>> I do appreciate you raising this especially as I was blissfully unaware,
>>> but I don't see how this patch can possibly work, sorry :(
>>>
>>> On Tue, Jan 14, 2025 at 08:53:01AM -0800, Yang Shi wrote:
>>>>
>>>> On 1/14/25 4:05 AM, Lorenzo Stoakes wrote:
>>>>> + Willy for the fs/weirdness elements of this.
>>>>>
>>>>> On Mon, Jan 13, 2025 at 02:30:33PM -0800, Yang Shi wrote:
>>>>>> When creating private mapping for /dev/zero, the driver makes it an
>>>>>> anonymous mapping by calling set_vma_anonymous(). But it just sets
>>>>>> vm_ops to NULL, vm_file is still valid and vm_pgoff is also file offset.
>>>>> Hm yikes.
>>>>>
>>>>>> This is a special case and the VMA doesn't look like either anonymous VMA
>>>>>> or file VMA. It confused other kernel subsystem, for example, khugepaged [1].
>>>>>>
>>>>>> It seems pointless to keep such special case. Making private /dev/zero
>>>>>> mapping a full anonymous mapping doesn't change the semantic of
>>>>>> /dev/zero either.
>>>>> My concern is that ostensibly there _is_ a file right? Are we certain that by
>>>>> not setting this we are not breaking something somewhere else?
>>>>>
>>>>> Are we not creating a sort of other type of 'non-such-beast' here?
>>>> But the file is /dev/zero. I don't see this could break the semantic of
>>>> /dev/zero. The shared mapping of /dev/zero is not affected by this change,
>>>> kernel already treated private mapping of /dev/zero as anonymous mapping,
>>>> but with some weird settings in VMA. When reading the mapping, it returns 0
>>>> with zero page, when writing the mapping, a new anonymous folio is
>>>> allocated.
>>> You're creating a new concept of an anon but not anon but also now with
>>> anon vm_pgoff and missing vm_file even though it does reference a file
>>> and... yeah.
>>>
>>> This is not usual :)
>> It does reference a file, but the file is /dev/zero... And if kernel already
>> treated it as anonymous mapping, it sounds like the file may not matter that
>> much, so why not make it as a real anonymous mapping? Then we end up having
>> anonymous VMA and file VMA only instead of anonymous VMA, file VMA and
>> hybrid special VMA. So we have less thing to worry about. If VMA is
>> anonymous VMA, it is guaranteed vm_file is NULL, vm_ops is NULL and vm_pgoff
>> is linear pgoff. But it is not true now.
> It's about user confusion for me really.
>
>>>>> I mean already setting it anon and setting vm_file non-NULL is really strange.
>>>>>
>>>>>> The user visible effect is the mapping entry shown in /proc/<PID>/smaps
>>>>>> and /proc/<PID>/maps.
>>>>>>
>>>>>> Before the change:
>>>>>> ffffb7190000-ffffb7590000 rw-p 00001000 00:06 8 /dev/zero
>>>>>>
>>>>>> After the change:
>>>>>> ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
>>>>>>
>>>>> Yeah this seems like it might break somebody to be honest, it's really
>>>>> really really strange to map a file then for it not to be mapped.
>>>> Yes, it is possible if someone really care whether the anonymous-like
>>>> mapping is mapped by /dev/zero or just created by malloc(). But I don't know
>>>> who really do...
>>>>
>>>>> But it's possibly EVEN WEIRDER to map a file and for it to seem mapped as a
>>>>> file but for it to be marked anonymous.
>>>>>
>>>>> God what a mess.
>>>>>
>>>>>> [1]: https://lore.kernel.org/linux-mm/20250111034511.2223353-1-liushixin2@huawei.com/
>>>>> I kind of hate that we have to mitigate like this for a case that should
>>>>> never ever happen so I'm inclined towards your solution but a lot more
>>>>> inclined towards us totally rethinking this.
>>>>>
>>>>> Do we _have_ to make this anonymous?? Why can't we just reference the zero
>>>>> page as if it were in the page cache (Willy - feel free to correct naive
>>>>> misapprehension here).
>>>> TBH, I don't see why page cache has to be involved. When reading, 0 is
>>>> returned by zero page. When writing a CoW is triggered if page cache is
>>>> involved, but the content of the page cache should be just 0, so we copy 0
>>>> to the new folio then write to it. It doesn't make too much sense. I think
>>>> this is why private /dev/zero mapping is treated as anonymous mapping in the
>>>> first place.
>>> I'm obviously not suggesting allocating a bunch of extra folios, I was
>>> thinking there would be some means of handing back the actual zero
>>> page. But I am not sure this is workable.
>> As I mentioned above, even handing back zero page should be not needed.
> Ack.
>
>>>>>> Signed-off-by: Yang Shi <yang@os.amperecomputing.com>
>>>>>> ---
>>>>>> drivers/char/mem.c | 4 ++++
>>>>>> 1 file changed, 4 insertions(+)
>>>>>>
>>>>>> diff --git a/drivers/char/mem.c b/drivers/char/mem.c
>>>>>> index 169eed162a7f..dae113f7fc1b 100644
>>>>>> --- a/drivers/char/mem.c
>>>>>> +++ b/drivers/char/mem.c
>>>>>> @@ -527,6 +527,10 @@ static int mmap_zero(struct file *file, struct vm_area_struct *vma)
>>>>>> if (vma->vm_flags & VM_SHARED)
>>>>>> return shmem_zero_setup(vma);
>>>>>> vma_set_anonymous(vma);
>>>>>> + fput(vma->vm_file);
>>>>>> + vma->vm_file = NULL;
>>>>>> + vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT;
>>> This is just not permitted. We maintain mmap state which contains the file
>>> and pgoff state which gets threaded through the mapping operation, and
>>> simply do not expect you to change these fields.
>>>
>>> In future we will assert on this or preferably, restrict users to only
>>> changing VMA flags, the private field and vm_ops.
>> Sure, hardening the VMA initialization code and making less surprising
>> corner case is definitely helpful.
> Yes and I've opened a can of worms and the worms have jumped out and on to
> my face and were not worms but in fact an alien facehugger :P
>
> In other words, I am going to be looking into this very seriously and
> auditing this whole thing... yay for making work for myself... :>)
Thank you for taking the action to kill the alien facehugger :-)
>
>>>>> Hmm, this might have been mremap()'d _potentially_ though? And then now
>>>>> this will be wrong? But then we'd have no way of tracking it correctly...
>>>> I'm not quite familiar with the subtle details and corner cases of
>>>> meremap(). But mmap_zero() should be called by mmap(), so the VMA has not
>>>> been visible to user yet at this point IIUC. How come mremap() could move
>>>> it?
>>> Ah OK, in that case fine on that front.
>>>
>>> But you are not permitted to touch this field (we need to enforce this...)
>>>
>>>>> I've not checked the function but do we mark this as a special mapping of
>>>>> some kind?
>>>>>
>>>>>> +
>>>>>> return 0;
>>>>>> }
>>>>>>
>>>>>> --
>>>>>> 2.47.0
>>>>>>
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-14 21:24 ` Yang Shi
@ 2025-01-15 12:10 ` Lorenzo Stoakes
2025-01-15 21:29 ` Yang Shi
0 siblings, 1 reply; 35+ messages in thread
From: Lorenzo Stoakes @ 2025-01-15 12:10 UTC (permalink / raw)
To: Yang Shi
Cc: arnd, gregkh, Liam.Howlett, vbabka, jannh, willy, liushixin2,
akpm, linux-mm, linux-kernel
On Tue, Jan 14, 2025 at 01:24:25PM -0800, Yang Shi wrote:
>
>
>
> On 1/14/25 11:13 AM, Lorenzo Stoakes wrote:
> > On Tue, Jan 14, 2025 at 11:03:48AM -0800, Yang Shi wrote:
> > >
> > >
> > > On 1/14/25 10:14 AM, Lorenzo Stoakes wrote:
> > > > This is getting into realms of discussion so to risk sounding rude - to be
> > > > clear - NACK.
> > > >
> > > > The user-visible change to /proc/$pid/[s]maps kills this patch dead. This
> > > > is regardless of any other discussed issue.
> > > I admit this is a concern, but I don't think this is really that bad to kill
> > > this patch. May this change result in userspace regression? Maybe, likely
> > > happens to some debugging and monitoring scripts, typically we don't worry
> > > them that much. Of course, I can't completely guarantee no regression for
> > > real life applications, it should just be unlikely IMHO.
> > Yeah, I don't think we can accept this unfortunately.
> >
> > This patch is SUPER important though even if rejected, because you've made
> > me realise we really need to audit all of these mmap handlers... so it's
> > all super appreciated regardless :)
>
> :-)
>
> >
> > > > But more importantly, I hadn't realise mmap_zero() was on the .mmap()
> > > > callback (sorry my mistake) - you're simply not permitted to change
> > > > vm_pgoff and vm_file fields here, the mapping logic doesn't expect it, and
> > > > it's broken.
> > > >
> > > > To me the alternative would be to have a custom fault handler that hands
> > > > back the zero page, but I"m not sure that's workable, you'd have to install
> > > > a special mapping etc. and huge pages are weird and...
> > > TBH, I don't think we need to make fault handler more complicated, it is
> > > just handled as anonymous fault handler.
> > >
> > > I understand your concern about changing those vma filed outside core mm. An
> > > alternative is to move such change to vma.c. For example:
> > >
> > > diff --git a/mm/vma.c b/mm/vma.c
> > > index bb2119e5a0d0..2a7ea9901f57 100644
> > > --- a/mm/vma.c
> > > +++ b/mm/vma.c
> > > @@ -2358,6 +2358,12 @@ static int __mmap_new_vma(struct mmap_state *map,
> > > struct vm_area_struct **vmap)
> > > else
> > > vma_set_anonymous(vma);
> > >
> > > + if (vma_is_anonymous(vma) && vma->vm_file) {
> > > + fput(vma->vm_file);
> > > + vma->vm_file = NULL;
> > > + vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT;
> > > + }
> > > +
> > OK that's more interesting. Though the user-facing thing remains...
> >
> > It's possiible we could detect that the underlying thing is a zero page and
> > manually print out /dev/zero, but can somebody create a zero page file
> > elsewhere? In which case they might find this confusing.
>
> I'm not sure about file mapping. However reading an anonymous mapping will
> instantiate zero page. It should not be marked as /dev/zero mapping.
>
> >
> > It's actually a nice idea to have this _explicitly_ covered off as we could
> > then also add a comment explaining 'hey there's this weird type of VMA' and
> > have it in a place where it's actually obvious to mm folk anyway.
> >
> > But this maps thing is just a killer. Somebody somewhere will be
> > confused. And it is not for us to judge whether that's silly or not...
>
> I just thought of named anonymous VMA may help. We can give the private
> /dev/zero mapping a name, for example, just "/dev/zero". However,
> "[anon:/dev/zero]" will show up in smaps/maps. We can't keep the device
> numbers and inode number either, but it seems it can tell the user this
> mapping comes from /dev/zero, and it also explicitly tells us it is
> specially treated by kernel. Hopefully setting anon_name is permitted.
But then that'd require CONFIG_ANON_VMA_NAME unfortunately :(
I think this maps thing is the killer here really.
It'd be nice to -specifically- have a means of expressing this kind of VMA,
we have a means of setting a VMA anon, so maybe we can 'set a VMA to
/dev/zero' and somehow explicitly know that we've done this and identify
this special case.
I'm not sure that the .mmap callback is the right place to do this and I"m
not sure how exactly this would work but this could be workable.
I agree the actual offset into the zero page is of no relevance and no
_sane_ user will care, but this way we could put /dev/zero in [s]maps,
treat this VMA as anon, but also add semantic information about the
existence of this weird corner case.
>
> >
> > > if (error)
> > > goto free_iter_vma;
> > >
> > >
> > > > I do appreciate you raising this especially as I was blissfully unaware,
> > > > but I don't see how this patch can possibly work, sorry :(
> > > >
> > > > On Tue, Jan 14, 2025 at 08:53:01AM -0800, Yang Shi wrote:
> > > > >
> > > > > On 1/14/25 4:05 AM, Lorenzo Stoakes wrote:
> > > > > > + Willy for the fs/weirdness elements of this.
> > > > > >
> > > > > > On Mon, Jan 13, 2025 at 02:30:33PM -0800, Yang Shi wrote:
> > > > > > > When creating private mapping for /dev/zero, the driver makes it an
> > > > > > > anonymous mapping by calling set_vma_anonymous(). But it just sets
> > > > > > > vm_ops to NULL, vm_file is still valid and vm_pgoff is also file offset.
> > > > > > Hm yikes.
> > > > > >
> > > > > > > This is a special case and the VMA doesn't look like either anonymous VMA
> > > > > > > or file VMA. It confused other kernel subsystem, for example, khugepaged [1].
> > > > > > >
> > > > > > > It seems pointless to keep such special case. Making private /dev/zero
> > > > > > > mapping a full anonymous mapping doesn't change the semantic of
> > > > > > > /dev/zero either.
> > > > > > My concern is that ostensibly there _is_ a file right? Are we certain that by
> > > > > > not setting this we are not breaking something somewhere else?
> > > > > >
> > > > > > Are we not creating a sort of other type of 'non-such-beast' here?
> > > > > But the file is /dev/zero. I don't see this could break the semantic of
> > > > > /dev/zero. The shared mapping of /dev/zero is not affected by this change,
> > > > > kernel already treated private mapping of /dev/zero as anonymous mapping,
> > > > > but with some weird settings in VMA. When reading the mapping, it returns 0
> > > > > with zero page, when writing the mapping, a new anonymous folio is
> > > > > allocated.
> > > > You're creating a new concept of an anon but not anon but also now with
> > > > anon vm_pgoff and missing vm_file even though it does reference a file
> > > > and... yeah.
> > > >
> > > > This is not usual :)
> > > It does reference a file, but the file is /dev/zero... And if kernel already
> > > treated it as anonymous mapping, it sounds like the file may not matter that
> > > much, so why not make it as a real anonymous mapping? Then we end up having
> > > anonymous VMA and file VMA only instead of anonymous VMA, file VMA and
> > > hybrid special VMA. So we have less thing to worry about. If VMA is
> > > anonymous VMA, it is guaranteed vm_file is NULL, vm_ops is NULL and vm_pgoff
> > > is linear pgoff. But it is not true now.
> > It's about user confusion for me really.
> >
> > > > > > I mean already setting it anon and setting vm_file non-NULL is really strange.
> > > > > >
> > > > > > > The user visible effect is the mapping entry shown in /proc/<PID>/smaps
> > > > > > > and /proc/<PID>/maps.
> > > > > > >
> > > > > > > Before the change:
> > > > > > > ffffb7190000-ffffb7590000 rw-p 00001000 00:06 8 /dev/zero
> > > > > > >
> > > > > > > After the change:
> > > > > > > ffffb6130000-ffffb6530000 rw-p 00000000 00:00 0
> > > > > > >
> > > > > > Yeah this seems like it might break somebody to be honest, it's really
> > > > > > really really strange to map a file then for it not to be mapped.
> > > > > Yes, it is possible if someone really care whether the anonymous-like
> > > > > mapping is mapped by /dev/zero or just created by malloc(). But I don't know
> > > > > who really do...
> > > > >
> > > > > > But it's possibly EVEN WEIRDER to map a file and for it to seem mapped as a
> > > > > > file but for it to be marked anonymous.
> > > > > >
> > > > > > God what a mess.
> > > > > >
> > > > > > > [1]: https://lore.kernel.org/linux-mm/20250111034511.2223353-1-liushixin2@huawei.com/
> > > > > > I kind of hate that we have to mitigate like this for a case that should
> > > > > > never ever happen so I'm inclined towards your solution but a lot more
> > > > > > inclined towards us totally rethinking this.
> > > > > >
> > > > > > Do we _have_ to make this anonymous?? Why can't we just reference the zero
> > > > > > page as if it were in the page cache (Willy - feel free to correct naive
> > > > > > misapprehension here).
> > > > > TBH, I don't see why page cache has to be involved. When reading, 0 is
> > > > > returned by zero page. When writing a CoW is triggered if page cache is
> > > > > involved, but the content of the page cache should be just 0, so we copy 0
> > > > > to the new folio then write to it. It doesn't make too much sense. I think
> > > > > this is why private /dev/zero mapping is treated as anonymous mapping in the
> > > > > first place.
> > > > I'm obviously not suggesting allocating a bunch of extra folios, I was
> > > > thinking there would be some means of handing back the actual zero
> > > > page. But I am not sure this is workable.
> > > As I mentioned above, even handing back zero page should be not needed.
> > Ack.
> >
> > > > > > > Signed-off-by: Yang Shi <yang@os.amperecomputing.com>
> > > > > > > ---
> > > > > > > drivers/char/mem.c | 4 ++++
> > > > > > > 1 file changed, 4 insertions(+)
> > > > > > >
> > > > > > > diff --git a/drivers/char/mem.c b/drivers/char/mem.c
> > > > > > > index 169eed162a7f..dae113f7fc1b 100644
> > > > > > > --- a/drivers/char/mem.c
> > > > > > > +++ b/drivers/char/mem.c
> > > > > > > @@ -527,6 +527,10 @@ static int mmap_zero(struct file *file, struct vm_area_struct *vma)
> > > > > > > if (vma->vm_flags & VM_SHARED)
> > > > > > > return shmem_zero_setup(vma);
> > > > > > > vma_set_anonymous(vma);
> > > > > > > + fput(vma->vm_file);
> > > > > > > + vma->vm_file = NULL;
> > > > > > > + vma->vm_pgoff = vma->vm_start >> PAGE_SHIFT;
> > > > This is just not permitted. We maintain mmap state which contains the file
> > > > and pgoff state which gets threaded through the mapping operation, and
> > > > simply do not expect you to change these fields.
> > > >
> > > > In future we will assert on this or preferably, restrict users to only
> > > > changing VMA flags, the private field and vm_ops.
> > > Sure, hardening the VMA initialization code and making less surprising
> > > corner case is definitely helpful.
> > Yes and I've opened a can of worms and the worms have jumped out and on to
> > my face and were not worms but in fact an alien facehugger :P
> >
> > In other words, I am going to be looking into this very seriously and
> > auditing this whole thing... yay for making work for myself... :>)
>
> Thank you for taking the action to kill the alien facehugger :-)
Haha thanks I'll do my best :))
>
> >
> > > > > > Hmm, this might have been mremap()'d _potentially_ though? And then now
> > > > > > this will be wrong? But then we'd have no way of tracking it correctly...
> > > > > I'm not quite familiar with the subtle details and corner cases of
> > > > > meremap(). But mmap_zero() should be called by mmap(), so the VMA has not
> > > > > been visible to user yet at this point IIUC. How come mremap() could move
> > > > > it?
> > > > Ah OK, in that case fine on that front.
> > > >
> > > > But you are not permitted to touch this field (we need to enforce this...)
> > > >
> > > > > > I've not checked the function but do we mark this as a special mapping of
> > > > > > some kind?
> > > > > >
> > > > > > > +
> > > > > > > return 0;
> > > > > > > }
> > > > > > >
> > > > > > > --
> > > > > > > 2.47.0
> > > > > > >
>
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-15 12:10 ` Lorenzo Stoakes
@ 2025-01-15 21:29 ` Yang Shi
2025-01-15 22:05 ` Christoph Lameter (Ampere)
0 siblings, 1 reply; 35+ messages in thread
From: Yang Shi @ 2025-01-15 21:29 UTC (permalink / raw)
To: Lorenzo Stoakes
Cc: arnd, gregkh, Liam.Howlett, vbabka, jannh, willy, liushixin2,
akpm, linux-mm, linux-kernel
>> I just thought of named anonymous VMA may help. We can give the private
>> /dev/zero mapping a name, for example, just "/dev/zero". However,
>> "[anon:/dev/zero]" will show up in smaps/maps. We can't keep the device
>> numbers and inode number either, but it seems it can tell the user this
>> mapping comes from /dev/zero, and it also explicitly tells us it is
>> specially treated by kernel. Hopefully setting anon_name is permitted.
> But then that'd require CONFIG_ANON_VMA_NAME unfortunately :(
Yes.
>
> I think this maps thing is the killer here really.
>
> It'd be nice to -specifically- have a means of expressing this kind of VMA,
> we have a means of setting a VMA anon, so maybe we can 'set a VMA to
> /dev/zero' and somehow explicitly know that we've done this and identify
> this special case.
>
> I'm not sure that the .mmap callback is the right place to do this and I"m
> not sure how exactly this would work but this could be workable.
A couple of potential approaches off the top of my head:
- A new vm flag
- Use vm_private_data
Both of them have pros and cons. The vm flag is simple enough, but it
needs to consume one bit for just one usecase. The vm_private_data is a
void pointer and a lot drivers use it to store driver specific data
structures, so using the pointer in a generic path (for example, smaps)
to tell us whether it is /dev/zero is not easy. We may be able to have a
special encoding to it, for example, set the last bit (the trick is not
unusual in core mm code).
>
> I agree the actual offset into the zero page is of no relevance and no
> _sane_ user will care, but this way we could put /dev/zero in [s]maps,
> treat this VMA as anon, but also add semantic information about the
> existence of this weird corner case.
>
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-15 21:29 ` Yang Shi
@ 2025-01-15 22:05 ` Christoph Lameter (Ampere)
0 siblings, 0 replies; 35+ messages in thread
From: Christoph Lameter (Ampere) @ 2025-01-15 22:05 UTC (permalink / raw)
To: Yang Shi
Cc: Lorenzo Stoakes, arnd, gregkh, Liam.Howlett, vbabka, jannh,
willy, liushixin2, akpm, linux-mm, linux-kernel
On Wed, 15 Jan 2025, Yang Shi wrote:
>
> > > I just thought of named anonymous VMA may help. We can give the private
> > > /dev/zero mapping a name, for example, just "/dev/zero". However,
> > > "[anon:/dev/zero]" will show up in smaps/maps. We can't keep the device
> > > numbers and inode number either, but it seems it can tell the user this
> > > mapping comes from /dev/zero, and it also explicitly tells us it is
> > > specially treated by kernel. Hopefully setting anon_name is permitted.
> > But then that'd require CONFIG_ANON_VMA_NAME unfortunately :(
>
> Yes.
Add a counter for NULL pages in smaps?
I.e.
Null: 4 kB
Both anonymous and file mappings could have NULL page references right?
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-13 22:30 [PATCH] /dev/zero: make private mapping full anonymous mapping Yang Shi
2025-01-14 12:05 ` Lorenzo Stoakes
2025-01-14 13:01 ` David Hildenbrand
@ 2025-01-28 3:14 ` kernel test robot
2025-01-31 18:38 ` Yang Shi
2 siblings, 1 reply; 35+ messages in thread
From: kernel test robot @ 2025-01-28 3:14 UTC (permalink / raw)
To: Yang Shi
Cc: oe-lkp, lkp, linux-kernel, arnd, gregkh, Liam.Howlett,
lorenzo.stoakes, vbabka, jannh, willy, liushixin2, akpm, yang,
linux-mm, oliver.sang
hi, All,
we don't have enough knowledge to understand fully the discussion for this
patch, we saw "NACK" but there were more discussions later.
so below report is just FYI what we observed in our tests. thanks
Hello,
kernel test robot noticed a 858.5% improvement of vm-scalability.throughput on:
commit: 7143ee2391f1ea15e6791e129870473543634de2 ("[PATCH] /dev/zero: make private mapping full anonymous mapping")
url: https://github.com/intel-lab-lkp/linux/commits/Yang-Shi/dev-zero-make-private-mapping-full-anonymous-mapping/20250114-063339
base: https://git.kernel.org/cgit/linux/kernel/git/gregkh/char-misc.git a68d3cbfade64392507302f3a920113b60dc811f
patch link: https://lore.kernel.org/all/20250113223033.4054534-1-yang@os.amperecomputing.com/
patch subject: [PATCH] /dev/zero: make private mapping full anonymous mapping
testcase: vm-scalability
config: x86_64-rhel-9.4
compiler: gcc-12
test machine: 224 threads 4 sockets Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with 192G memory
parameters:
runtime: 300s
test: small-allocs
cpufreq_governor: performance
Details are as below:
-------------------------------------------------------------------------------------------------->
The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20250128/202501281038.617c6b60-lkp@intel.com
=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
commit:
a68d3cbfad ("memstick: core: fix kernel-doc notation")
7143ee2391 ("/dev/zero: make private mapping full anonymous mapping")
a68d3cbfade64392 7143ee2391f1ea15e6791e12987
---------------- ---------------------------
%stddev %change %stddev
\ | \
5.262e+09 ± 3% -67.6% 1.705e+09 ± 3% cpuidle..time
7924008 ± 3% -88.9% 875849 ± 3% cpuidle..usage
1585617 ± 5% +13.5% 1799302 ± 2% numa-numastat.node1.local_node
1667793 ± 4% +13.2% 1887467 ± 2% numa-numastat.node1.numa_hit
399.52 -78.0% 87.79 uptime.boot
14507 -24.4% 10963 uptime.idle
3408 ± 5% -99.6% 13.00 ± 40% perf-c2c.DRAM.local
18076 ± 3% -99.8% 38.67 ± 36% perf-c2c.DRAM.remote
8082 ± 5% -99.8% 19.33 ± 52% perf-c2c.HITM.local
6544 ± 6% -99.8% 14.17 ± 35% perf-c2c.HITM.remote
14627 ± 4% -99.8% 33.50 ± 34% perf-c2c.HITM.total
6.49 ± 3% +10.5 17.04 ± 7% mpstat.cpu.all.idle%
0.63 -0.3 0.35 ± 2% mpstat.cpu.all.irq%
0.03 ± 2% +0.2 0.18 ± 6% mpstat.cpu.all.soft%
91.17 -29.6 61.57 ± 2% mpstat.cpu.all.sys%
1.68 ± 2% +19.2 20.86 ± 2% mpstat.cpu.all.usr%
337.33 -95.3% 15.83 ± 35% mpstat.max_utilization.seconds
6.99 ± 3% +190.2% 20.30 ± 5% vmstat.cpu.id
91.35 -34.8% 59.59 ± 2% vmstat.cpu.sy
1.71 +1073.6% 20.04 ± 2% vmstat.cpu.us
210.36 -12.7% 183.65 vmstat.procs.r
34204 ± 5% -41.8% 19899 ± 6% vmstat.system.cs
266575 -23.1% 205001 vmstat.system.in
1609925 -50.9% 790974 meminfo.Active
1609925 -50.9% 790974 meminfo.Active(anon)
160837 ± 33% -77.3% 36534 ± 11% meminfo.AnonHugePages
4435665 -18.7% 3606310 meminfo.Cached
1775547 -44.6% 983546 meminfo.Committed_AS
148539 -47.7% 77658 ± 2% meminfo.Mapped
25332110 ± 3% -7.7% 23373667 meminfo.Memused
4245538 ± 4% -26.2% 3134309 meminfo.PageTables
14166291 ± 4% -11.9% 12484042 meminfo.SUnreclaim
929777 -89.1% 100886 meminfo.Shmem
14315492 ± 4% -11.8% 12624243 meminfo.Slab
1063552 ± 4% -27.8% 767817 ± 12% numa-meminfo.node0.PageTables
125455 ±106% -83.3% 20992 ±155% numa-meminfo.node0.Shmem
48482 ± 67% -44.8% 26748 ±127% numa-meminfo.node1.Mapped
1062709 ± 4% -21.9% 829672 numa-meminfo.node1.PageTables
1058901 ± 4% -27.5% 767469 ± 14% numa-meminfo.node2.PageTables
770405 ± 30% -74.0% 200464 ± 77% numa-meminfo.node3.Active
770405 ± 30% -74.0% 200464 ± 77% numa-meminfo.node3.Active(anon)
1146977 ±108% -94.5% 63226 ±114% numa-meminfo.node3.FilePages
52663 ± 47% -97.8% 1141 ± 55% numa-meminfo.node3.Mapped
6368902 ± 20% -23.5% 4869231 ± 12% numa-meminfo.node3.MemUsed
1058539 ± 4% -27.8% 764243 ± 12% numa-meminfo.node3.PageTables
558943 ± 14% -97.0% 16946 ±195% numa-meminfo.node3.Shmem
64129 ± 4% +885.2% 631788 ± 3% vm-scalability.median
45.40 ± 5% +1368.7 1414 ± 5% vm-scalability.stddev%
14364828 ± 4% +858.5% 1.377e+08 ± 3% vm-scalability.throughput
352.76 -88.2% 41.52 ± 3% vm-scalability.time.elapsed_time
352.76 -88.2% 41.52 ± 3% vm-scalability.time.elapsed_time.max
225965 ± 7% +62.0% 365969 ± 2% vm-scalability.time.involuntary_context_switches
9.592e+08 ± 4% +11.9% 1.074e+09 vm-scalability.time.minor_page_faults
20852 -9.7% 18831 vm-scalability.time.percent_of_cpu_this_job_got
72302 -91.9% 5866 ± 4% vm-scalability.time.system_time
1260 ± 3% +54.9% 1953 vm-scalability.time.user_time
5393707 ± 5% -99.6% 21840 ± 49% vm-scalability.time.voluntary_context_switches
4.316e+09 ± 4% +11.9% 4.832e+09 vm-scalability.workload
265763 ± 4% -27.8% 191828 ± 11% numa-vmstat.node0.nr_page_table_pages
31364 ±106% -83.0% 5332 ±156% numa-vmstat.node0.nr_shmem
12205 ± 67% -44.4% 6791 ±127% numa-vmstat.node1.nr_mapped
265546 ± 4% -21.8% 207663 numa-vmstat.node1.nr_page_table_pages
1667048 ± 4% +13.2% 1886422 ± 2% numa-vmstat.node1.numa_hit
1584872 ± 5% +13.5% 1798258 ± 2% numa-vmstat.node1.numa_local
264589 ± 4% -27.1% 192920 ± 14% numa-vmstat.node2.nr_page_table_pages
192683 ± 30% -73.9% 50195 ± 76% numa-vmstat.node3.nr_active_anon
286819 ±108% -94.5% 15799 ±114% numa-vmstat.node3.nr_file_pages
13124 ± 49% -97.8% 285.03 ± 55% numa-vmstat.node3.nr_mapped
264499 ± 4% -27.4% 192027 ± 12% numa-vmstat.node3.nr_page_table_pages
139810 ± 14% -97.0% 4229 ±195% numa-vmstat.node3.nr_shmem
192683 ± 30% -73.9% 50195 ± 76% numa-vmstat.node3.nr_zone_active_anon
402515 -50.8% 197849 proc-vmstat.nr_active_anon
170568 +1.8% 173597 proc-vmstat.nr_anon_pages
78.63 ± 33% -77.4% 17.80 ± 11% proc-vmstat.nr_anon_transparent_hugepages
4257257 +1.1% 4305540 proc-vmstat.nr_dirty_background_threshold
8524925 +1.1% 8621607 proc-vmstat.nr_dirty_threshold
1109246 -18.7% 901907 proc-vmstat.nr_file_pages
42815276 +1.1% 43299295 proc-vmstat.nr_free_pages
37525 -47.6% 19653 ± 2% proc-vmstat.nr_mapped
1059932 ± 4% -26.0% 784175 proc-vmstat.nr_page_table_pages
232507 -89.1% 25298 proc-vmstat.nr_shmem
37297 -6.0% 35048 proc-vmstat.nr_slab_reclaimable
3537843 ± 4% -11.8% 3120130 proc-vmstat.nr_slab_unreclaimable
402515 -50.8% 197849 proc-vmstat.nr_zone_active_anon
61931 ± 8% -73.8% 16233 ± 34% proc-vmstat.numa_hint_faults
15755 ± 21% -89.8% 1609 ±117% proc-vmstat.numa_hint_faults_local
293942 ± 3% -66.1% 99500 ± 20% proc-vmstat.numa_pte_updates
9.608e+08 ± 4% +11.8% 1.074e+09 proc-vmstat.pgfault
55981 ± 2% -69.0% 17375 ± 8% proc-vmstat.pgreuse
0.82 ± 4% -60.7% 0.32 ± 3% perf-stat.i.MPKI
2.714e+10 ± 2% +413.1% 1.393e+11 ± 3% perf-stat.i.branch-instructions
0.11 ± 3% +0.1 0.19 ± 2% perf-stat.i.branch-miss-rate%
24932893 +321.8% 1.052e+08 ± 3% perf-stat.i.branch-misses
64.93 -7.4 57.53 perf-stat.i.cache-miss-rate%
88563288 ± 3% +50.5% 1.333e+08 ± 3% perf-stat.i.cache-misses
1.369e+08 ± 3% +55.8% 2.134e+08 ± 3% perf-stat.i.cache-references
34508 ± 4% -39.5% 20864 ± 6% perf-stat.i.context-switches
7.67 -79.6% 1.57 ± 2% perf-stat.i.cpi
7.989e+11 -7.6% 7.383e+11 ± 2% perf-stat.i.cpu-cycles
696.35 ± 2% -52.8% 328.76 ± 2% perf-stat.i.cpu-migrations
10834 ± 4% -32.9% 7272 ± 4% perf-stat.i.cycles-between-cache-misses
1.102e+11 +310.6% 4.525e+11 ± 3% perf-stat.i.instructions
0.14 +426.9% 0.75 ± 2% perf-stat.i.ipc
24.25 ± 3% +855.3% 231.63 ± 3% perf-stat.i.metric.K/sec
2722043 ± 3% +867.7% 26340617 ± 3% perf-stat.i.minor-faults
2722043 ± 3% +867.7% 26340616 ± 3% perf-stat.i.page-faults
0.81 ± 3% -63.3% 0.30 ± 2% perf-stat.overall.MPKI
0.09 -0.0 0.07 ± 2% perf-stat.overall.branch-miss-rate%
64.81 -2.1 62.72 perf-stat.overall.cache-miss-rate%
7.24 -77.5% 1.63 ± 3% perf-stat.overall.cpi
8933 ± 4% -38.7% 5479 ± 4% perf-stat.overall.cycles-between-cache-misses
0.14 +344.4% 0.61 ± 3% perf-stat.overall.ipc
9012 ± 2% -57.9% 3797 perf-stat.overall.path-length
2.701e+10 ± 2% +396.9% 1.342e+11 ± 3% perf-stat.ps.branch-instructions
24708939 +305.5% 1.002e+08 ± 4% perf-stat.ps.branch-misses
89032538 ± 3% +45.9% 1.299e+08 ± 3% perf-stat.ps.cache-misses
1.374e+08 ± 3% +50.8% 2.071e+08 ± 3% perf-stat.ps.cache-references
34266 ± 5% -41.1% 20179 ± 7% perf-stat.ps.context-switches
223334 -2.2% 218529 perf-stat.ps.cpu-clock
7.941e+11 -10.5% 7.11e+11 perf-stat.ps.cpu-cycles
693.54 ± 2% -54.7% 314.08 ± 2% perf-stat.ps.cpu-migrations
1.097e+11 +297.8% 4.362e+11 ± 3% perf-stat.ps.instructions
2710577 ± 3% +836.2% 25375552 ± 3% perf-stat.ps.minor-faults
2710577 ± 3% +836.2% 25375552 ± 3% perf-stat.ps.page-faults
223334 -2.2% 218529 perf-stat.ps.task-clock
3.886e+13 ± 2% -52.8% 1.835e+13 perf-stat.total.instructions
64052898 ± 5% -99.8% 124999 ± 22% sched_debug.cfs_rq:/.avg_vruntime.avg
95701822 ± 7% -96.4% 3453252 ± 6% sched_debug.cfs_rq:/.avg_vruntime.max
43098762 ± 6% -100.0% 148.27 ± 21% sched_debug.cfs_rq:/.avg_vruntime.min
9223270 ± 9% -94.6% 495929 ± 17% sched_debug.cfs_rq:/.avg_vruntime.stddev
0.78 ± 2% -94.6% 0.04 ± 22% sched_debug.cfs_rq:/.h_nr_running.avg
0.28 ± 7% -28.9% 0.20 ± 10% sched_debug.cfs_rq:/.h_nr_running.stddev
411536 ± 58% -100.0% 3.77 ±141% sched_debug.cfs_rq:/.left_deadline.avg
43049468 ± 22% -100.0% 844.45 ±141% sched_debug.cfs_rq:/.left_deadline.max
3836405 ± 37% -100.0% 56.30 ±141% sched_debug.cfs_rq:/.left_deadline.stddev
411536 ± 58% -100.0% 3.62 ±141% sched_debug.cfs_rq:/.left_vruntime.avg
43049467 ± 22% -100.0% 809.82 ±141% sched_debug.cfs_rq:/.left_vruntime.max
3836405 ± 37% -100.0% 53.99 ±141% sched_debug.cfs_rq:/.left_vruntime.stddev
8792 ± 28% -81.8% 1600 ±106% sched_debug.cfs_rq:/.load.avg
64052901 ± 5% -99.8% 124999 ± 22% sched_debug.cfs_rq:/.min_vruntime.avg
95701822 ± 7% -96.4% 3453252 ± 6% sched_debug.cfs_rq:/.min_vruntime.max
43098762 ± 6% -100.0% 148.27 ± 21% sched_debug.cfs_rq:/.min_vruntime.min
9223270 ± 9% -94.6% 495929 ± 17% sched_debug.cfs_rq:/.min_vruntime.stddev
0.77 ± 2% -94.6% 0.04 ± 22% sched_debug.cfs_rq:/.nr_running.avg
0.26 ± 10% -22.4% 0.20 ± 10% sched_debug.cfs_rq:/.nr_running.stddev
411536 ± 58% -100.0% 3.62 ±141% sched_debug.cfs_rq:/.right_vruntime.avg
43049467 ± 22% -100.0% 809.82 ±141% sched_debug.cfs_rq:/.right_vruntime.max
3836405 ± 37% -100.0% 53.99 ±141% sched_debug.cfs_rq:/.right_vruntime.stddev
286633 ± 43% +421.0% 1493420 ± 42% sched_debug.cfs_rq:/.runnable_avg.avg
34728895 ± 30% +380.1% 1.667e+08 ± 27% sched_debug.cfs_rq:/.runnable_avg.max
2845573 ± 30% +406.5% 14411856 ± 30% sched_debug.cfs_rq:/.runnable_avg.stddev
769.03 -85.4% 112.18 ± 6% sched_debug.cfs_rq:/.util_avg.avg
1621 ± 5% -39.3% 983.67 ± 9% sched_debug.cfs_rq:/.util_avg.max
159.12 ± 8% +26.6% 201.45 ± 6% sched_debug.cfs_rq:/.util_avg.stddev
724.17 ± 2% -98.8% 8.91 ± 43% sched_debug.cfs_rq:/.util_est.avg
1360 ± 15% -52.9% 640.17 ± 13% sched_debug.cfs_rq:/.util_est.max
234.34 ± 9% -71.0% 67.88 ± 27% sched_debug.cfs_rq:/.util_est.stddev
766944 ± 3% +18.9% 911838 sched_debug.cpu.avg_idle.avg
1067639 ± 5% +31.7% 1406047 ± 12% sched_debug.cpu.avg_idle.max
321459 ± 2% -37.0% 202531 ± 7% sched_debug.cpu.avg_idle.stddev
195573 -76.7% 45494 sched_debug.cpu.clock.avg
195596 -76.7% 45510 sched_debug.cpu.clock.max
195548 -76.7% 45471 sched_debug.cpu.clock.min
13.79 ± 3% -36.2% 8.80 ± 2% sched_debug.cpu.clock.stddev
194424 -76.7% 45370 sched_debug.cpu.clock_task.avg
194608 -76.6% 45496 sched_debug.cpu.clock_task.max
181834 -81.8% 33106 sched_debug.cpu.clock_task.min
4241 ± 2% -96.8% 134.16 ± 27% sched_debug.cpu.curr->pid.avg
9799 ± 2% -59.8% 3941 sched_debug.cpu.curr->pid.max
1365 ± 10% -49.6% 688.63 ± 13% sched_debug.cpu.curr->pid.stddev
537665 ± 4% +31.3% 705893 ± 9% sched_debug.cpu.max_idle_balance_cost.max
3119 ± 56% +590.3% 21534 ± 34% sched_debug.cpu.max_idle_balance_cost.stddev
0.00 ± 12% -70.8% 0.00 ± 12% sched_debug.cpu.next_balance.stddev
0.78 ± 2% -95.2% 0.04 ± 25% sched_debug.cpu.nr_running.avg
2.17 ± 8% -46.2% 1.17 ± 31% sched_debug.cpu.nr_running.max
0.29 ± 8% -34.0% 0.19 ± 12% sched_debug.cpu.nr_running.stddev
25773 ± 5% -97.0% 783.41 ± 5% sched_debug.cpu.nr_switches.avg
48669 ± 10% -76.8% 11301 ± 18% sched_debug.cpu.nr_switches.max
19006 ± 7% -99.2% 156.50 ± 11% sched_debug.cpu.nr_switches.min
4142 ± 8% -68.9% 1290 ± 12% sched_debug.cpu.nr_switches.stddev
0.07 ± 23% -94.0% 0.00 ± 57% sched_debug.cpu.nr_uninterruptible.avg
240.19 ± 16% -81.7% 44.00 ± 19% sched_debug.cpu.nr_uninterruptible.max
-77.92 -84.6% -12.00 sched_debug.cpu.nr_uninterruptible.min
37.87 ± 5% -85.2% 5.60 ± 12% sched_debug.cpu.nr_uninterruptible.stddev
195549 -76.7% 45480 sched_debug.cpu_clk
194699 -77.1% 44630 sched_debug.ktime
0.00 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.avg
0.17 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.max
0.01 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.stddev
196368 -76.4% 46311 sched_debug.sched_clk
95.59 -95.6 0.00 perf-profile.calltrace.cycles-pp.__mmap
95.54 -95.5 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
95.54 -95.5 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap
94.54 -94.5 0.00 perf-profile.calltrace.cycles-pp.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
94.46 -94.1 0.31 ±101% perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
94.14 -93.8 0.37 ±105% perf-profile.calltrace.cycles-pp.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff
93.79 -93.6 0.16 ±223% perf-profile.calltrace.cycles-pp.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff
93.44 -93.4 0.00 perf-profile.calltrace.cycles-pp.down_write.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap
93.40 -93.4 0.00 perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma.__mmap_region
93.33 -93.3 0.00 perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma
94.25 -93.3 0.98 ± 82% perf-profile.calltrace.cycles-pp.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
94.45 -93.0 1.40 ± 51% perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
92.89 -92.9 0.00 perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file
0.00 +1.7 1.73 ± 34% perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exec_mmap.begin_new_exec.load_elf_binary
0.00 +1.8 1.82 ± 56% perf-profile.calltrace.cycles-pp.do_pte_missing.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
0.00 +1.9 1.85 ± 31% perf-profile.calltrace.cycles-pp.__mmput.exec_mmap.begin_new_exec.load_elf_binary.search_binary_handler
0.00 +1.9 1.85 ± 31% perf-profile.calltrace.cycles-pp.begin_new_exec.load_elf_binary.search_binary_handler.exec_binprm.bprm_execve
0.00 +1.9 1.85 ± 31% perf-profile.calltrace.cycles-pp.exec_mmap.begin_new_exec.load_elf_binary.search_binary_handler.exec_binprm
0.00 +2.3 2.28 ± 38% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
0.00 +2.5 2.48 ± 25% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
0.00 +2.5 2.48 ± 25% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
0.00 +2.5 2.50 ± 48% perf-profile.calltrace.cycles-pp.zap_present_ptes.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
0.00 +2.5 2.52 ± 31% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +2.5 2.52 ± 31% perf-profile.calltrace.cycles-pp.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +2.7 2.68 ± 27% perf-profile.calltrace.cycles-pp.asm_exc_page_fault
0.00 +2.7 2.71 ± 40% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64
0.00 +2.7 2.71 ± 40% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +2.8 2.76 ± 59% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
0.00 +2.8 2.85 ± 54% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry
0.00 +2.8 2.85 ± 54% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt
0.00 +3.0 2.96 ± 53% perf-profile.calltrace.cycles-pp.unmap_vmas.exit_mmap.__mmput.exit_mm.do_exit
0.00 +3.0 2.99 ± 53% perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
0.00 +3.0 2.99 ± 53% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
0.00 +3.0 2.99 ± 53% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe._Fork
0.00 +3.0 2.99 ± 53% perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
0.00 +3.0 3.02 ± 31% perf-profile.calltrace.cycles-pp.exec_binprm.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64
0.00 +3.0 3.02 ± 31% perf-profile.calltrace.cycles-pp.load_elf_binary.search_binary_handler.exec_binprm.bprm_execve.do_execveat_common
0.00 +3.0 3.02 ± 31% perf-profile.calltrace.cycles-pp.search_binary_handler.exec_binprm.bprm_execve.do_execveat_common.__x64_sys_execve
0.00 +3.0 3.03 ± 52% perf-profile.calltrace.cycles-pp._Fork
0.00 +3.3 3.31 ± 26% perf-profile.calltrace.cycles-pp.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +3.5 3.52 ± 20% perf-profile.calltrace.cycles-pp.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
0.00 +3.5 3.52 ± 20% perf-profile.calltrace.cycles-pp.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
0.00 +3.5 3.52 ± 20% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
0.00 +3.5 3.52 ± 20% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.execve
0.00 +3.5 3.52 ± 20% perf-profile.calltrace.cycles-pp.execve
0.00 +3.5 3.54 ± 41% perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap.__mmput
0.00 +3.5 3.54 ± 41% perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap
0.00 +3.7 3.69 ± 37% perf-profile.calltrace.cycles-pp.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +3.7 3.69 ± 37% perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64
0.00 +3.9 3.89 ± 50% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter
0.00 +3.9 3.94 ± 44% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
0.00 +4.2 4.18 ± 91% perf-profile.calltrace.cycles-pp.kthread.ret_from_fork.ret_from_fork_asm
0.00 +4.2 4.18 ± 91% perf-profile.calltrace.cycles-pp.ret_from_fork.ret_from_fork_asm
0.00 +4.2 4.18 ± 91% perf-profile.calltrace.cycles-pp.ret_from_fork_asm
0.00 +5.5 5.54 ± 38% perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit
0.00 +5.8 5.85 ± 27% perf-profile.calltrace.cycles-pp.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter
0.00 +6.5 6.50 ± 62% perf-profile.calltrace.cycles-pp.handle_internal_command.main
0.00 +6.5 6.50 ± 62% perf-profile.calltrace.cycles-pp.main
0.00 +6.5 6.50 ± 62% perf-profile.calltrace.cycles-pp.run_builtin.handle_internal_command.main
0.00 +9.1 9.05 ± 54% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
0.00 +9.1 9.05 ± 54% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
0.00 +9.4 9.38 ± 52% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
0.00 +9.5 9.48 ± 52% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
0.00 +9.9 9.92 ± 57% perf-profile.calltrace.cycles-pp.read
0.00 +12.0 11.98 ± 50% perf-profile.calltrace.cycles-pp.asm_sysvec_reschedule_ipi.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
0.00 +18.8 18.83 ± 38% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +18.8 18.83 ± 38% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
1.21 ± 3% +34.3 35.50 ± 18% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
1.21 ± 3% +34.8 35.97 ± 18% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.common_startup_64
1.21 ± 3% +35.0 36.19 ± 16% perf-profile.calltrace.cycles-pp.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
1.21 ± 3% +35.1 36.30 ± 16% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
1.21 ± 3% +35.1 36.30 ± 16% perf-profile.calltrace.cycles-pp.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
1.22 ± 3% +35.5 36.71 ± 18% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.common_startup_64
1.22 ± 3% +35.5 36.71 ± 18% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.common_startup_64
1.22 ± 3% +35.5 36.71 ± 18% perf-profile.calltrace.cycles-pp.start_secondary.common_startup_64
1.22 ± 3% +36.4 37.61 ± 15% perf-profile.calltrace.cycles-pp.common_startup_64
2.19 ± 3% +49.9 52.08 ± 18% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
95.60 -95.2 0.42 ±113% perf-profile.children.cycles-pp.__mmap
94.14 -93.6 0.54 ±106% perf-profile.children.cycles-pp.__mmap_new_vma
93.79 -93.6 0.21 ±171% perf-profile.children.cycles-pp.vma_link_file
93.40 -93.4 0.00 perf-profile.children.cycles-pp.rwsem_down_write_slowpath
93.33 -93.3 0.00 perf-profile.children.cycles-pp.rwsem_optimistic_spin
93.44 -93.2 0.24 ±178% perf-profile.children.cycles-pp.down_write
94.55 -93.1 1.40 ± 51% perf-profile.children.cycles-pp.ksys_mmap_pgoff
94.25 -93.0 1.30 ± 59% perf-profile.children.cycles-pp.__mmap_region
92.91 -92.9 0.00 perf-profile.children.cycles-pp.osq_lock
94.45 -92.7 1.72 ± 34% perf-profile.children.cycles-pp.do_mmap
94.46 -92.6 1.83 ± 31% perf-profile.children.cycles-pp.vm_mmap_pgoff
95.58 -45.3 50.30 ± 6% perf-profile.children.cycles-pp.do_syscall_64
95.58 -45.2 50.40 ± 6% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
0.00 +1.2 1.22 ± 40% perf-profile.children.cycles-pp._raw_spin_lock
0.00 +1.3 1.26 ± 34% perf-profile.children.cycles-pp.seq_printf
0.00 +1.3 1.32 ± 78% perf-profile.children.cycles-pp.kmem_cache_free
0.00 +1.6 1.60 ± 42% perf-profile.children.cycles-pp.sched_balance_rq
0.00 +1.7 1.73 ± 41% perf-profile.children.cycles-pp.open_last_lookups
0.00 +1.9 1.85 ± 31% perf-profile.children.cycles-pp.begin_new_exec
0.00 +1.9 1.85 ± 31% perf-profile.children.cycles-pp.exec_mmap
0.00 +2.1 2.09 ± 40% perf-profile.children.cycles-pp.do_pte_missing
0.46 +2.4 2.85 ± 54% perf-profile.children.cycles-pp.__hrtimer_run_queues
0.53 +2.4 2.94 ± 49% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
0.53 +2.4 2.94 ± 49% perf-profile.children.cycles-pp.hrtimer_interrupt
0.00 +2.4 2.44 ±101% perf-profile.children.cycles-pp.__evlist__enable
0.00 +2.5 2.54 ± 45% perf-profile.children.cycles-pp.zap_present_ptes
0.00 +2.6 2.58 ± 54% perf-profile.children.cycles-pp.mutex_unlock
0.00 +2.7 2.68 ± 67% perf-profile.children.cycles-pp.evlist_cpu_iterator__next
0.00 +2.7 2.71 ± 40% perf-profile.children.cycles-pp.__x64_sys_exit_group
0.00 +2.7 2.71 ± 40% perf-profile.children.cycles-pp.x64_sys_call
0.00 +3.0 2.99 ± 53% perf-profile.children.cycles-pp.__do_sys_clone
0.00 +3.0 2.99 ± 53% perf-profile.children.cycles-pp.kernel_clone
0.00 +3.0 3.02 ± 31% perf-profile.children.cycles-pp.exec_binprm
0.00 +3.0 3.02 ± 31% perf-profile.children.cycles-pp.load_elf_binary
0.00 +3.0 3.02 ± 31% perf-profile.children.cycles-pp.search_binary_handler
0.00 +3.0 3.03 ± 52% perf-profile.children.cycles-pp._Fork
0.00 +3.3 3.31 ± 26% perf-profile.children.cycles-pp.bprm_execve
0.58 ± 2% +3.4 3.98 ± 47% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
0.00 +3.5 3.52 ± 20% perf-profile.children.cycles-pp.execve
0.04 ± 44% +3.7 3.72 ± 18% perf-profile.children.cycles-pp.__schedule
0.00 +3.7 3.72 ± 14% perf-profile.children.cycles-pp.__x64_sys_execve
0.00 +3.7 3.72 ± 14% perf-profile.children.cycles-pp.do_execveat_common
0.51 ± 6% +3.7 4.25 ± 31% perf-profile.children.cycles-pp.handle_mm_fault
0.00 +3.8 3.79 ± 40% perf-profile.children.cycles-pp.zap_pte_range
0.00 +3.9 3.90 ± 26% perf-profile.children.cycles-pp.do_filp_open
0.00 +3.9 3.90 ± 26% perf-profile.children.cycles-pp.path_openat
0.00 +3.9 3.91 ± 43% perf-profile.children.cycles-pp.unmap_page_range
0.00 +3.9 3.91 ± 43% perf-profile.children.cycles-pp.zap_pmd_range
1.18 +4.0 5.20 ± 19% perf-profile.children.cycles-pp.asm_exc_page_fault
0.19 ± 23% +4.0 4.21 ± 32% perf-profile.children.cycles-pp.__handle_mm_fault
0.77 ± 3% +4.0 4.79 ± 27% perf-profile.children.cycles-pp.exc_page_fault
0.76 ± 3% +4.0 4.79 ± 27% perf-profile.children.cycles-pp.do_user_addr_fault
0.00 +4.1 4.13 ± 38% perf-profile.children.cycles-pp.do_sys_openat2
0.00 +4.2 4.15 ± 35% perf-profile.children.cycles-pp.unmap_vmas
0.00 +4.2 4.18 ± 91% perf-profile.children.cycles-pp.kthread
0.00 +4.2 4.22 ± 91% perf-profile.children.cycles-pp.ret_from_fork
0.00 +4.2 4.22 ± 91% perf-profile.children.cycles-pp.ret_from_fork_asm
0.00 +4.3 4.25 ± 37% perf-profile.children.cycles-pp.__x64_sys_openat
0.00 +5.5 5.54 ± 38% perf-profile.children.cycles-pp.exit_mm
0.00 +6.1 6.09 ± 48% perf-profile.children.cycles-pp.asm_sysvec_reschedule_ipi
0.02 ±141% +6.5 6.50 ± 62% perf-profile.children.cycles-pp.__cmd_record
0.02 ±141% +6.5 6.50 ± 62% perf-profile.children.cycles-pp.cmd_record
0.02 ±141% +6.5 6.50 ± 62% perf-profile.children.cycles-pp.handle_internal_command
0.02 ±141% +6.5 6.50 ± 62% perf-profile.children.cycles-pp.main
0.02 ±141% +6.5 6.50 ± 62% perf-profile.children.cycles-pp.run_builtin
0.00 +7.3 7.28 ± 26% perf-profile.children.cycles-pp.exit_mmap
0.00 +7.4 7.40 ± 27% perf-profile.children.cycles-pp.__mmput
0.00 +8.5 8.52 ± 58% perf-profile.children.cycles-pp.seq_read_iter
0.00 +8.6 8.56 ± 52% perf-profile.children.cycles-pp.__fput
0.00 +9.1 9.05 ± 54% perf-profile.children.cycles-pp.ksys_read
0.00 +9.1 9.05 ± 54% perf-profile.children.cycles-pp.vfs_read
0.00 +9.7 9.72 ± 54% perf-profile.children.cycles-pp.read
0.00 +16.0 16.03 ± 41% perf-profile.children.cycles-pp.do_exit
0.00 +16.0 16.03 ± 41% perf-profile.children.cycles-pp.do_group_exit
1.70 ± 2% +26.7 28.38 ± 16% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
1.21 ± 3% +35.0 36.19 ± 16% perf-profile.children.cycles-pp.acpi_idle_do_entry
1.21 ± 3% +35.0 36.19 ± 16% perf-profile.children.cycles-pp.acpi_safe_halt
1.21 ± 3% +35.1 36.30 ± 16% perf-profile.children.cycles-pp.acpi_idle_enter
1.21 ± 3% +35.1 36.30 ± 16% perf-profile.children.cycles-pp.cpuidle_enter_state
1.21 ± 3% +35.2 36.40 ± 15% perf-profile.children.cycles-pp.cpuidle_enter
1.22 ± 3% +35.5 36.71 ± 18% perf-profile.children.cycles-pp.start_secondary
1.22 ± 3% +35.7 36.87 ± 15% perf-profile.children.cycles-pp.cpuidle_idle_call
1.22 ± 3% +36.4 37.61 ± 15% perf-profile.children.cycles-pp.common_startup_64
1.22 ± 3% +36.4 37.61 ± 15% perf-profile.children.cycles-pp.cpu_startup_entry
1.22 ± 3% +36.4 37.61 ± 15% perf-profile.children.cycles-pp.do_idle
92.37 -92.4 0.00 perf-profile.self.cycles-pp.osq_lock
1.19 ± 3% +29.6 30.75 ± 22% perf-profile.self.cycles-pp.acpi_safe_halt
0.17 ±142% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
0.19 ± 34% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
0.14 ± 55% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
0.14 ± 73% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
0.10 ± 66% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
0.11 ± 59% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
0.04 ±132% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
0.07 ±101% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
0.02 ± 31% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
0.02 ±143% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
0.10 ± 44% -99.5% 0.00 ±223% perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
0.12 ±145% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
0.04 ± 55% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.25 ± 41% -95.8% 0.01 ±144% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
0.11 ± 59% -99.1% 0.00 ±115% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
0.40 ± 50% -99.6% 0.00 ±223% perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.32 ±104% -100.0% 0.00 perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
0.01 ± 12% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
0.08 ± 28% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
0.01 ± 42% -90.6% 0.00 ±223% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown].[unknown]
0.18 ± 57% -99.8% 0.00 ±223% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.03 ± 83% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
0.01 ± 20% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
0.32 ± 47% -97.1% 0.01 ± 55% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
0.07 ± 20% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
0.26 ± 17% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.02 ± 60% -83.3% 0.00 ±141% perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
0.01 ±128% -100.0% 0.00 perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
0.06 ± 31% +1806.3% 1.16 ±127% perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
1.00 ±151% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
25.45 ± 94% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
4.56 ± 67% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
3.55 ± 97% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
2.13 ± 67% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
3.16 ± 78% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
0.30 ±159% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
1.61 ±100% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
0.03 ± 86% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
0.20 ±182% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
3.51 ± 21% -100.0% 0.00 ±223% perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
0.83 ±160% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
0.09 ± 31% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
3.59 ± 11% -99.6% 0.01 ±158% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
1.60 ± 69% -99.9% 0.00 ±104% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
0.81 ± 43% -99.8% 0.00 ±223% perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
1.02 ± 88% -100.0% 0.00 perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
0.02 ± 7% -100.0% 0.00 perf-sched.sch_delay.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
9.68 ± 32% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
0.01 ± 49% -92.3% 0.00 ±223% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown].[unknown]
12.26 ±109% -100.0% 0.00 ±223% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
5.60 ±139% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
0.03 ±106% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
2.11 ± 61% -99.6% 0.01 ±160% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
171.77 ±217% -99.7% 0.54 ±195% perf-sched.sch_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
3.67 ± 25% -99.7% 0.01 ± 47% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
37.84 ± 47% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
4.68 ± 36% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.21 ±169% -98.4% 0.00 ±145% perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
7.92 ±131% -99.6% 0.03 ± 75% perf-sched.sch_delay.max.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.36 ±186% -100.0% 0.00 perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
33.45 ± 3% -88.6% 3.82 ± 80% perf-sched.total_wait_and_delay.average.ms
97903 ± 4% -98.0% 1998 ± 22% perf-sched.total_wait_and_delay.count.ms
2942 ± 23% -96.3% 109.30 ± 43% perf-sched.total_wait_and_delay.max.ms
33.37 ± 3% -88.9% 3.71 ± 83% perf-sched.total_wait_time.average.ms
2942 ± 23% -97.2% 81.62 ± 52% perf-sched.total_wait_time.max.ms
3.97 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
3.08 ± 4% -96.4% 0.11 ± 94% perf-sched.wait_and_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
119.91 ± 38% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
433.73 ± 41% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
302.41 ± 5% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
1.48 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
23.24 ± 25% -95.7% 1.01 ± 23% perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
327.16 ± 9% -97.5% 8.12 ±202% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
369.37 ± 2% -96.6% 12.56 ± 89% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
0.96 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
453.60 -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
187.66 -95.3% 8.75 ± 90% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
750.07 -99.0% 7.40 ± 73% perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
1831 ± 9% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
1269 ± 8% -43.3% 719.33 ± 26% perf-sched.wait_and_delay.count.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
6.17 ± 45% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
5.00 -100.0% 0.00 perf-sched.wait_and_delay.count.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
14.33 ± 5% -100.0% 0.00 perf-sched.wait_and_delay.count.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
810.00 ± 10% -100.0% 0.00 perf-sched.wait_and_delay.count.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
3112 ± 24% -96.8% 100.67 ± 72% perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64
40.50 ± 8% -97.5% 1.00 ±100% perf-sched.wait_and_delay.count.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
13.17 ± 2% -44.3% 7.33 ± 28% perf-sched.wait_and_delay.count.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
73021 ± 3% -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
40.00 -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_timeout.kcompactd.kthread.ret_from_fork
1122 -98.5% 16.33 ± 78% perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
11323 ± 3% -93.3% 756.17 ± 25% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1887 ± 45% -99.9% 2.33 ±117% perf-sched.wait_and_delay.count.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
1238 -93.4% 81.50 ± 64% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
35.19 ± 57% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
1002 -96.9% 31.26 ± 97% perf-sched.wait_and_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
318.48 ± 65% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1000 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
966.90 ± 7% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
20.79 ± 19% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
1043 -97.6% 24.88 ±123% perf-sched.wait_and_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
1240 ± 20% -98.7% 16.23 ±202% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
500.34 -90.4% 47.79 ± 94% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
58.83 ± 39% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
505.17 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
19.77 ± 55% -68.0% 6.33 ± 54% perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
1237 ± 34% -93.3% 83.40 ± 33% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1001 -97.3% 27.51 ±141% perf-sched.wait_and_delay.max.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
2794 ± 24% -97.4% 73.62 ± 55% perf-sched.wait_and_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
49.27 ±119% -100.0% 0.00 ±223% perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
58.17 ±187% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
3.78 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
2.99 ± 4% -98.1% 0.06 ± 95% perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
3.92 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
4.71 ± 8% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
1.67 ± 20% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
2.10 ± 27% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
0.01 ± 44% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
1.67 ± 21% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
0.04 ±133% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
67.14 ± 73% -96.0% 2.67 ±208% perf-sched.wait_time.avg.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
1.65 ± 67% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
2.30 ± 14% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
42.44 ±200% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
119.87 ± 38% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
3.80 ± 18% -99.7% 0.01 ±144% perf-sched.wait_time.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
433.32 ± 41% -100.0% 0.00 perf-sched.wait_time.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
250.23 ±107% -100.0% 0.00 perf-sched.wait_time.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
29.19 ± 5% -99.0% 0.30 ± 28% perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
302.40 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
1.40 ± 6% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
4.03 ± 8% -96.6% 0.14 ±223% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
35.38 ±192% -99.9% 0.05 ±223% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
0.05 ± 40% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
0.72 ±220% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
1.00 ±120% -98.0% 0.02 ±193% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
23.07 ± 24% -95.7% 1.00 ± 23% perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
326.84 ± 9% -97.5% 8.14 ±201% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
369.18 ± 2% -98.0% 7.39 ±103% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
0.89 ± 6% -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
1.17 ± 16% -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
453.58 -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
4.42 -27.8% 3.19 ± 26% perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
187.58 -95.4% 8.69 ± 91% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.01 ±156% -100.0% 0.00 perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
750.01 -99.2% 6.24 ± 99% perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
340.69 ±135% -100.0% 0.00 ±223% perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
535.09 ±128% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
22.04 ± 32% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
1001 -98.4% 15.63 ± 97% perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
13.57 ± 17% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
13.54 ± 10% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
10.17 ± 19% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
11.35 ± 25% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
0.01 ± 32% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
10.62 ± 9% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
0.20 ±199% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
1559 ± 64% -99.8% 2.67 ±208% perf-sched.wait_time.max.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
6.93 ± 53% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
14.42 ± 22% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
159.10 ±148% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
391.02 ±171% -99.3% 2.80 ±223% perf-sched.wait_time.max.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin
318.43 ± 65% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
13.14 ± 21% -99.9% 0.01 ±158% perf-sched.wait_time.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
1000 -100.0% 0.00 perf-sched.wait_time.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
500.84 ± 99% -100.0% 0.00 perf-sched.wait_time.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
641.50 ± 23% -99.0% 6.41 ± 48% perf-sched.wait_time.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
10.75 ± 98% -93.5% 0.70 ± 9% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
966.89 ± 7% -100.0% 0.00 perf-sched.wait_time.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
15.80 ± 8% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
16.69 ± 10% -99.2% 0.14 ±223% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
41.71 ±158% -99.9% 0.05 ±223% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
11.64 ± 61% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
2.94 ±213% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
175.70 ±210% -100.0% 0.06 ±213% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
1043 -97.6% 24.88 ±123% perf-sched.wait_time.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
1240 ± 20% -98.7% 16.28 ±201% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
500.11 -94.3% 28.64 ±118% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
32.65 ± 33% -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
22.94 ± 56% -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
505.00 -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
12.20 ± 43% -60.5% 4.82 ± 7% perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
1237 ± 34% -94.0% 74.19 ± 53% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1000 -97.2% 27.51 ±141% perf-sched.wait_time.max.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.36 ±190% -100.0% 0.00 perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
2794 ± 24% -98.0% 56.88 ± 94% perf-sched.wait_time.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
--
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-28 3:14 ` kernel test robot
@ 2025-01-31 18:38 ` Yang Shi
2025-02-06 8:02 ` Oliver Sang
0 siblings, 1 reply; 35+ messages in thread
From: Yang Shi @ 2025-01-31 18:38 UTC (permalink / raw)
To: kernel test robot
Cc: oe-lkp, lkp, linux-kernel, arnd, gregkh, Liam.Howlett,
lorenzo.stoakes, vbabka, jannh, willy, liushixin2, akpm,
linux-mm
On 1/27/25 7:14 PM, kernel test robot wrote:
> hi, All,
>
> we don't have enough knowledge to understand fully the discussion for this
> patch, we saw "NACK" but there were more discussions later.
> so below report is just FYI what we observed in our tests. thanks
Thanks for the report. It was nack'ed because of the change to
smaps/maps files in proc.
>
> Hello,
>
> kernel test robot noticed a 858.5% improvement of vm-scalability.throughput on:
>
>
> commit: 7143ee2391f1ea15e6791e129870473543634de2 ("[PATCH] /dev/zero: make private mapping full anonymous mapping")
> url: https://github.com/intel-lab-lkp/linux/commits/Yang-Shi/dev-zero-make-private-mapping-full-anonymous-mapping/20250114-063339
> base: https://git.kernel.org/cgit/linux/kernel/git/gregkh/char-misc.git a68d3cbfade64392507302f3a920113b60dc811f
> patch link: https://lore.kernel.org/all/20250113223033.4054534-1-yang@os.amperecomputing.com/
> patch subject: [PATCH] /dev/zero: make private mapping full anonymous mapping
>
> testcase: vm-scalability
> config: x86_64-rhel-9.4
> compiler: gcc-12
> test machine: 224 threads 4 sockets Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with 192G memory
> parameters:
>
> runtime: 300s
> test: small-allocs
It seems this benchmark allocates huge amount of small areas (each area
is as big as 40K) by mmap'ing /dev/zero.
This patch makes /dev/zero mapping a full anonymous mapping, so the
later vma_link_file() is actually skipped, which needs acquire file rmap
lock then insert the mapping into file rmap tree. The below profiling
also showed this.
Quoted here so that we don't have to scroll down:
> 95.60 -95.2 0.42 ±113% perf-profile.children.cycles-pp.__mmap
> 94.14 -93.6 0.54 ±106% perf-profile.children.cycles-pp.__mmap_new_vma
> 93.79 -93.6 0.21 ±171% perf-profile.children.cycles-pp.vma_link_file
> 93.40 -93.4 0.00 perf-profile.children.cycles-pp.rwsem_down_write_slowpath
> 93.33 -93.3 0.00 perf-profile.children.cycles-pp.rwsem_optimistic_spin
> 93.44 -93.2 0.24 ±178% perf-profile.children.cycles-pp.down_write
> 94.55 -93.1 1.40 ± 51% perf-profile.children.cycles-pp.ksys_mmap_pgoff
> 94.25 -93.0 1.30 ± 59% perf-profile.children.cycles-pp.__mmap_region
> 92.91 -92.9 0.00 perf-profile.children.cycles-pp.osq_lock
> 94.45 -92.7 1.72 ± 34% perf-profile.children.cycles-pp.do_mmap
> 94.46 -92.6 1.83 ± 31% perf-profile.children.cycles-pp.vm_mmap_pgoff
It significantly speed up mmap for this benchmark and the rmap lock
contention is reduced significantly for both multi-processes and
multi-threads.
The benchmark itself may exaggerate the improvement, but it may really
speed up some real life workloads. For example, multiple applications
which may allocate anonymous mapping by mmap'ing /dev/zero, then they
may have contention on /dev/zero's rmap lock.
It doesn't make too much sense to link /dev/zero anonymous vmas to the
file rmap tree. So the below patch should be able to speed up the
benchmark too.
Oliver, can you please give this patch a try?
diff --git a/mm/vma.c b/mm/vma.c
index bb2119e5a0d0..1092222c40ae 100644
--- a/mm/vma.c
+++ b/mm/vma.c
@@ -1633,6 +1633,9 @@ static void unlink_file_vma_batch_process(struct
unlink_vma_file_batch *vb)
void unlink_file_vma_batch_add(struct unlink_vma_file_batch *vb,
struct vm_area_struct *vma)
{
+ if (vma_is_anonymous(vma))
+ return;
+
if (vma->vm_file == NULL)
return;
@@ -1658,6 +1661,9 @@ void unlink_file_vma(struct vm_area_struct *vma)
{
struct file *file = vma->vm_file;
+ if (vma_is_anonymous(vma))
+ return;
+
if (file) {
struct address_space *mapping = file->f_mapping;
@@ -1672,6 +1678,9 @@ void vma_link_file(struct vm_area_struct *vma)
struct file *file = vma->vm_file;
struct address_space *mapping;
+ if (vma_is_anonymous(vma))
+ return;
+
if (file) {
mapping = file->f_mapping;
i_mmap_lock_write(mapping);
Because /dev/zero's private mapping is an anonymous mapping with valid
vm_file, so we need to bail out early if the vma is anonymous even
though it has vm_file. IMHO, making /dev/zero private mapping a full
anonymous mapping looks more clean.
> cpufreq_governor: performance
>
>
>
> Details are as below:
> -------------------------------------------------------------------------------------------------->
>
>
> The kernel config and materials to reproduce are available at:
> https://download.01.org/0day-ci/archive/20250128/202501281038.617c6b60-lkp@intel.com
>
> =========================================================================================
> compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
> gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
>
> commit:
> a68d3cbfad ("memstick: core: fix kernel-doc notation")
> 7143ee2391 ("/dev/zero: make private mapping full anonymous mapping")
>
> a68d3cbfade64392 7143ee2391f1ea15e6791e12987
> ---------------- ---------------------------
> %stddev %change %stddev
> \ | \
> 5.262e+09 ± 3% -67.6% 1.705e+09 ± 3% cpuidle..time
> 7924008 ± 3% -88.9% 875849 ± 3% cpuidle..usage
> 1585617 ± 5% +13.5% 1799302 ± 2% numa-numastat.node1.local_node
> 1667793 ± 4% +13.2% 1887467 ± 2% numa-numastat.node1.numa_hit
> 399.52 -78.0% 87.79 uptime.boot
> 14507 -24.4% 10963 uptime.idle
> 3408 ± 5% -99.6% 13.00 ± 40% perf-c2c.DRAM.local
> 18076 ± 3% -99.8% 38.67 ± 36% perf-c2c.DRAM.remote
> 8082 ± 5% -99.8% 19.33 ± 52% perf-c2c.HITM.local
> 6544 ± 6% -99.8% 14.17 ± 35% perf-c2c.HITM.remote
> 14627 ± 4% -99.8% 33.50 ± 34% perf-c2c.HITM.total
> 6.49 ± 3% +10.5 17.04 ± 7% mpstat.cpu.all.idle%
> 0.63 -0.3 0.35 ± 2% mpstat.cpu.all.irq%
> 0.03 ± 2% +0.2 0.18 ± 6% mpstat.cpu.all.soft%
> 91.17 -29.6 61.57 ± 2% mpstat.cpu.all.sys%
> 1.68 ± 2% +19.2 20.86 ± 2% mpstat.cpu.all.usr%
> 337.33 -95.3% 15.83 ± 35% mpstat.max_utilization.seconds
> 6.99 ± 3% +190.2% 20.30 ± 5% vmstat.cpu.id
> 91.35 -34.8% 59.59 ± 2% vmstat.cpu.sy
> 1.71 +1073.6% 20.04 ± 2% vmstat.cpu.us
> 210.36 -12.7% 183.65 vmstat.procs.r
> 34204 ± 5% -41.8% 19899 ± 6% vmstat.system.cs
> 266575 -23.1% 205001 vmstat.system.in
> 1609925 -50.9% 790974 meminfo.Active
> 1609925 -50.9% 790974 meminfo.Active(anon)
> 160837 ± 33% -77.3% 36534 ± 11% meminfo.AnonHugePages
> 4435665 -18.7% 3606310 meminfo.Cached
> 1775547 -44.6% 983546 meminfo.Committed_AS
> 148539 -47.7% 77658 ± 2% meminfo.Mapped
> 25332110 ± 3% -7.7% 23373667 meminfo.Memused
> 4245538 ± 4% -26.2% 3134309 meminfo.PageTables
> 14166291 ± 4% -11.9% 12484042 meminfo.SUnreclaim
> 929777 -89.1% 100886 meminfo.Shmem
> 14315492 ± 4% -11.8% 12624243 meminfo.Slab
> 1063552 ± 4% -27.8% 767817 ± 12% numa-meminfo.node0.PageTables
> 125455 ±106% -83.3% 20992 ±155% numa-meminfo.node0.Shmem
> 48482 ± 67% -44.8% 26748 ±127% numa-meminfo.node1.Mapped
> 1062709 ± 4% -21.9% 829672 numa-meminfo.node1.PageTables
> 1058901 ± 4% -27.5% 767469 ± 14% numa-meminfo.node2.PageTables
> 770405 ± 30% -74.0% 200464 ± 77% numa-meminfo.node3.Active
> 770405 ± 30% -74.0% 200464 ± 77% numa-meminfo.node3.Active(anon)
> 1146977 ±108% -94.5% 63226 ±114% numa-meminfo.node3.FilePages
> 52663 ± 47% -97.8% 1141 ± 55% numa-meminfo.node3.Mapped
> 6368902 ± 20% -23.5% 4869231 ± 12% numa-meminfo.node3.MemUsed
> 1058539 ± 4% -27.8% 764243 ± 12% numa-meminfo.node3.PageTables
> 558943 ± 14% -97.0% 16946 ±195% numa-meminfo.node3.Shmem
> 64129 ± 4% +885.2% 631788 ± 3% vm-scalability.median
> 45.40 ± 5% +1368.7 1414 ± 5% vm-scalability.stddev%
> 14364828 ± 4% +858.5% 1.377e+08 ± 3% vm-scalability.throughput
> 352.76 -88.2% 41.52 ± 3% vm-scalability.time.elapsed_time
> 352.76 -88.2% 41.52 ± 3% vm-scalability.time.elapsed_time.max
> 225965 ± 7% +62.0% 365969 ± 2% vm-scalability.time.involuntary_context_switches
> 9.592e+08 ± 4% +11.9% 1.074e+09 vm-scalability.time.minor_page_faults
> 20852 -9.7% 18831 vm-scalability.time.percent_of_cpu_this_job_got
> 72302 -91.9% 5866 ± 4% vm-scalability.time.system_time
> 1260 ± 3% +54.9% 1953 vm-scalability.time.user_time
> 5393707 ± 5% -99.6% 21840 ± 49% vm-scalability.time.voluntary_context_switches
> 4.316e+09 ± 4% +11.9% 4.832e+09 vm-scalability.workload
> 265763 ± 4% -27.8% 191828 ± 11% numa-vmstat.node0.nr_page_table_pages
> 31364 ±106% -83.0% 5332 ±156% numa-vmstat.node0.nr_shmem
> 12205 ± 67% -44.4% 6791 ±127% numa-vmstat.node1.nr_mapped
> 265546 ± 4% -21.8% 207663 numa-vmstat.node1.nr_page_table_pages
> 1667048 ± 4% +13.2% 1886422 ± 2% numa-vmstat.node1.numa_hit
> 1584872 ± 5% +13.5% 1798258 ± 2% numa-vmstat.node1.numa_local
> 264589 ± 4% -27.1% 192920 ± 14% numa-vmstat.node2.nr_page_table_pages
> 192683 ± 30% -73.9% 50195 ± 76% numa-vmstat.node3.nr_active_anon
> 286819 ±108% -94.5% 15799 ±114% numa-vmstat.node3.nr_file_pages
> 13124 ± 49% -97.8% 285.03 ± 55% numa-vmstat.node3.nr_mapped
> 264499 ± 4% -27.4% 192027 ± 12% numa-vmstat.node3.nr_page_table_pages
> 139810 ± 14% -97.0% 4229 ±195% numa-vmstat.node3.nr_shmem
> 192683 ± 30% -73.9% 50195 ± 76% numa-vmstat.node3.nr_zone_active_anon
> 402515 -50.8% 197849 proc-vmstat.nr_active_anon
> 170568 +1.8% 173597 proc-vmstat.nr_anon_pages
> 78.63 ± 33% -77.4% 17.80 ± 11% proc-vmstat.nr_anon_transparent_hugepages
> 4257257 +1.1% 4305540 proc-vmstat.nr_dirty_background_threshold
> 8524925 +1.1% 8621607 proc-vmstat.nr_dirty_threshold
> 1109246 -18.7% 901907 proc-vmstat.nr_file_pages
> 42815276 +1.1% 43299295 proc-vmstat.nr_free_pages
> 37525 -47.6% 19653 ± 2% proc-vmstat.nr_mapped
> 1059932 ± 4% -26.0% 784175 proc-vmstat.nr_page_table_pages
> 232507 -89.1% 25298 proc-vmstat.nr_shmem
> 37297 -6.0% 35048 proc-vmstat.nr_slab_reclaimable
> 3537843 ± 4% -11.8% 3120130 proc-vmstat.nr_slab_unreclaimable
> 402515 -50.8% 197849 proc-vmstat.nr_zone_active_anon
> 61931 ± 8% -73.8% 16233 ± 34% proc-vmstat.numa_hint_faults
> 15755 ± 21% -89.8% 1609 ±117% proc-vmstat.numa_hint_faults_local
> 293942 ± 3% -66.1% 99500 ± 20% proc-vmstat.numa_pte_updates
> 9.608e+08 ± 4% +11.8% 1.074e+09 proc-vmstat.pgfault
> 55981 ± 2% -69.0% 17375 ± 8% proc-vmstat.pgreuse
> 0.82 ± 4% -60.7% 0.32 ± 3% perf-stat.i.MPKI
> 2.714e+10 ± 2% +413.1% 1.393e+11 ± 3% perf-stat.i.branch-instructions
> 0.11 ± 3% +0.1 0.19 ± 2% perf-stat.i.branch-miss-rate%
> 24932893 +321.8% 1.052e+08 ± 3% perf-stat.i.branch-misses
> 64.93 -7.4 57.53 perf-stat.i.cache-miss-rate%
> 88563288 ± 3% +50.5% 1.333e+08 ± 3% perf-stat.i.cache-misses
> 1.369e+08 ± 3% +55.8% 2.134e+08 ± 3% perf-stat.i.cache-references
> 34508 ± 4% -39.5% 20864 ± 6% perf-stat.i.context-switches
> 7.67 -79.6% 1.57 ± 2% perf-stat.i.cpi
> 7.989e+11 -7.6% 7.383e+11 ± 2% perf-stat.i.cpu-cycles
> 696.35 ± 2% -52.8% 328.76 ± 2% perf-stat.i.cpu-migrations
> 10834 ± 4% -32.9% 7272 ± 4% perf-stat.i.cycles-between-cache-misses
> 1.102e+11 +310.6% 4.525e+11 ± 3% perf-stat.i.instructions
> 0.14 +426.9% 0.75 ± 2% perf-stat.i.ipc
> 24.25 ± 3% +855.3% 231.63 ± 3% perf-stat.i.metric.K/sec
> 2722043 ± 3% +867.7% 26340617 ± 3% perf-stat.i.minor-faults
> 2722043 ± 3% +867.7% 26340616 ± 3% perf-stat.i.page-faults
> 0.81 ± 3% -63.3% 0.30 ± 2% perf-stat.overall.MPKI
> 0.09 -0.0 0.07 ± 2% perf-stat.overall.branch-miss-rate%
> 64.81 -2.1 62.72 perf-stat.overall.cache-miss-rate%
> 7.24 -77.5% 1.63 ± 3% perf-stat.overall.cpi
> 8933 ± 4% -38.7% 5479 ± 4% perf-stat.overall.cycles-between-cache-misses
> 0.14 +344.4% 0.61 ± 3% perf-stat.overall.ipc
> 9012 ± 2% -57.9% 3797 perf-stat.overall.path-length
> 2.701e+10 ± 2% +396.9% 1.342e+11 ± 3% perf-stat.ps.branch-instructions
> 24708939 +305.5% 1.002e+08 ± 4% perf-stat.ps.branch-misses
> 89032538 ± 3% +45.9% 1.299e+08 ± 3% perf-stat.ps.cache-misses
> 1.374e+08 ± 3% +50.8% 2.071e+08 ± 3% perf-stat.ps.cache-references
> 34266 ± 5% -41.1% 20179 ± 7% perf-stat.ps.context-switches
> 223334 -2.2% 218529 perf-stat.ps.cpu-clock
> 7.941e+11 -10.5% 7.11e+11 perf-stat.ps.cpu-cycles
> 693.54 ± 2% -54.7% 314.08 ± 2% perf-stat.ps.cpu-migrations
> 1.097e+11 +297.8% 4.362e+11 ± 3% perf-stat.ps.instructions
> 2710577 ± 3% +836.2% 25375552 ± 3% perf-stat.ps.minor-faults
> 2710577 ± 3% +836.2% 25375552 ± 3% perf-stat.ps.page-faults
> 223334 -2.2% 218529 perf-stat.ps.task-clock
> 3.886e+13 ± 2% -52.8% 1.835e+13 perf-stat.total.instructions
> 64052898 ± 5% -99.8% 124999 ± 22% sched_debug.cfs_rq:/.avg_vruntime.avg
> 95701822 ± 7% -96.4% 3453252 ± 6% sched_debug.cfs_rq:/.avg_vruntime.max
> 43098762 ± 6% -100.0% 148.27 ± 21% sched_debug.cfs_rq:/.avg_vruntime.min
> 9223270 ± 9% -94.6% 495929 ± 17% sched_debug.cfs_rq:/.avg_vruntime.stddev
> 0.78 ± 2% -94.6% 0.04 ± 22% sched_debug.cfs_rq:/.h_nr_running.avg
> 0.28 ± 7% -28.9% 0.20 ± 10% sched_debug.cfs_rq:/.h_nr_running.stddev
> 411536 ± 58% -100.0% 3.77 ±141% sched_debug.cfs_rq:/.left_deadline.avg
> 43049468 ± 22% -100.0% 844.45 ±141% sched_debug.cfs_rq:/.left_deadline.max
> 3836405 ± 37% -100.0% 56.30 ±141% sched_debug.cfs_rq:/.left_deadline.stddev
> 411536 ± 58% -100.0% 3.62 ±141% sched_debug.cfs_rq:/.left_vruntime.avg
> 43049467 ± 22% -100.0% 809.82 ±141% sched_debug.cfs_rq:/.left_vruntime.max
> 3836405 ± 37% -100.0% 53.99 ±141% sched_debug.cfs_rq:/.left_vruntime.stddev
> 8792 ± 28% -81.8% 1600 ±106% sched_debug.cfs_rq:/.load.avg
> 64052901 ± 5% -99.8% 124999 ± 22% sched_debug.cfs_rq:/.min_vruntime.avg
> 95701822 ± 7% -96.4% 3453252 ± 6% sched_debug.cfs_rq:/.min_vruntime.max
> 43098762 ± 6% -100.0% 148.27 ± 21% sched_debug.cfs_rq:/.min_vruntime.min
> 9223270 ± 9% -94.6% 495929 ± 17% sched_debug.cfs_rq:/.min_vruntime.stddev
> 0.77 ± 2% -94.6% 0.04 ± 22% sched_debug.cfs_rq:/.nr_running.avg
> 0.26 ± 10% -22.4% 0.20 ± 10% sched_debug.cfs_rq:/.nr_running.stddev
> 411536 ± 58% -100.0% 3.62 ±141% sched_debug.cfs_rq:/.right_vruntime.avg
> 43049467 ± 22% -100.0% 809.82 ±141% sched_debug.cfs_rq:/.right_vruntime.max
> 3836405 ± 37% -100.0% 53.99 ±141% sched_debug.cfs_rq:/.right_vruntime.stddev
> 286633 ± 43% +421.0% 1493420 ± 42% sched_debug.cfs_rq:/.runnable_avg.avg
> 34728895 ± 30% +380.1% 1.667e+08 ± 27% sched_debug.cfs_rq:/.runnable_avg.max
> 2845573 ± 30% +406.5% 14411856 ± 30% sched_debug.cfs_rq:/.runnable_avg.stddev
> 769.03 -85.4% 112.18 ± 6% sched_debug.cfs_rq:/.util_avg.avg
> 1621 ± 5% -39.3% 983.67 ± 9% sched_debug.cfs_rq:/.util_avg.max
> 159.12 ± 8% +26.6% 201.45 ± 6% sched_debug.cfs_rq:/.util_avg.stddev
> 724.17 ± 2% -98.8% 8.91 ± 43% sched_debug.cfs_rq:/.util_est.avg
> 1360 ± 15% -52.9% 640.17 ± 13% sched_debug.cfs_rq:/.util_est.max
> 234.34 ± 9% -71.0% 67.88 ± 27% sched_debug.cfs_rq:/.util_est.stddev
> 766944 ± 3% +18.9% 911838 sched_debug.cpu.avg_idle.avg
> 1067639 ± 5% +31.7% 1406047 ± 12% sched_debug.cpu.avg_idle.max
> 321459 ± 2% -37.0% 202531 ± 7% sched_debug.cpu.avg_idle.stddev
> 195573 -76.7% 45494 sched_debug.cpu.clock.avg
> 195596 -76.7% 45510 sched_debug.cpu.clock.max
> 195548 -76.7% 45471 sched_debug.cpu.clock.min
> 13.79 ± 3% -36.2% 8.80 ± 2% sched_debug.cpu.clock.stddev
> 194424 -76.7% 45370 sched_debug.cpu.clock_task.avg
> 194608 -76.6% 45496 sched_debug.cpu.clock_task.max
> 181834 -81.8% 33106 sched_debug.cpu.clock_task.min
> 4241 ± 2% -96.8% 134.16 ± 27% sched_debug.cpu.curr->pid.avg
> 9799 ± 2% -59.8% 3941 sched_debug.cpu.curr->pid.max
> 1365 ± 10% -49.6% 688.63 ± 13% sched_debug.cpu.curr->pid.stddev
> 537665 ± 4% +31.3% 705893 ± 9% sched_debug.cpu.max_idle_balance_cost.max
> 3119 ± 56% +590.3% 21534 ± 34% sched_debug.cpu.max_idle_balance_cost.stddev
> 0.00 ± 12% -70.8% 0.00 ± 12% sched_debug.cpu.next_balance.stddev
> 0.78 ± 2% -95.2% 0.04 ± 25% sched_debug.cpu.nr_running.avg
> 2.17 ± 8% -46.2% 1.17 ± 31% sched_debug.cpu.nr_running.max
> 0.29 ± 8% -34.0% 0.19 ± 12% sched_debug.cpu.nr_running.stddev
> 25773 ± 5% -97.0% 783.41 ± 5% sched_debug.cpu.nr_switches.avg
> 48669 ± 10% -76.8% 11301 ± 18% sched_debug.cpu.nr_switches.max
> 19006 ± 7% -99.2% 156.50 ± 11% sched_debug.cpu.nr_switches.min
> 4142 ± 8% -68.9% 1290 ± 12% sched_debug.cpu.nr_switches.stddev
> 0.07 ± 23% -94.0% 0.00 ± 57% sched_debug.cpu.nr_uninterruptible.avg
> 240.19 ± 16% -81.7% 44.00 ± 19% sched_debug.cpu.nr_uninterruptible.max
> -77.92 -84.6% -12.00 sched_debug.cpu.nr_uninterruptible.min
> 37.87 ± 5% -85.2% 5.60 ± 12% sched_debug.cpu.nr_uninterruptible.stddev
> 195549 -76.7% 45480 sched_debug.cpu_clk
> 194699 -77.1% 44630 sched_debug.ktime
> 0.00 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.avg
> 0.17 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.max
> 0.01 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.stddev
> 196368 -76.4% 46311 sched_debug.sched_clk
> 95.59 -95.6 0.00 perf-profile.calltrace.cycles-pp.__mmap
> 95.54 -95.5 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
> 95.54 -95.5 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap
> 94.54 -94.5 0.00 perf-profile.calltrace.cycles-pp.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
> 94.46 -94.1 0.31 ±101% perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
> 94.14 -93.8 0.37 ±105% perf-profile.calltrace.cycles-pp.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff
> 93.79 -93.6 0.16 ±223% perf-profile.calltrace.cycles-pp.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff
> 93.44 -93.4 0.00 perf-profile.calltrace.cycles-pp.down_write.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap
> 93.40 -93.4 0.00 perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 93.33 -93.3 0.00 perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma
> 94.25 -93.3 0.98 ± 82% perf-profile.calltrace.cycles-pp.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 94.45 -93.0 1.40 ± 51% perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 92.89 -92.9 0.00 perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file
> 0.00 +1.7 1.73 ± 34% perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exec_mmap.begin_new_exec.load_elf_binary
> 0.00 +1.8 1.82 ± 56% perf-profile.calltrace.cycles-pp.do_pte_missing.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
> 0.00 +1.9 1.85 ± 31% perf-profile.calltrace.cycles-pp.__mmput.exec_mmap.begin_new_exec.load_elf_binary.search_binary_handler
> 0.00 +1.9 1.85 ± 31% perf-profile.calltrace.cycles-pp.begin_new_exec.load_elf_binary.search_binary_handler.exec_binprm.bprm_execve
> 0.00 +1.9 1.85 ± 31% perf-profile.calltrace.cycles-pp.exec_mmap.begin_new_exec.load_elf_binary.search_binary_handler.exec_binprm
> 0.00 +2.3 2.28 ± 38% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
> 0.00 +2.5 2.48 ± 25% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
> 0.00 +2.5 2.48 ± 25% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
> 0.00 +2.5 2.50 ± 48% perf-profile.calltrace.cycles-pp.zap_present_ptes.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
> 0.00 +2.5 2.52 ± 31% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +2.5 2.52 ± 31% perf-profile.calltrace.cycles-pp.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +2.7 2.68 ± 27% perf-profile.calltrace.cycles-pp.asm_exc_page_fault
> 0.00 +2.7 2.71 ± 40% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64
> 0.00 +2.7 2.71 ± 40% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +2.8 2.76 ± 59% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
> 0.00 +2.8 2.85 ± 54% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry
> 0.00 +2.8 2.85 ± 54% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt
> 0.00 +3.0 2.96 ± 53% perf-profile.calltrace.cycles-pp.unmap_vmas.exit_mmap.__mmput.exit_mm.do_exit
> 0.00 +3.0 2.99 ± 53% perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
> 0.00 +3.0 2.99 ± 53% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
> 0.00 +3.0 2.99 ± 53% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe._Fork
> 0.00 +3.0 2.99 ± 53% perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
> 0.00 +3.0 3.02 ± 31% perf-profile.calltrace.cycles-pp.exec_binprm.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64
> 0.00 +3.0 3.02 ± 31% perf-profile.calltrace.cycles-pp.load_elf_binary.search_binary_handler.exec_binprm.bprm_execve.do_execveat_common
> 0.00 +3.0 3.02 ± 31% perf-profile.calltrace.cycles-pp.search_binary_handler.exec_binprm.bprm_execve.do_execveat_common.__x64_sys_execve
> 0.00 +3.0 3.03 ± 52% perf-profile.calltrace.cycles-pp._Fork
> 0.00 +3.3 3.31 ± 26% perf-profile.calltrace.cycles-pp.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +3.5 3.52 ± 20% perf-profile.calltrace.cycles-pp.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> 0.00 +3.5 3.52 ± 20% perf-profile.calltrace.cycles-pp.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> 0.00 +3.5 3.52 ± 20% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> 0.00 +3.5 3.52 ± 20% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.execve
> 0.00 +3.5 3.52 ± 20% perf-profile.calltrace.cycles-pp.execve
> 0.00 +3.5 3.54 ± 41% perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap.__mmput
> 0.00 +3.5 3.54 ± 41% perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap
> 0.00 +3.7 3.69 ± 37% perf-profile.calltrace.cycles-pp.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +3.7 3.69 ± 37% perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64
> 0.00 +3.9 3.89 ± 50% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter
> 0.00 +3.9 3.94 ± 44% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
> 0.00 +4.2 4.18 ± 91% perf-profile.calltrace.cycles-pp.kthread.ret_from_fork.ret_from_fork_asm
> 0.00 +4.2 4.18 ± 91% perf-profile.calltrace.cycles-pp.ret_from_fork.ret_from_fork_asm
> 0.00 +4.2 4.18 ± 91% perf-profile.calltrace.cycles-pp.ret_from_fork_asm
> 0.00 +5.5 5.54 ± 38% perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit
> 0.00 +5.8 5.85 ± 27% perf-profile.calltrace.cycles-pp.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter
> 0.00 +6.5 6.50 ± 62% perf-profile.calltrace.cycles-pp.handle_internal_command.main
> 0.00 +6.5 6.50 ± 62% perf-profile.calltrace.cycles-pp.main
> 0.00 +6.5 6.50 ± 62% perf-profile.calltrace.cycles-pp.run_builtin.handle_internal_command.main
> 0.00 +9.1 9.05 ± 54% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
> 0.00 +9.1 9.05 ± 54% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
> 0.00 +9.4 9.38 ± 52% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
> 0.00 +9.5 9.48 ± 52% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
> 0.00 +9.9 9.92 ± 57% perf-profile.calltrace.cycles-pp.read
> 0.00 +12.0 11.98 ± 50% perf-profile.calltrace.cycles-pp.asm_sysvec_reschedule_ipi.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
> 0.00 +18.8 18.83 ± 38% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +18.8 18.83 ± 38% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
> 1.21 ± 3% +34.3 35.50 ± 18% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
> 1.21 ± 3% +34.8 35.97 ± 18% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.common_startup_64
> 1.21 ± 3% +35.0 36.19 ± 16% perf-profile.calltrace.cycles-pp.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
> 1.21 ± 3% +35.1 36.30 ± 16% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
> 1.21 ± 3% +35.1 36.30 ± 16% perf-profile.calltrace.cycles-pp.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
> 1.22 ± 3% +35.5 36.71 ± 18% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.common_startup_64
> 1.22 ± 3% +35.5 36.71 ± 18% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.common_startup_64
> 1.22 ± 3% +35.5 36.71 ± 18% perf-profile.calltrace.cycles-pp.start_secondary.common_startup_64
> 1.22 ± 3% +36.4 37.61 ± 15% perf-profile.calltrace.cycles-pp.common_startup_64
> 2.19 ± 3% +49.9 52.08 ± 18% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
> 95.60 -95.2 0.42 ±113% perf-profile.children.cycles-pp.__mmap
> 94.14 -93.6 0.54 ±106% perf-profile.children.cycles-pp.__mmap_new_vma
> 93.79 -93.6 0.21 ±171% perf-profile.children.cycles-pp.vma_link_file
> 93.40 -93.4 0.00 perf-profile.children.cycles-pp.rwsem_down_write_slowpath
> 93.33 -93.3 0.00 perf-profile.children.cycles-pp.rwsem_optimistic_spin
> 93.44 -93.2 0.24 ±178% perf-profile.children.cycles-pp.down_write
> 94.55 -93.1 1.40 ± 51% perf-profile.children.cycles-pp.ksys_mmap_pgoff
> 94.25 -93.0 1.30 ± 59% perf-profile.children.cycles-pp.__mmap_region
> 92.91 -92.9 0.00 perf-profile.children.cycles-pp.osq_lock
> 94.45 -92.7 1.72 ± 34% perf-profile.children.cycles-pp.do_mmap
> 94.46 -92.6 1.83 ± 31% perf-profile.children.cycles-pp.vm_mmap_pgoff
> 95.58 -45.3 50.30 ± 6% perf-profile.children.cycles-pp.do_syscall_64
> 95.58 -45.2 50.40 ± 6% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
> 0.00 +1.2 1.22 ± 40% perf-profile.children.cycles-pp._raw_spin_lock
> 0.00 +1.3 1.26 ± 34% perf-profile.children.cycles-pp.seq_printf
> 0.00 +1.3 1.32 ± 78% perf-profile.children.cycles-pp.kmem_cache_free
> 0.00 +1.6 1.60 ± 42% perf-profile.children.cycles-pp.sched_balance_rq
> 0.00 +1.7 1.73 ± 41% perf-profile.children.cycles-pp.open_last_lookups
> 0.00 +1.9 1.85 ± 31% perf-profile.children.cycles-pp.begin_new_exec
> 0.00 +1.9 1.85 ± 31% perf-profile.children.cycles-pp.exec_mmap
> 0.00 +2.1 2.09 ± 40% perf-profile.children.cycles-pp.do_pte_missing
> 0.46 +2.4 2.85 ± 54% perf-profile.children.cycles-pp.__hrtimer_run_queues
> 0.53 +2.4 2.94 ± 49% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
> 0.53 +2.4 2.94 ± 49% perf-profile.children.cycles-pp.hrtimer_interrupt
> 0.00 +2.4 2.44 ±101% perf-profile.children.cycles-pp.__evlist__enable
> 0.00 +2.5 2.54 ± 45% perf-profile.children.cycles-pp.zap_present_ptes
> 0.00 +2.6 2.58 ± 54% perf-profile.children.cycles-pp.mutex_unlock
> 0.00 +2.7 2.68 ± 67% perf-profile.children.cycles-pp.evlist_cpu_iterator__next
> 0.00 +2.7 2.71 ± 40% perf-profile.children.cycles-pp.__x64_sys_exit_group
> 0.00 +2.7 2.71 ± 40% perf-profile.children.cycles-pp.x64_sys_call
> 0.00 +3.0 2.99 ± 53% perf-profile.children.cycles-pp.__do_sys_clone
> 0.00 +3.0 2.99 ± 53% perf-profile.children.cycles-pp.kernel_clone
> 0.00 +3.0 3.02 ± 31% perf-profile.children.cycles-pp.exec_binprm
> 0.00 +3.0 3.02 ± 31% perf-profile.children.cycles-pp.load_elf_binary
> 0.00 +3.0 3.02 ± 31% perf-profile.children.cycles-pp.search_binary_handler
> 0.00 +3.0 3.03 ± 52% perf-profile.children.cycles-pp._Fork
> 0.00 +3.3 3.31 ± 26% perf-profile.children.cycles-pp.bprm_execve
> 0.58 ± 2% +3.4 3.98 ± 47% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
> 0.00 +3.5 3.52 ± 20% perf-profile.children.cycles-pp.execve
> 0.04 ± 44% +3.7 3.72 ± 18% perf-profile.children.cycles-pp.__schedule
> 0.00 +3.7 3.72 ± 14% perf-profile.children.cycles-pp.__x64_sys_execve
> 0.00 +3.7 3.72 ± 14% perf-profile.children.cycles-pp.do_execveat_common
> 0.51 ± 6% +3.7 4.25 ± 31% perf-profile.children.cycles-pp.handle_mm_fault
> 0.00 +3.8 3.79 ± 40% perf-profile.children.cycles-pp.zap_pte_range
> 0.00 +3.9 3.90 ± 26% perf-profile.children.cycles-pp.do_filp_open
> 0.00 +3.9 3.90 ± 26% perf-profile.children.cycles-pp.path_openat
> 0.00 +3.9 3.91 ± 43% perf-profile.children.cycles-pp.unmap_page_range
> 0.00 +3.9 3.91 ± 43% perf-profile.children.cycles-pp.zap_pmd_range
> 1.18 +4.0 5.20 ± 19% perf-profile.children.cycles-pp.asm_exc_page_fault
> 0.19 ± 23% +4.0 4.21 ± 32% perf-profile.children.cycles-pp.__handle_mm_fault
> 0.77 ± 3% +4.0 4.79 ± 27% perf-profile.children.cycles-pp.exc_page_fault
> 0.76 ± 3% +4.0 4.79 ± 27% perf-profile.children.cycles-pp.do_user_addr_fault
> 0.00 +4.1 4.13 ± 38% perf-profile.children.cycles-pp.do_sys_openat2
> 0.00 +4.2 4.15 ± 35% perf-profile.children.cycles-pp.unmap_vmas
> 0.00 +4.2 4.18 ± 91% perf-profile.children.cycles-pp.kthread
> 0.00 +4.2 4.22 ± 91% perf-profile.children.cycles-pp.ret_from_fork
> 0.00 +4.2 4.22 ± 91% perf-profile.children.cycles-pp.ret_from_fork_asm
> 0.00 +4.3 4.25 ± 37% perf-profile.children.cycles-pp.__x64_sys_openat
> 0.00 +5.5 5.54 ± 38% perf-profile.children.cycles-pp.exit_mm
> 0.00 +6.1 6.09 ± 48% perf-profile.children.cycles-pp.asm_sysvec_reschedule_ipi
> 0.02 ±141% +6.5 6.50 ± 62% perf-profile.children.cycles-pp.__cmd_record
> 0.02 ±141% +6.5 6.50 ± 62% perf-profile.children.cycles-pp.cmd_record
> 0.02 ±141% +6.5 6.50 ± 62% perf-profile.children.cycles-pp.handle_internal_command
> 0.02 ±141% +6.5 6.50 ± 62% perf-profile.children.cycles-pp.main
> 0.02 ±141% +6.5 6.50 ± 62% perf-profile.children.cycles-pp.run_builtin
> 0.00 +7.3 7.28 ± 26% perf-profile.children.cycles-pp.exit_mmap
> 0.00 +7.4 7.40 ± 27% perf-profile.children.cycles-pp.__mmput
> 0.00 +8.5 8.52 ± 58% perf-profile.children.cycles-pp.seq_read_iter
> 0.00 +8.6 8.56 ± 52% perf-profile.children.cycles-pp.__fput
> 0.00 +9.1 9.05 ± 54% perf-profile.children.cycles-pp.ksys_read
> 0.00 +9.1 9.05 ± 54% perf-profile.children.cycles-pp.vfs_read
> 0.00 +9.7 9.72 ± 54% perf-profile.children.cycles-pp.read
> 0.00 +16.0 16.03 ± 41% perf-profile.children.cycles-pp.do_exit
> 0.00 +16.0 16.03 ± 41% perf-profile.children.cycles-pp.do_group_exit
> 1.70 ± 2% +26.7 28.38 ± 16% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
> 1.21 ± 3% +35.0 36.19 ± 16% perf-profile.children.cycles-pp.acpi_idle_do_entry
> 1.21 ± 3% +35.0 36.19 ± 16% perf-profile.children.cycles-pp.acpi_safe_halt
> 1.21 ± 3% +35.1 36.30 ± 16% perf-profile.children.cycles-pp.acpi_idle_enter
> 1.21 ± 3% +35.1 36.30 ± 16% perf-profile.children.cycles-pp.cpuidle_enter_state
> 1.21 ± 3% +35.2 36.40 ± 15% perf-profile.children.cycles-pp.cpuidle_enter
> 1.22 ± 3% +35.5 36.71 ± 18% perf-profile.children.cycles-pp.start_secondary
> 1.22 ± 3% +35.7 36.87 ± 15% perf-profile.children.cycles-pp.cpuidle_idle_call
> 1.22 ± 3% +36.4 37.61 ± 15% perf-profile.children.cycles-pp.common_startup_64
> 1.22 ± 3% +36.4 37.61 ± 15% perf-profile.children.cycles-pp.cpu_startup_entry
> 1.22 ± 3% +36.4 37.61 ± 15% perf-profile.children.cycles-pp.do_idle
> 92.37 -92.4 0.00 perf-profile.self.cycles-pp.osq_lock
> 1.19 ± 3% +29.6 30.75 ± 22% perf-profile.self.cycles-pp.acpi_safe_halt
> 0.17 ±142% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
> 0.19 ± 34% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 0.14 ± 55% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
> 0.14 ± 73% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
> 0.10 ± 66% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
> 0.11 ± 59% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 0.04 ±132% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
> 0.07 ±101% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 0.02 ± 31% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
> 0.02 ±143% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
> 0.10 ± 44% -99.5% 0.00 ±223% perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
> 0.12 ±145% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
> 0.04 ± 55% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 0.25 ± 41% -95.8% 0.01 ±144% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
> 0.11 ± 59% -99.1% 0.00 ±115% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
> 0.40 ± 50% -99.6% 0.00 ±223% perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 0.32 ±104% -100.0% 0.00 perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
> 0.01 ± 12% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 0.08 ± 28% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 0.01 ± 42% -90.6% 0.00 ±223% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown].[unknown]
> 0.18 ± 57% -99.8% 0.00 ±223% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
> 0.03 ± 83% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
> 0.01 ± 20% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
> 0.32 ± 47% -97.1% 0.01 ± 55% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 0.07 ± 20% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 0.26 ± 17% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
> 0.02 ± 60% -83.3% 0.00 ±141% perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
> 0.01 ±128% -100.0% 0.00 perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
> 0.06 ± 31% +1806.3% 1.16 ±127% perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 1.00 ±151% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
> 25.45 ± 94% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 4.56 ± 67% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
> 3.55 ± 97% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
> 2.13 ± 67% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
> 3.16 ± 78% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 0.30 ±159% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
> 1.61 ±100% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 0.03 ± 86% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
> 0.20 ±182% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
> 3.51 ± 21% -100.0% 0.00 ±223% perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
> 0.83 ±160% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
> 0.09 ± 31% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 3.59 ± 11% -99.6% 0.01 ±158% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
> 1.60 ± 69% -99.9% 0.00 ±104% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
> 0.81 ± 43% -99.8% 0.00 ±223% perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 1.02 ± 88% -100.0% 0.00 perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
> 0.02 ± 7% -100.0% 0.00 perf-sched.sch_delay.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 9.68 ± 32% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 0.01 ± 49% -92.3% 0.00 ±223% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown].[unknown]
> 12.26 ±109% -100.0% 0.00 ±223% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
> 5.60 ±139% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
> 0.03 ±106% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
> 2.11 ± 61% -99.6% 0.01 ±160% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
> 171.77 ±217% -99.7% 0.54 ±195% perf-sched.sch_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
> 3.67 ± 25% -99.7% 0.01 ± 47% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 37.84 ± 47% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 4.68 ± 36% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
> 0.21 ±169% -98.4% 0.00 ±145% perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
> 7.92 ±131% -99.6% 0.03 ± 75% perf-sched.sch_delay.max.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 0.36 ±186% -100.0% 0.00 perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
> 33.45 ± 3% -88.6% 3.82 ± 80% perf-sched.total_wait_and_delay.average.ms
> 97903 ± 4% -98.0% 1998 ± 22% perf-sched.total_wait_and_delay.count.ms
> 2942 ± 23% -96.3% 109.30 ± 43% perf-sched.total_wait_and_delay.max.ms
> 33.37 ± 3% -88.9% 3.71 ± 83% perf-sched.total_wait_time.average.ms
> 2942 ± 23% -97.2% 81.62 ± 52% perf-sched.total_wait_time.max.ms
> 3.97 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 3.08 ± 4% -96.4% 0.11 ± 94% perf-sched.wait_and_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
> 119.91 ± 38% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 433.73 ± 41% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 302.41 ± 5% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 1.48 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 23.24 ± 25% -95.7% 1.01 ± 23% perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
> 327.16 ± 9% -97.5% 8.12 ±202% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 369.37 ± 2% -96.6% 12.56 ± 89% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 0.96 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 453.60 -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
> 187.66 -95.3% 8.75 ± 90% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 750.07 -99.0% 7.40 ± 73% perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 1831 ± 9% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 1269 ± 8% -43.3% 719.33 ± 26% perf-sched.wait_and_delay.count.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
> 6.17 ± 45% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 5.00 -100.0% 0.00 perf-sched.wait_and_delay.count.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 14.33 ± 5% -100.0% 0.00 perf-sched.wait_and_delay.count.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 810.00 ± 10% -100.0% 0.00 perf-sched.wait_and_delay.count.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 3112 ± 24% -96.8% 100.67 ± 72% perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64
> 40.50 ± 8% -97.5% 1.00 ±100% perf-sched.wait_and_delay.count.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 13.17 ± 2% -44.3% 7.33 ± 28% perf-sched.wait_and_delay.count.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 73021 ± 3% -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 40.00 -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_timeout.kcompactd.kthread.ret_from_fork
> 1122 -98.5% 16.33 ± 78% perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
> 11323 ± 3% -93.3% 756.17 ± 25% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 1887 ± 45% -99.9% 2.33 ±117% perf-sched.wait_and_delay.count.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 1238 -93.4% 81.50 ± 64% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 35.19 ± 57% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 1002 -96.9% 31.26 ± 97% perf-sched.wait_and_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
> 318.48 ± 65% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 1000 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 966.90 ± 7% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 20.79 ± 19% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 1043 -97.6% 24.88 ±123% perf-sched.wait_and_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
> 1240 ± 20% -98.7% 16.23 ±202% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 500.34 -90.4% 47.79 ± 94% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 58.83 ± 39% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 505.17 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
> 19.77 ± 55% -68.0% 6.33 ± 54% perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
> 1237 ± 34% -93.3% 83.40 ± 33% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 1001 -97.3% 27.51 ±141% perf-sched.wait_and_delay.max.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 2794 ± 24% -97.4% 73.62 ± 55% perf-sched.wait_and_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 49.27 ±119% -100.0% 0.00 ±223% perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
> 58.17 ±187% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
> 3.78 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 2.99 ± 4% -98.1% 0.06 ± 95% perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
> 3.92 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
> 4.71 ± 8% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
> 1.67 ± 20% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
> 2.10 ± 27% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 0.01 ± 44% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
> 1.67 ± 21% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 0.04 ±133% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
> 67.14 ± 73% -96.0% 2.67 ±208% perf-sched.wait_time.avg.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
> 1.65 ± 67% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
> 2.30 ± 14% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
> 42.44 ±200% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
> 119.87 ± 38% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 3.80 ± 18% -99.7% 0.01 ±144% perf-sched.wait_time.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
> 433.32 ± 41% -100.0% 0.00 perf-sched.wait_time.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 250.23 ±107% -100.0% 0.00 perf-sched.wait_time.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
> 29.19 ± 5% -99.0% 0.30 ± 28% perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
> 302.40 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 1.40 ± 6% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 4.03 ± 8% -96.6% 0.14 ±223% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
> 35.38 ±192% -99.9% 0.05 ±223% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
> 0.05 ± 40% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
> 0.72 ±220% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
> 1.00 ±120% -98.0% 0.02 ±193% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
> 23.07 ± 24% -95.7% 1.00 ± 23% perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
> 326.84 ± 9% -97.5% 8.14 ±201% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 369.18 ± 2% -98.0% 7.39 ±103% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 0.89 ± 6% -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 1.17 ± 16% -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
> 453.58 -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
> 4.42 -27.8% 3.19 ± 26% perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
> 187.58 -95.4% 8.69 ± 91% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 0.01 ±156% -100.0% 0.00 perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
> 750.01 -99.2% 6.24 ± 99% perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 340.69 ±135% -100.0% 0.00 ±223% perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
> 535.09 ±128% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
> 22.04 ± 32% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 1001 -98.4% 15.63 ± 97% perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
> 13.57 ± 17% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
> 13.54 ± 10% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
> 10.17 ± 19% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
> 11.35 ± 25% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 0.01 ± 32% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
> 10.62 ± 9% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 0.20 ±199% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
> 1559 ± 64% -99.8% 2.67 ±208% perf-sched.wait_time.max.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
> 6.93 ± 53% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
> 14.42 ± 22% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
> 159.10 ±148% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
> 391.02 ±171% -99.3% 2.80 ±223% perf-sched.wait_time.max.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin
> 318.43 ± 65% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 13.14 ± 21% -99.9% 0.01 ±158% perf-sched.wait_time.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
> 1000 -100.0% 0.00 perf-sched.wait_time.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 500.84 ± 99% -100.0% 0.00 perf-sched.wait_time.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
> 641.50 ± 23% -99.0% 6.41 ± 48% perf-sched.wait_time.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
> 10.75 ± 98% -93.5% 0.70 ± 9% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 966.89 ± 7% -100.0% 0.00 perf-sched.wait_time.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 15.80 ± 8% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 16.69 ± 10% -99.2% 0.14 ±223% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
> 41.71 ±158% -99.9% 0.05 ±223% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
> 11.64 ± 61% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
> 2.94 ±213% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
> 175.70 ±210% -100.0% 0.06 ±213% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
> 1043 -97.6% 24.88 ±123% perf-sched.wait_time.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
> 1240 ± 20% -98.7% 16.28 ±201% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 500.11 -94.3% 28.64 ±118% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 32.65 ± 33% -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 22.94 ± 56% -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
> 505.00 -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
> 12.20 ± 43% -60.5% 4.82 ± 7% perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
> 1237 ± 34% -94.0% 74.19 ± 53% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 1000 -97.2% 27.51 ±141% perf-sched.wait_time.max.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 0.36 ±190% -100.0% 0.00 perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
> 2794 ± 24% -98.0% 56.88 ± 94% perf-sched.wait_time.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
>
>
>
>
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or software
> design or configuration may affect actual performance.
>
>
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-01-31 18:38 ` Yang Shi
@ 2025-02-06 8:02 ` Oliver Sang
2025-02-07 18:10 ` Yang Shi
0 siblings, 1 reply; 35+ messages in thread
From: Oliver Sang @ 2025-02-06 8:02 UTC (permalink / raw)
To: Yang Shi
Cc: oe-lkp, lkp, linux-kernel, arnd, gregkh, Liam.Howlett,
lorenzo.stoakes, vbabka, jannh, willy, liushixin2, akpm,
linux-mm, oliver.sang
hi, Yang Shi,
On Fri, Jan 31, 2025 at 10:38:03AM -0800, Yang Shi wrote:
>
>
>
> On 1/27/25 7:14 PM, kernel test robot wrote:
> > hi, All,
> >
> > we don't have enough knowledge to understand fully the discussion for this
> > patch, we saw "NACK" but there were more discussions later.
> > so below report is just FYI what we observed in our tests. thanks
>
> Thanks for the report. It was nack'ed because of the change to smaps/maps
> files in proc.
>
> >
> > Hello,
> >
> > kernel test robot noticed a 858.5% improvement of vm-scalability.throughput on:
> >
> >
> > commit: 7143ee2391f1ea15e6791e129870473543634de2 ("[PATCH] /dev/zero: make private mapping full anonymous mapping")
> > url: https://github.com/intel-lab-lkp/linux/commits/Yang-Shi/dev-zero-make-private-mapping-full-anonymous-mapping/20250114-063339
> > base: https://git.kernel.org/cgit/linux/kernel/git/gregkh/char-misc.git a68d3cbfade64392507302f3a920113b60dc811f
> > patch link: https://lore.kernel.org/all/20250113223033.4054534-1-yang@os.amperecomputing.com/
> > patch subject: [PATCH] /dev/zero: make private mapping full anonymous mapping
> >
> > testcase: vm-scalability
> > config: x86_64-rhel-9.4
> > compiler: gcc-12
> > test machine: 224 threads 4 sockets Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with 192G memory
> > parameters:
> >
> > runtime: 300s
> > test: small-allocs
>
> It seems this benchmark allocates huge amount of small areas (each area is
> as big as 40K) by mmap'ing /dev/zero.
>
> This patch makes /dev/zero mapping a full anonymous mapping, so the later
> vma_link_file() is actually skipped, which needs acquire file rmap lock then
> insert the mapping into file rmap tree. The below profiling also showed
> this.
> Quoted here so that we don't have to scroll down:
>
> > 95.60 -95.2 0.42 ±113% perf-profile.children.cycles-pp.__mmap
> > 94.14 -93.6 0.54 ±106% perf-profile.children.cycles-pp.__mmap_new_vma
> > 93.79 -93.6 0.21 ±171% perf-profile.children.cycles-pp.vma_link_file
> > 93.40 -93.4 0.00 perf-profile.children.cycles-pp.rwsem_down_write_slowpath
> > 93.33 -93.3 0.00 perf-profile.children.cycles-pp.rwsem_optimistic_spin
> > 93.44 -93.2 0.24 ±178% perf-profile.children.cycles-pp.down_write
> > 94.55 -93.1 1.40 ± 51% perf-profile.children.cycles-pp.ksys_mmap_pgoff
> > 94.25 -93.0 1.30 ± 59% perf-profile.children.cycles-pp.__mmap_region
> > 92.91 -92.9 0.00 perf-profile.children.cycles-pp.osq_lock
> > 94.45 -92.7 1.72 ± 34% perf-profile.children.cycles-pp.do_mmap
> > 94.46 -92.6 1.83 ± 31% perf-profile.children.cycles-pp.vm_mmap_pgoff
>
> It significantly speed up mmap for this benchmark and the rmap lock
> contention is reduced significantly for both multi-processes and
> multi-threads.
>
> The benchmark itself may exaggerate the improvement, but it may really speed
> up some real life workloads. For example, multiple applications which may
> allocate anonymous mapping by mmap'ing /dev/zero, then they may have
> contention on /dev/zero's rmap lock.
>
> It doesn't make too much sense to link /dev/zero anonymous vmas to the file
> rmap tree. So the below patch should be able to speed up the benchmark too.
sorry for late and thanks a lot for information!
>
> Oliver, can you please give this patch a try?
it seems this is an alternative patch?
since we applied your "/dev/zero: make private mapping full anonymous mapping"
patch upon a68d3cbfad like below:
* 7143ee2391f1e /dev/zero: make private mapping full anonymous mapping
* a68d3cbfade64 memstick: core: fix kernel-doc notation
so I applied below patch also upon a68d3cbfad.
we saw big improvement but not that big.
=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
commit:
a68d3cbfad ("memstick: core: fix kernel-doc notation")
52ec85cb99 <--- your patch
a68d3cbfade64392 52ec85cb99e9b31dc304eae965a
---------------- ---------------------------
%stddev %change %stddev
\ | \
14364828 ± 4% +410.6% 73349239 ± 3% vm-scalability.throughput
full comparison as below [1] just FYI.
>
>
> diff --git a/mm/vma.c b/mm/vma.c
> index bb2119e5a0d0..1092222c40ae 100644
> --- a/mm/vma.c
> +++ b/mm/vma.c
> @@ -1633,6 +1633,9 @@ static void unlink_file_vma_batch_process(struct
> unlink_vma_file_batch *vb)
> void unlink_file_vma_batch_add(struct unlink_vma_file_batch *vb,
> struct vm_area_struct *vma)
> {
> + if (vma_is_anonymous(vma))
> + return;
> +
> if (vma->vm_file == NULL)
> return;
>
> @@ -1658,6 +1661,9 @@ void unlink_file_vma(struct vm_area_struct *vma)
> {
> struct file *file = vma->vm_file;
>
> + if (vma_is_anonymous(vma))
> + return;
> +
> if (file) {
> struct address_space *mapping = file->f_mapping;
>
> @@ -1672,6 +1678,9 @@ void vma_link_file(struct vm_area_struct *vma)
> struct file *file = vma->vm_file;
> struct address_space *mapping;
>
> + if (vma_is_anonymous(vma))
> + return;
> +
> if (file) {
> mapping = file->f_mapping;
> i_mmap_lock_write(mapping);
>
>
> Because /dev/zero's private mapping is an anonymous mapping with valid
> vm_file, so we need to bail out early if the vma is anonymous even though it
> has vm_file. IMHO, making /dev/zero private mapping a full anonymous mapping
> looks more clean.
>
[1]
=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
commit:
a68d3cbfad ("memstick: core: fix kernel-doc notation")
52ec85cb99 <--- your patch
a68d3cbfade64392 52ec85cb99e9b31dc304eae965a
---------------- ---------------------------
%stddev %change %stddev
\ | \
5.262e+09 ± 3% -45.0% 2.896e+09 ± 6% cpuidle..time
7924008 ± 3% -79.3% 1643339 ± 11% cpuidle..usage
1871164 ± 4% -22.4% 1452554 ± 12% numa-numastat.node3.local_node
1952164 ± 3% -20.1% 1560294 ± 12% numa-numastat.node3.numa_hit
399.52 -68.2% 126.86 uptime.boot
14507 -15.7% 12232 uptime.idle
6.99 ± 3% +147.9% 17.34 ± 4% vmstat.cpu.id
1.71 +473.6% 9.79 ± 2% vmstat.cpu.us
34204 ± 5% -72.9% 9272 ± 7% vmstat.system.cs
266575 -21.2% 210191 vmstat.system.in
3408 ± 5% -99.8% 8.38 ± 48% perf-c2c.DRAM.local
18076 ± 3% -99.8% 32.25 ± 27% perf-c2c.DRAM.remote
8082 ± 5% -99.8% 15.50 ± 64% perf-c2c.HITM.local
6544 ± 6% -99.8% 13.62 ± 51% perf-c2c.HITM.remote
14627 ± 4% -99.8% 29.12 ± 53% perf-c2c.HITM.total
6.49 ± 3% +8.8 15.24 ± 5% mpstat.cpu.all.idle%
0.63 -0.3 0.32 ± 4% mpstat.cpu.all.irq%
0.03 ± 2% +0.2 0.26 ± 2% mpstat.cpu.all.soft%
91.17 -17.0 74.15 mpstat.cpu.all.sys%
1.68 ± 2% +8.3 10.03 ± 2% mpstat.cpu.all.usr%
337.33 -97.4% 8.88 ± 75% mpstat.max_utilization.seconds
352.76 -77.3% 79.95 ± 2% time.elapsed_time
352.76 -77.3% 79.95 ± 2% time.elapsed_time.max
225965 ± 7% -16.0% 189844 ± 6% time.involuntary_context_switches
9.592e+08 ± 4% +11.9% 1.074e+09 time.minor_page_faults
20852 -8.8% 19012 time.percent_of_cpu_this_job_got
72302 -81.4% 13425 ± 3% time.system_time
1260 ± 3% +41.0% 1777 time.user_time
5393707 ± 5% -98.4% 86880 ± 17% time.voluntary_context_switches
1609925 -50.3% 800493 meminfo.Active
1609925 -50.3% 800493 meminfo.Active(anon)
160837 ± 33% -63.9% 58119 ± 13% meminfo.AnonHugePages
4435665 -18.5% 3614714 meminfo.Cached
1775547 -43.8% 998415 meminfo.Committed_AS
148539 -43.7% 83699 ± 4% meminfo.Mapped
4245538 ± 4% -20.9% 3356561 meminfo.PageTables
14166291 ± 4% -9.6% 12806082 meminfo.SUnreclaim
929777 -88.2% 109274 ± 3% meminfo.Shmem
14315492 ± 4% -9.6% 12947821 meminfo.Slab
64129 ± 4% +418.9% 332751 ± 3% vm-scalability.median
45.40 ± 5% +1961.8 2007 ± 8% vm-scalability.stddev%
14364828 ± 4% +410.6% 73349239 ± 3% vm-scalability.throughput
352.76 -77.3% 79.95 ± 2% vm-scalability.time.elapsed_time
352.76 -77.3% 79.95 ± 2% vm-scalability.time.elapsed_time.max
225965 ± 7% -16.0% 189844 ± 6% vm-scalability.time.involuntary_context_switches
9.592e+08 ± 4% +11.9% 1.074e+09 vm-scalability.time.minor_page_faults
20852 -8.8% 19012 vm-scalability.time.percent_of_cpu_this_job_got
72302 -81.4% 13425 ± 3% vm-scalability.time.system_time
1260 ± 3% +41.0% 1777 vm-scalability.time.user_time
5393707 ± 5% -98.4% 86880 ± 17% vm-scalability.time.voluntary_context_switches
4.316e+09 ± 4% +11.9% 4.832e+09 vm-scalability.workload
265763 ± 4% -20.5% 211398 ± 4% numa-vmstat.node0.nr_page_table_pages
31364 ±106% -85.0% 4690 ±169% numa-vmstat.node0.nr_shmem
12205 ± 67% -74.1% 3161 ±199% numa-vmstat.node1.nr_mapped
265546 ± 4% -21.8% 207742 ± 4% numa-vmstat.node1.nr_page_table_pages
44052 ± 71% -86.0% 6163 ±161% numa-vmstat.node1.nr_shmem
885590 ± 4% -9.9% 797649 ± 4% numa-vmstat.node1.nr_slab_unreclaimable
264589 ± 4% -21.2% 208598 ± 4% numa-vmstat.node2.nr_page_table_pages
881598 ± 4% -10.0% 793829 ± 4% numa-vmstat.node2.nr_slab_unreclaimable
192683 ± 30% -61.0% 75078 ± 70% numa-vmstat.node3.nr_active_anon
286819 ±108% -93.0% 19993 ± 39% numa-vmstat.node3.nr_file_pages
13124 ± 49% -92.3% 1006 ± 57% numa-vmstat.node3.nr_mapped
264499 ± 4% -22.1% 206135 ± 2% numa-vmstat.node3.nr_page_table_pages
139810 ± 14% -90.5% 13229 ± 89% numa-vmstat.node3.nr_shmem
880199 ± 4% -11.8% 776210 ± 5% numa-vmstat.node3.nr_slab_unreclaimable
192683 ± 30% -61.0% 75077 ± 70% numa-vmstat.node3.nr_zone_active_anon
1951359 ± 3% -20.1% 1558936 ± 12% numa-vmstat.node3.numa_hit
1870359 ± 4% -22.4% 1451195 ± 12% numa-vmstat.node3.numa_local
402515 -50.3% 200150 proc-vmstat.nr_active_anon
170568 +1.9% 173746 proc-vmstat.nr_anon_pages
4257257 +0.9% 4296664 proc-vmstat.nr_dirty_background_threshold
8524925 +0.9% 8603835 proc-vmstat.nr_dirty_threshold
1109246 -18.5% 903959 proc-vmstat.nr_file_pages
42815276 +0.9% 43210344 proc-vmstat.nr_free_pages
37525 -43.6% 21164 ± 4% proc-vmstat.nr_mapped
1059932 ± 4% -21.1% 836810 proc-vmstat.nr_page_table_pages
232507 -88.2% 27341 ± 3% proc-vmstat.nr_shmem
37297 -5.0% 35436 proc-vmstat.nr_slab_reclaimable
3537843 ± 4% -9.8% 3192506 proc-vmstat.nr_slab_unreclaimable
402515 -50.3% 200150 proc-vmstat.nr_zone_active_anon
61931 ± 8% -83.8% 10023 ± 45% proc-vmstat.numa_hint_faults
15755 ± 21% -87.1% 2039 ± 97% proc-vmstat.numa_hint_faults_local
6916516 ± 3% -7.1% 6425430 proc-vmstat.numa_hit
6568542 ± 3% -7.5% 6077764 proc-vmstat.numa_local
293942 ± 3% -69.6% 89435 ± 49% proc-vmstat.numa_pte_updates
9.608e+08 ± 4% +11.8% 1.074e+09 proc-vmstat.pgfault
55981 ± 2% -63.1% 20641 ± 2% proc-vmstat.pgreuse
1063552 ± 4% -20.3% 847673 ± 4% numa-meminfo.node0.PageTables
3565610 ± 4% -8.0% 3279375 ± 3% numa-meminfo.node0.SUnreclaim
125455 ±106% -85.2% 18620 ±168% numa-meminfo.node0.Shmem
3592377 ± 4% -7.1% 3336072 ± 4% numa-meminfo.node0.Slab
48482 ± 67% -74.3% 12475 ±199% numa-meminfo.node1.Mapped
1062709 ± 4% -21.7% 831966 ± 4% numa-meminfo.node1.PageTables
3543793 ± 4% -10.0% 3189589 ± 4% numa-meminfo.node1.SUnreclaim
176171 ± 71% -86.0% 24677 ±161% numa-meminfo.node1.Shmem
3593431 ± 4% -10.4% 3220352 ± 4% numa-meminfo.node1.Slab
1058901 ± 4% -21.3% 833124 ± 4% numa-meminfo.node2.PageTables
3527862 ± 4% -10.2% 3168666 ± 5% numa-meminfo.node2.SUnreclaim
3565750 ± 4% -10.3% 3200248 ± 5% numa-meminfo.node2.Slab
770405 ± 30% -61.0% 300435 ± 70% numa-meminfo.node3.Active
770405 ± 30% -61.0% 300435 ± 70% numa-meminfo.node3.Active(anon)
1146977 ±108% -93.0% 80110 ± 40% numa-meminfo.node3.FilePages
52663 ± 47% -91.6% 4397 ± 56% numa-meminfo.node3.Mapped
6368902 ± 20% -21.2% 5021246 ± 2% numa-meminfo.node3.MemUsed
1058539 ± 4% -22.2% 823061 ± 3% numa-meminfo.node3.PageTables
3522496 ± 4% -12.1% 3096728 ± 6% numa-meminfo.node3.SUnreclaim
558943 ± 14% -90.5% 53054 ± 89% numa-meminfo.node3.Shmem
3557392 ± 4% -12.3% 3119454 ± 6% numa-meminfo.node3.Slab
0.82 ± 4% -39.7% 0.50 ± 12% perf-stat.i.MPKI
2.714e+10 ± 2% +185.7% 7.755e+10 ± 6% perf-stat.i.branch-instructions
0.11 ± 3% +0.1 0.20 ± 5% perf-stat.i.branch-miss-rate%
24932893 +156.6% 63980942 ± 5% perf-stat.i.branch-misses
64.93 -10.1 54.87 ± 2% perf-stat.i.cache-miss-rate%
34508 ± 4% -61.4% 13315 ± 10% perf-stat.i.context-switches
7.67 -63.7% 2.79 ± 6% perf-stat.i.cpi
224605 +10.8% 248972 ± 4% perf-stat.i.cpu-clock
696.35 ± 2% -57.4% 296.79 ± 3% perf-stat.i.cpu-migrations
1.102e+11 +128.5% 2.518e+11 ± 6% perf-stat.i.instructions
0.14 +198.2% 0.42 ± 5% perf-stat.i.ipc
24.25 ± 3% +375.8% 115.36 ± 3% perf-stat.i.metric.K/sec
2722043 ± 3% +439.7% 14690226 ± 6% perf-stat.i.minor-faults
2722043 ± 3% +439.7% 14690226 ± 6% perf-stat.i.page-faults
224605 +10.8% 248972 ± 4% perf-stat.i.task-clock
0.81 ± 3% -52.5% 0.39 ± 14% perf-stat.overall.MPKI
0.09 -0.0 0.08 ± 2% perf-stat.overall.branch-miss-rate%
64.81 -6.4 58.40 perf-stat.overall.cache-miss-rate%
7.24 -56.3% 3.17 ± 3% perf-stat.overall.cpi
0.14 +129.0% 0.32 ± 3% perf-stat.overall.ipc
9012 ± 2% -57.5% 3827 perf-stat.overall.path-length
2.701e+10 ± 2% +159.6% 7.012e+10 ± 2% perf-stat.ps.branch-instructions
24708939 +119.2% 54173035 perf-stat.ps.branch-misses
34266 ± 5% -73.9% 8949 ± 7% perf-stat.ps.context-switches
7.941e+11 -9.1% 7.219e+11 perf-stat.ps.cpu-cycles
693.54 ± 2% -68.6% 217.73 ± 5% perf-stat.ps.cpu-migrations
1.097e+11 +108.1% 2.282e+11 ± 2% perf-stat.ps.instructions
2710577 ± 3% +388.7% 13246535 ± 2% perf-stat.ps.minor-faults
2710577 ± 3% +388.7% 13246536 ± 2% perf-stat.ps.page-faults
3.886e+13 ± 2% -52.4% 1.849e+13 perf-stat.total.instructions
64052898 ± 5% -96.2% 2460331 ±166% sched_debug.cfs_rq:/.avg_vruntime.avg
95701822 ± 7% -85.1% 14268127 ±116% sched_debug.cfs_rq:/.avg_vruntime.max
43098762 ± 6% -96.0% 1715136 ±173% sched_debug.cfs_rq:/.avg_vruntime.min
9223270 ± 9% -84.2% 1457904 ±122% sched_debug.cfs_rq:/.avg_vruntime.stddev
0.78 ± 2% -77.0% 0.18 ±130% sched_debug.cfs_rq:/.h_nr_running.avg
43049468 ± 22% -89.3% 4590302 ±180% sched_debug.cfs_rq:/.left_deadline.max
3836405 ± 37% -85.6% 550773 ±176% sched_debug.cfs_rq:/.left_deadline.stddev
43049467 ± 22% -89.3% 4590279 ±180% sched_debug.cfs_rq:/.left_vruntime.max
3836405 ± 37% -85.6% 550772 ±176% sched_debug.cfs_rq:/.left_vruntime.stddev
64052901 ± 5% -96.2% 2460341 ±166% sched_debug.cfs_rq:/.min_vruntime.avg
95701822 ± 7% -85.1% 14268127 ±116% sched_debug.cfs_rq:/.min_vruntime.max
43098762 ± 6% -96.0% 1715136 ±173% sched_debug.cfs_rq:/.min_vruntime.min
9223270 ± 9% -84.2% 1457902 ±122% sched_debug.cfs_rq:/.min_vruntime.stddev
0.77 ± 2% -77.4% 0.17 ±128% sched_debug.cfs_rq:/.nr_running.avg
1.61 ± 24% +396.0% 7.96 ± 62% sched_debug.cfs_rq:/.removed.runnable_avg.avg
86.69 +424.4% 454.62 ± 24% sched_debug.cfs_rq:/.removed.runnable_avg.max
11.14 ± 13% +409.8% 56.79 ± 35% sched_debug.cfs_rq:/.removed.runnable_avg.stddev
1.61 ± 24% +396.0% 7.96 ± 62% sched_debug.cfs_rq:/.removed.util_avg.avg
86.69 +424.4% 454.62 ± 24% sched_debug.cfs_rq:/.removed.util_avg.max
11.14 ± 13% +409.8% 56.79 ± 35% sched_debug.cfs_rq:/.removed.util_avg.stddev
43049467 ± 22% -89.3% 4590282 ±180% sched_debug.cfs_rq:/.right_vruntime.max
3836405 ± 37% -85.6% 550772 ±176% sched_debug.cfs_rq:/.right_vruntime.stddev
286633 ± 43% +262.3% 1038592 ± 36% sched_debug.cfs_rq:/.runnable_avg.avg
34728895 ± 30% +349.2% 1.56e+08 ± 26% sched_debug.cfs_rq:/.runnable_avg.max
2845573 ± 30% +325.9% 12119045 ± 26% sched_debug.cfs_rq:/.runnable_avg.stddev
769.03 -69.9% 231.86 ± 84% sched_debug.cfs_rq:/.util_avg.avg
1621 ± 5% -31.5% 1111 ± 8% sched_debug.cfs_rq:/.util_avg.max
724.17 ± 2% -89.6% 75.66 ±147% sched_debug.cfs_rq:/.util_est.avg
1360 ± 15% -39.2% 826.88 ± 37% sched_debug.cfs_rq:/.util_est.max
766944 ± 3% +18.1% 905901 sched_debug.cpu.avg_idle.avg
321459 ± 2% -35.6% 207172 ± 10% sched_debug.cpu.avg_idle.stddev
195573 -72.7% 53401 ± 24% sched_debug.cpu.clock.avg
195596 -72.7% 53442 ± 24% sched_debug.cpu.clock.max
195548 -72.7% 53352 ± 24% sched_debug.cpu.clock.min
194424 -72.6% 53229 ± 24% sched_debug.cpu.clock_task.avg
194608 -72.6% 53383 ± 24% sched_debug.cpu.clock_task.max
181834 -77.5% 40964 ± 31% sched_debug.cpu.clock_task.min
4241 ± 2% -80.6% 821.65 ±142% sched_debug.cpu.curr->pid.avg
9799 ± 2% -55.4% 4365 ± 17% sched_debug.cpu.curr->pid.max
1365 ± 10% -48.0% 709.44 ± 5% sched_debug.cpu.curr->pid.stddev
537665 ± 4% +31.2% 705318 ± 14% sched_debug.cpu.max_idle_balance_cost.max
3119 ± 56% +579.1% 21184 ± 39% sched_debug.cpu.max_idle_balance_cost.stddev
0.78 ± 2% -76.3% 0.18 ±135% sched_debug.cpu.nr_running.avg
25773 ± 5% -96.1% 1007 ± 41% sched_debug.cpu.nr_switches.avg
48669 ± 10% -76.5% 11448 ± 13% sched_debug.cpu.nr_switches.max
19006 ± 7% -98.6% 258.81 ± 64% sched_debug.cpu.nr_switches.min
4142 ± 8% -66.3% 1396 ± 17% sched_debug.cpu.nr_switches.stddev
0.07 ± 23% -92.9% 0.01 ± 41% sched_debug.cpu.nr_uninterruptible.avg
240.19 ± 16% -82.1% 42.94 ± 41% sched_debug.cpu.nr_uninterruptible.max
-77.92 -88.1% -9.25 sched_debug.cpu.nr_uninterruptible.min
37.87 ± 5% -85.8% 5.36 ± 13% sched_debug.cpu.nr_uninterruptible.stddev
195549 -72.7% 53356 ± 24% sched_debug.cpu_clk
194699 -73.0% 52506 ± 25% sched_debug.ktime
0.00 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.avg
0.17 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.max
0.01 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.stddev
196368 -72.4% 54191 ± 24% sched_debug.sched_clk
0.17 ±142% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
0.19 ± 34% -51.3% 0.09 ± 37% perf-sched.sch_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
0.14 ± 55% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
0.14 ± 73% -82.5% 0.03 ±168% perf-sched.sch_delay.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
0.11 ± 59% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
0.04 ±132% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
0.02 ± 31% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
0.00 ±223% +51950.0% 0.26 ±212% perf-sched.sch_delay.avg.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit
0.25 ± 59% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
0.12 ±145% -99.1% 0.00 ±141% perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
0.25 ± 41% -81.6% 0.05 ± 69% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
0.11 ± 59% -87.1% 0.01 ±198% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
0.40 ± 50% -97.8% 0.01 ± 30% perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
2.25 ±138% -99.6% 0.01 ± 7% perf-sched.sch_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
0.32 ±104% -97.3% 0.01 ± 38% perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
0.01 ± 12% -34.9% 0.01 ± 18% perf-sched.sch_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
0.01 ± 20% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
0.19 ±185% -95.6% 0.01 ± 44% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
0.07 ± 20% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
0.26 ± 17% -98.8% 0.00 ± 10% perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.03 ± 51% -69.7% 0.01 ± 67% perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
0.01 ± 55% +721.9% 0.10 ± 29% perf-sched.sch_delay.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.01 ±128% -83.6% 0.00 ± 20% perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
0.06 ± 31% +1921.5% 1.23 ±165% perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
1.00 ±151% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
25.45 ± 94% -98.6% 0.36 ± 61% perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
4.56 ± 67% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
3.55 ± 97% -98.9% 0.04 ±189% perf-sched.sch_delay.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
3.16 ± 78% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
0.30 ±159% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
0.03 ± 86% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
0.00 ±223% +3.2e+06% 15.79 ±259% perf-sched.sch_delay.max.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit
3.09 ± 45% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
3.51 ± 21% -86.1% 0.49 ± 72% perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
3.59 ± 11% -92.0% 0.29 ±165% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
1.60 ± 69% -95.7% 0.07 ±243% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
0.81 ± 43% -98.5% 0.01 ± 43% perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
1.02 ± 88% -98.1% 0.02 ± 47% perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
9.68 ± 32% -92.2% 0.76 ± 72% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
12.26 ±109% -92.9% 0.87 ±101% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.03 ±106% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
37.84 ± 47% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
4.68 ± 36% -99.8% 0.01 ± 65% perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.36 ±186% -96.3% 0.01 ± 90% perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
97903 ± 4% -38.3% 60433 ± 29% perf-sched.total_wait_and_delay.count.ms
3.97 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
302.41 ± 5% -27.4% 219.54 ± 14% perf-sched.wait_and_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
1.48 ± 6% -90.9% 0.14 ± 79% perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
327.16 ± 9% -46.6% 174.81 ± 24% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
369.37 ± 2% -75.3% 91.05 ± 35% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
0.96 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
187.66 +120.6% 413.97 ± 14% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1831 ± 9% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
6.17 ± 45% -79.7% 1.25 ±142% perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
40.50 ± 8% +245.7% 140.00 ± 23% perf-sched.wait_and_delay.count.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
13.17 ± 2% +624.4% 95.38 ± 19% perf-sched.wait_and_delay.count.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
73021 ± 3% -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
11323 ± 3% -75.9% 2725 ± 28% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1887 ± 45% -96.1% 73.88 ± 78% perf-sched.wait_and_delay.count.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
1238 -34.5% 811.25 ± 13% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
35.19 ± 57% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
20.79 ± 19% -95.9% 0.84 ± 93% perf-sched.wait_and_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
1240 ± 20% -14.4% 1062 ± 10% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
500.34 +31.2% 656.38 ± 39% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
58.83 ± 39% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
1237 ± 34% +151.7% 3114 ± 25% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
49.27 ±119% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
58.17 ±187% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
3.78 ± 5% -97.6% 0.09 ± 37% perf-sched.wait_time.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
2.99 ± 4% +15.4% 3.45 ± 10% perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
3.92 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
4.71 ± 8% -99.5% 0.02 ±170% perf-sched.wait_time.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
1.67 ± 20% -92.7% 0.12 ± 30% perf-sched.wait_time.avg.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
2.10 ± 27% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
0.01 ± 44% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
1.67 ± 21% -94.3% 0.10 ± 35% perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
0.04 ±133% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
2.30 ± 14% -95.5% 0.10 ± 42% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
2.00 ± 74% +2917.4% 60.44 ± 33% perf-sched.wait_time.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
29.19 ± 5% -38.5% 17.96 ± 28% perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
0.37 ± 30% +5524.5% 20.95 ± 30% perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
302.40 ± 5% -27.4% 219.53 ± 14% perf-sched.wait_time.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
1.40 ± 6% -92.7% 0.10 ± 18% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
0.72 ±220% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
326.84 ± 9% -46.6% 174.54 ± 24% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
369.18 ± 2% -75.3% 91.04 ± 35% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
0.89 ± 6% -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
187.58 +120.6% 413.77 ± 14% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
2.36 ± 29% +1759.6% 43.80 ± 33% perf-sched.wait_time.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.01 ±156% -97.9% 0.00 ±264% perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
340.69 ±135% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
535.09 ±128% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
22.04 ± 32% -98.4% 0.36 ± 61% perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
13.57 ± 17% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
13.54 ± 10% -99.7% 0.04 ±189% perf-sched.wait_time.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
10.17 ± 19% -95.2% 0.49 ± 56% perf-sched.wait_time.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
11.35 ± 25% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
0.01 ± 32% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
10.62 ± 9% -96.5% 0.38 ± 72% perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
0.20 ±199% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
14.42 ± 22% -96.6% 0.49 ± 72% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
4.00 ± 74% +19182.5% 772.23 ± 40% perf-sched.wait_time.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
10.75 ± 98% +6512.2% 710.88 ± 56% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
15.80 ± 8% -95.2% 0.76 ± 72% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
11.64 ± 61% -98.9% 0.13 ±132% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
2.94 ±213% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
1240 ± 20% -14.3% 1062 ± 10% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
500.11 +31.2% 656.37 ± 39% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
32.65 ± 33% -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
1237 ± 34% +151.6% 3113 ± 25% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
95.59 -95.6 0.00 perf-profile.calltrace.cycles-pp.__mmap
95.54 -95.5 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
95.54 -95.5 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap
94.54 -94.5 0.00 perf-profile.calltrace.cycles-pp.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
94.46 -94.0 0.41 ±138% perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
94.14 -93.7 0.40 ±136% perf-profile.calltrace.cycles-pp.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff
93.79 -93.5 0.31 ±134% perf-profile.calltrace.cycles-pp.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff
93.40 -93.4 0.00 perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma.__mmap_region
93.33 -93.3 0.00 perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma
93.44 -93.3 0.14 ±264% perf-profile.calltrace.cycles-pp.down_write.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap
94.45 -93.0 1.42 ± 60% perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
94.25 -92.9 1.33 ± 61% perf-profile.calltrace.cycles-pp.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
92.89 -92.9 0.00 perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file
0.00 +1.1 1.09 ± 33% perf-profile.calltrace.cycles-pp.dup_mmap.dup_mm.copy_process.kernel_clone.__do_sys_clone
0.00 +1.4 1.37 ± 49% perf-profile.calltrace.cycles-pp.setlocale
0.00 +1.6 1.64 ± 47% perf-profile.calltrace.cycles-pp.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry
0.00 +1.6 1.64 ± 47% perf-profile.calltrace.cycles-pp.handle_softirqs.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt
0.00 +1.6 1.65 ± 43% perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +1.8 1.76 ± 44% perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +1.9 1.93 ± 26% perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64
0.00 +2.2 2.16 ± 44% perf-profile.calltrace.cycles-pp.do_pte_missing.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
0.00 +2.2 2.23 ± 33% perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +2.4 2.37 ± 36% perf-profile.calltrace.cycles-pp.zap_present_ptes.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
0.00 +2.5 2.48 ± 32% perf-profile.calltrace.cycles-pp.get_cpu_sleep_time_us.get_idle_time.uptime_proc_show.seq_read_iter.vfs_read
0.00 +2.5 2.50 ± 45% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
0.00 +2.5 2.54 ± 47% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group
0.00 +2.5 2.54 ± 47% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
0.00 +2.6 2.62 ± 35% perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
0.00 +2.6 2.62 ± 35% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
0.00 +2.6 2.62 ± 35% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe._Fork
0.00 +2.6 2.62 ± 35% perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
0.00 +2.7 2.68 ± 35% perf-profile.calltrace.cycles-pp.get_idle_time.uptime_proc_show.seq_read_iter.vfs_read.ksys_read
0.00 +2.8 2.77 ± 33% perf-profile.calltrace.cycles-pp.uptime_proc_show.seq_read_iter.vfs_read.ksys_read.do_syscall_64
0.00 +2.8 2.82 ± 32% perf-profile.calltrace.cycles-pp._Fork
0.00 +2.8 2.84 ± 45% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
0.00 +2.8 2.84 ± 45% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
0.00 +2.9 2.89 ± 39% perf-profile.calltrace.cycles-pp.event_function_call.perf_event_release_kernel.perf_release.__fput.task_work_run
0.00 +2.9 2.89 ± 39% perf-profile.calltrace.cycles-pp.smp_call_function_single.event_function_call.perf_event_release_kernel.perf_release.__fput
0.00 +3.1 3.10 ± 64% perf-profile.calltrace.cycles-pp.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +3.1 3.10 ± 64% perf-profile.calltrace.cycles-pp.seq_read_iter.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64
0.00 +3.1 3.13 ± 33% perf-profile.calltrace.cycles-pp.asm_exc_page_fault
0.00 +3.2 3.18 ± 37% perf-profile.calltrace.cycles-pp.seq_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +3.2 3.20 ± 28% perf-profile.calltrace.cycles-pp.mutex_unlock.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release
0.00 +3.2 3.24 ± 39% perf-profile.calltrace.cycles-pp.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +3.2 3.24 ± 36% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +3.2 3.24 ± 36% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64
0.00 +3.2 3.24 ± 36% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +3.2 3.24 ± 36% perf-profile.calltrace.cycles-pp.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +3.8 3.85 ± 39% perf-profile.calltrace.cycles-pp.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
0.00 +3.8 3.85 ± 39% perf-profile.calltrace.cycles-pp.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
0.00 +3.8 3.85 ± 39% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
0.00 +3.8 3.85 ± 39% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.execve
0.00 +3.8 3.85 ± 39% perf-profile.calltrace.cycles-pp.execve
0.00 +4.0 4.04 ± 43% perf-profile.calltrace.cycles-pp.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +4.0 4.04 ± 43% perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64
0.00 +4.1 4.10 ± 30% perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.exit_mmap.__mmput.exit_mm
0.00 +4.2 4.18 ± 31% perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap.__mmput
0.00 +4.2 4.18 ± 31% perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap
0.00 +4.2 4.20 ± 28% perf-profile.calltrace.cycles-pp.unmap_vmas.exit_mmap.__mmput.exit_mm.do_exit
0.00 +4.2 4.25 ± 65% perf-profile.calltrace.cycles-pp.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64
0.00 +4.3 4.27 ± 26% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
0.00 +4.3 4.30 ± 22% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.get_signal
0.00 +4.3 4.30 ± 22% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
0.00 +4.5 4.46 ± 59% perf-profile.calltrace.cycles-pp.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +4.6 4.57 ± 58% perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen
0.00 +4.7 4.68 ± 55% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn
0.00 +4.7 4.68 ± 55% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn.perf_mmap__push
0.00 +4.7 4.68 ± 55% perf-profile.calltrace.cycles-pp.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record
0.00 +4.7 4.68 ± 55% perf-profile.calltrace.cycles-pp.write.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist
0.00 +4.7 4.68 ± 55% perf-profile.calltrace.cycles-pp.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record
0.00 +4.9 4.90 ± 57% perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
0.00 +4.9 4.92 ± 26% perf-profile.calltrace.cycles-pp.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release.__fput
0.00 +5.0 4.99 ±100% perf-profile.calltrace.cycles-pp.__intel_pmu_enable_all.perf_rotate_context.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt
0.00 +5.0 4.99 ±100% perf-profile.calltrace.cycles-pp.perf_rotate_context.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt
0.00 +5.1 5.08 ±102% perf-profile.calltrace.cycles-pp.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt
0.00 +5.1 5.14 ± 28% perf-profile.calltrace.cycles-pp.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin
0.00 +5.1 5.14 ± 28% perf-profile.calltrace.cycles-pp.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin.handle_internal_command
0.00 +5.4 5.43 ± 25% perf-profile.calltrace.cycles-pp._free_event.perf_event_release_kernel.perf_release.__fput.task_work_run
0.00 +5.8 5.82 ± 94% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry
0.00 +5.8 5.82 ± 94% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt
0.00 +6.1 6.07 ± 90% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
0.00 +6.6 6.62 ± 24% perf-profile.calltrace.cycles-pp.__cmd_record.cmd_record.run_builtin.handle_internal_command.main
0.00 +6.6 6.62 ± 24% perf-profile.calltrace.cycles-pp.cmd_record.run_builtin.handle_internal_command.main
0.00 +6.8 6.76 ± 18% perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit
0.00 +7.6 7.56 ± 76% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter
0.00 +8.0 8.03 ± 27% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
0.00 +8.0 8.03 ± 27% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
0.00 +8.0 8.05 ± 68% perf-profile.calltrace.cycles-pp.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter
0.00 +8.1 8.13 ± 28% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
0.00 +8.1 8.13 ± 28% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
0.00 +8.1 8.13 ± 28% perf-profile.calltrace.cycles-pp.read
0.00 +9.1 9.05 ± 35% perf-profile.calltrace.cycles-pp.handle_internal_command.main
0.00 +9.1 9.05 ± 35% perf-profile.calltrace.cycles-pp.main
0.00 +9.1 9.05 ± 35% perf-profile.calltrace.cycles-pp.run_builtin.handle_internal_command.main
0.00 +9.3 9.26 ± 30% perf-profile.calltrace.cycles-pp.perf_event_release_kernel.perf_release.__fput.task_work_run.do_exit
0.00 +9.3 9.26 ± 30% perf-profile.calltrace.cycles-pp.perf_release.__fput.task_work_run.do_exit.do_group_exit
0.00 +10.1 10.14 ± 28% perf-profile.calltrace.cycles-pp.__fput.task_work_run.do_exit.do_group_exit.get_signal
0.00 +10.2 10.23 ± 27% perf-profile.calltrace.cycles-pp.task_work_run.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
0.00 +11.0 10.98 ± 55% perf-profile.calltrace.cycles-pp.asm_sysvec_reschedule_ipi.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
0.00 +20.6 20.64 ± 30% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +20.6 20.64 ± 30% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
1.21 ± 3% +36.6 37.80 ± 12% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
1.21 ± 3% +36.6 37.80 ± 12% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.common_startup_64
1.22 ± 3% +36.8 38.00 ± 13% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.common_startup_64
1.22 ± 3% +36.9 38.10 ± 13% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.common_startup_64
1.22 ± 3% +36.9 38.10 ± 13% perf-profile.calltrace.cycles-pp.start_secondary.common_startup_64
1.21 ± 3% +37.2 38.43 ± 11% perf-profile.calltrace.cycles-pp.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
1.21 ± 3% +37.2 38.43 ± 11% perf-profile.calltrace.cycles-pp.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
1.21 ± 3% +37.3 38.54 ± 12% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
1.22 ± 3% +37.6 38.84 ± 12% perf-profile.calltrace.cycles-pp.common_startup_64
2.19 ± 3% +53.9 56.10 ± 19% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
95.60 -95.2 0.41 ±138% perf-profile.children.cycles-pp.__mmap
94.14 -93.7 0.49 ±130% perf-profile.children.cycles-pp.__mmap_new_vma
93.79 -93.5 0.31 ±134% perf-profile.children.cycles-pp.vma_link_file
93.40 -93.4 0.00 perf-profile.children.cycles-pp.rwsem_down_write_slowpath
93.33 -93.3 0.00 perf-profile.children.cycles-pp.rwsem_optimistic_spin
94.55 -93.1 1.42 ± 60% perf-profile.children.cycles-pp.ksys_mmap_pgoff
92.91 -92.9 0.00 perf-profile.children.cycles-pp.osq_lock
93.44 -92.7 0.75 ±109% perf-profile.children.cycles-pp.down_write
94.46 -92.6 1.84 ± 34% perf-profile.children.cycles-pp.vm_mmap_pgoff
94.45 -92.6 1.84 ± 34% perf-profile.children.cycles-pp.do_mmap
94.25 -92.6 1.66 ± 37% perf-profile.children.cycles-pp.__mmap_region
95.58 -44.8 50.78 ± 11% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
95.58 -44.8 50.78 ± 11% perf-profile.children.cycles-pp.do_syscall_64
0.00 +1.1 1.09 ± 33% perf-profile.children.cycles-pp.dup_mmap
0.00 +1.4 1.37 ± 49% perf-profile.children.cycles-pp.setlocale
0.00 +1.9 1.93 ± 26% perf-profile.children.cycles-pp.dup_mm
0.03 ± 70% +2.0 1.99 ± 36% perf-profile.children.cycles-pp.handle_softirqs
0.00 +2.0 1.99 ± 36% perf-profile.children.cycles-pp.__irq_exit_rcu
0.00 +2.0 2.02 ± 38% perf-profile.children.cycles-pp.folios_put_refs
0.00 +2.1 2.06 ± 52% perf-profile.children.cycles-pp._raw_spin_lock
0.00 +2.2 2.16 ± 44% perf-profile.children.cycles-pp.do_pte_missing
0.00 +2.2 2.21 ± 68% perf-profile.children.cycles-pp.link_path_walk
0.00 +2.2 2.23 ± 33% perf-profile.children.cycles-pp.copy_process
0.00 +2.3 2.30 ± 40% perf-profile.children.cycles-pp.__tlb_batch_free_encoded_pages
0.00 +2.3 2.30 ± 40% perf-profile.children.cycles-pp.free_pages_and_swap_cache
0.00 +2.3 2.34 ± 46% perf-profile.children.cycles-pp.walk_component
0.00 +2.4 2.37 ± 36% perf-profile.children.cycles-pp.zap_present_ptes
0.00 +2.5 2.48 ± 32% perf-profile.children.cycles-pp.get_cpu_sleep_time_us
0.00 +2.6 2.62 ± 35% perf-profile.children.cycles-pp.__do_sys_clone
0.00 +2.6 2.62 ± 35% perf-profile.children.cycles-pp.kernel_clone
0.00 +2.7 2.68 ± 35% perf-profile.children.cycles-pp.get_idle_time
0.00 +2.8 2.77 ± 33% perf-profile.children.cycles-pp.uptime_proc_show
0.00 +2.9 2.91 ± 32% perf-profile.children.cycles-pp._Fork
0.00 +3.1 3.10 ± 64% perf-profile.children.cycles-pp.proc_reg_read_iter
0.00 +3.2 3.24 ± 39% perf-profile.children.cycles-pp.bprm_execve
0.00 +3.2 3.24 ± 36% perf-profile.children.cycles-pp.__x64_sys_exit_group
0.00 +3.2 3.24 ± 36% perf-profile.children.cycles-pp.x64_sys_call
0.00 +3.8 3.85 ± 39% perf-profile.children.cycles-pp.__x64_sys_execve
0.00 +3.8 3.85 ± 39% perf-profile.children.cycles-pp.do_execveat_common
0.00 +3.8 3.85 ± 39% perf-profile.children.cycles-pp.execve
0.00 +4.0 3.99 ± 38% perf-profile.children.cycles-pp.mutex_unlock
0.00 +4.2 4.19 ± 31% perf-profile.children.cycles-pp.zap_pte_range
0.00 +4.2 4.25 ± 65% perf-profile.children.cycles-pp.generic_perform_write
0.00 +4.3 4.29 ± 29% perf-profile.children.cycles-pp.unmap_page_range
0.00 +4.3 4.29 ± 29% perf-profile.children.cycles-pp.zap_pmd_range
0.00 +4.3 4.31 ± 51% perf-profile.children.cycles-pp.do_filp_open
0.00 +4.3 4.31 ± 51% perf-profile.children.cycles-pp.path_openat
0.19 ± 23% +4.4 4.60 ± 26% perf-profile.children.cycles-pp.__handle_mm_fault
0.00 +4.5 4.46 ± 59% perf-profile.children.cycles-pp.shmem_file_write_iter
0.00 +4.5 4.55 ± 24% perf-profile.children.cycles-pp.event_function_call
0.00 +4.5 4.55 ± 24% perf-profile.children.cycles-pp.smp_call_function_single
0.00 +4.6 4.58 ± 30% perf-profile.children.cycles-pp.unmap_vmas
0.51 ± 6% +4.6 5.14 ± 24% perf-profile.children.cycles-pp.handle_mm_fault
0.00 +4.7 4.68 ± 55% perf-profile.children.cycles-pp.record__pushfn
0.00 +4.7 4.68 ± 55% perf-profile.children.cycles-pp.writen
0.00 +4.8 4.80 ± 48% perf-profile.children.cycles-pp.do_sys_openat2
0.77 ± 3% +4.8 5.59 ± 21% perf-profile.children.cycles-pp.exc_page_fault
0.76 ± 3% +4.8 5.59 ± 21% perf-profile.children.cycles-pp.do_user_addr_fault
0.00 +4.9 4.90 ± 57% perf-profile.children.cycles-pp.ksys_write
0.00 +4.9 4.90 ± 57% perf-profile.children.cycles-pp.vfs_write
0.00 +4.9 4.90 ± 48% perf-profile.children.cycles-pp.__x64_sys_openat
0.00 +4.9 4.92 ± 26% perf-profile.children.cycles-pp.sw_perf_event_destroy
0.00 +5.0 4.99 ±100% perf-profile.children.cycles-pp.perf_rotate_context
0.00 +5.0 5.01 ± 54% perf-profile.children.cycles-pp.write
0.00 +5.1 5.09 ±102% perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
0.00 +5.4 5.43 ± 25% perf-profile.children.cycles-pp._free_event
1.18 +5.6 6.78 ± 20% perf-profile.children.cycles-pp.asm_exc_page_fault
0.46 +5.6 6.07 ± 90% perf-profile.children.cycles-pp.__hrtimer_run_queues
0.00 +5.7 5.75 ± 39% perf-profile.children.cycles-pp.perf_mmap__push
0.00 +5.7 5.75 ± 39% perf-profile.children.cycles-pp.record__mmap_read_evlist
0.53 +5.8 6.28 ± 89% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
0.53 +5.8 6.28 ± 89% perf-profile.children.cycles-pp.hrtimer_interrupt
0.00 +6.6 6.65 ± 77% perf-profile.children.cycles-pp.__intel_pmu_enable_all
0.00 +6.8 6.85 ± 20% perf-profile.children.cycles-pp.exit_mm
0.58 ± 2% +7.6 8.14 ± 75% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
0.00 +7.7 7.67 ± 23% perf-profile.children.cycles-pp.exit_mmap
0.00 +7.7 7.67 ± 30% perf-profile.children.cycles-pp.seq_read_iter
0.00 +7.7 7.72 ± 80% perf-profile.children.cycles-pp.asm_sysvec_reschedule_ipi
0.00 +7.8 7.75 ± 23% perf-profile.children.cycles-pp.__mmput
0.00 +8.0 8.03 ± 27% perf-profile.children.cycles-pp.ksys_read
0.00 +8.0 8.03 ± 27% perf-profile.children.cycles-pp.vfs_read
0.00 +8.1 8.13 ± 28% perf-profile.children.cycles-pp.read
0.02 ±141% +9.0 9.05 ± 35% perf-profile.children.cycles-pp.__cmd_record
0.02 ±141% +9.0 9.05 ± 35% perf-profile.children.cycles-pp.cmd_record
0.02 ±141% +9.0 9.05 ± 35% perf-profile.children.cycles-pp.handle_internal_command
0.02 ±141% +9.0 9.05 ± 35% perf-profile.children.cycles-pp.main
0.02 ±141% +9.0 9.05 ± 35% perf-profile.children.cycles-pp.run_builtin
0.00 +9.3 9.26 ± 30% perf-profile.children.cycles-pp.perf_event_release_kernel
0.00 +9.3 9.26 ± 30% perf-profile.children.cycles-pp.perf_release
1.02 ± 4% +9.3 10.33 ± 27% perf-profile.children.cycles-pp.task_work_run
0.00 +11.0 11.05 ± 28% perf-profile.children.cycles-pp.__fput
0.00 +15.8 15.85 ± 25% perf-profile.children.cycles-pp.arch_do_signal_or_restart
0.00 +15.8 15.85 ± 25% perf-profile.children.cycles-pp.get_signal
0.00 +19.1 19.09 ± 19% perf-profile.children.cycles-pp.do_exit
0.00 +19.1 19.09 ± 19% perf-profile.children.cycles-pp.do_group_exit
1.70 ± 2% +30.7 32.41 ± 21% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
1.22 ± 3% +36.9 38.10 ± 13% perf-profile.children.cycles-pp.start_secondary
1.21 ± 3% +37.2 38.43 ± 11% perf-profile.children.cycles-pp.acpi_idle_do_entry
1.21 ± 3% +37.2 38.43 ± 11% perf-profile.children.cycles-pp.acpi_idle_enter
1.21 ± 3% +37.2 38.43 ± 11% perf-profile.children.cycles-pp.acpi_safe_halt
1.22 ± 3% +37.3 38.54 ± 12% perf-profile.children.cycles-pp.cpuidle_idle_call
1.21 ± 3% +37.3 38.54 ± 12% perf-profile.children.cycles-pp.cpuidle_enter
1.21 ± 3% +37.3 38.54 ± 12% perf-profile.children.cycles-pp.cpuidle_enter_state
1.22 ± 3% +37.6 38.84 ± 12% perf-profile.children.cycles-pp.common_startup_64
1.22 ± 3% +37.6 38.84 ± 12% perf-profile.children.cycles-pp.cpu_startup_entry
1.22 ± 3% +37.6 38.84 ± 12% perf-profile.children.cycles-pp.do_idle
92.37 -92.4 0.00 perf-profile.self.cycles-pp.osq_lock
0.00 +2.1 2.06 ± 52% perf-profile.self.cycles-pp._raw_spin_lock
0.00 +2.6 2.61 ± 36% perf-profile.self.cycles-pp.smp_call_function_single
0.00 +3.7 3.68 ± 37% perf-profile.self.cycles-pp.mutex_unlock
0.00 +6.6 6.65 ± 77% perf-profile.self.cycles-pp.__intel_pmu_enable_all
1.19 ± 3% +29.2 30.38 ± 15% perf-profile.self.cycles-pp.acpi_safe_halt
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-02-06 8:02 ` Oliver Sang
@ 2025-02-07 18:10 ` Yang Shi
2025-02-13 2:04 ` Oliver Sang
0 siblings, 1 reply; 35+ messages in thread
From: Yang Shi @ 2025-02-07 18:10 UTC (permalink / raw)
To: Oliver Sang
Cc: oe-lkp, lkp, linux-kernel, arnd, gregkh, Liam.Howlett,
lorenzo.stoakes, vbabka, jannh, willy, liushixin2, akpm,
linux-mm
On 2/6/25 12:02 AM, Oliver Sang wrote:
> hi, Yang Shi,
>
> On Fri, Jan 31, 2025 at 10:38:03AM -0800, Yang Shi wrote:
>>
>>
>> On 1/27/25 7:14 PM, kernel test robot wrote:
>>> hi, All,
>>>
>>> we don't have enough knowledge to understand fully the discussion for this
>>> patch, we saw "NACK" but there were more discussions later.
>>> so below report is just FYI what we observed in our tests. thanks
>> Thanks for the report. It was nack'ed because of the change to smaps/maps
>> files in proc.
>>
>>> Hello,
>>>
>>> kernel test robot noticed a 858.5% improvement of vm-scalability.throughput on:
>>>
>>>
>>> commit: 7143ee2391f1ea15e6791e129870473543634de2 ("[PATCH] /dev/zero: make private mapping full anonymous mapping")
>>> url: https://github.com/intel-lab-lkp/linux/commits/Yang-Shi/dev-zero-make-private-mapping-full-anonymous-mapping/20250114-063339
>>> base: https://git.kernel.org/cgit/linux/kernel/git/gregkh/char-misc.git a68d3cbfade64392507302f3a920113b60dc811f
>>> patch link: https://lore.kernel.org/all/20250113223033.4054534-1-yang@os.amperecomputing.com/
>>> patch subject: [PATCH] /dev/zero: make private mapping full anonymous mapping
>>>
>>> testcase: vm-scalability
>>> config: x86_64-rhel-9.4
>>> compiler: gcc-12
>>> test machine: 224 threads 4 sockets Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with 192G memory
>>> parameters:
>>>
>>> runtime: 300s
>>> test: small-allocs
>> It seems this benchmark allocates huge amount of small areas (each area is
>> as big as 40K) by mmap'ing /dev/zero.
>>
>> This patch makes /dev/zero mapping a full anonymous mapping, so the later
>> vma_link_file() is actually skipped, which needs acquire file rmap lock then
>> insert the mapping into file rmap tree. The below profiling also showed
>> this.
>> Quoted here so that we don't have to scroll down:
>>
>>> 95.60 -95.2 0.42 ±113% perf-profile.children.cycles-pp.__mmap
>>> 94.14 -93.6 0.54 ±106% perf-profile.children.cycles-pp.__mmap_new_vma
>>> 93.79 -93.6 0.21 ±171% perf-profile.children.cycles-pp.vma_link_file
>>> 93.40 -93.4 0.00 perf-profile.children.cycles-pp.rwsem_down_write_slowpath
>>> 93.33 -93.3 0.00 perf-profile.children.cycles-pp.rwsem_optimistic_spin
>>> 93.44 -93.2 0.24 ±178% perf-profile.children.cycles-pp.down_write
>>> 94.55 -93.1 1.40 ± 51% perf-profile.children.cycles-pp.ksys_mmap_pgoff
>>> 94.25 -93.0 1.30 ± 59% perf-profile.children.cycles-pp.__mmap_region
>>> 92.91 -92.9 0.00 perf-profile.children.cycles-pp.osq_lock
>>> 94.45 -92.7 1.72 ± 34% perf-profile.children.cycles-pp.do_mmap
>>> 94.46 -92.6 1.83 ± 31% perf-profile.children.cycles-pp.vm_mmap_pgoff
>> It significantly speed up mmap for this benchmark and the rmap lock
>> contention is reduced significantly for both multi-processes and
>> multi-threads.
>>
>> The benchmark itself may exaggerate the improvement, but it may really speed
>> up some real life workloads. For example, multiple applications which may
>> allocate anonymous mapping by mmap'ing /dev/zero, then they may have
>> contention on /dev/zero's rmap lock.
>>
>> It doesn't make too much sense to link /dev/zero anonymous vmas to the file
>> rmap tree. So the below patch should be able to speed up the benchmark too.
> sorry for late and thanks a lot for information!
>
>> Oliver, can you please give this patch a try?
> it seems this is an alternative patch?
Yes
> since we applied your "/dev/zero: make private mapping full anonymous mapping"
> patch upon a68d3cbfad like below:
>
> * 7143ee2391f1e /dev/zero: make private mapping full anonymous mapping
> * a68d3cbfade64 memstick: core: fix kernel-doc notation
>
> so I applied below patch also upon a68d3cbfad.
>
> we saw big improvement but not that big.
>
> =========================================================================================
> compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
> gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
>
> commit:
> a68d3cbfad ("memstick: core: fix kernel-doc notation")
> 52ec85cb99 <--- your patch
>
>
> a68d3cbfade64392 52ec85cb99e9b31dc304eae965a
> ---------------- ---------------------------
> %stddev %change %stddev
> \ | \
> 14364828 ± 4% +410.6% 73349239 ± 3% vm-scalability.throughput
>
> full comparison as below [1] just FYI.
Thanks for the update. I stared at the profiling report for a whole day,
but I didn't figure out where that 400% lost. I just saw the number of
page faults was fewer. And it seems like the reduction of page faults
match the 400% loss. So I did more trace and profiling.
The test case did the below stuff in a tight loop:
mmap 40K memory from /dev/zero (read only)
read the area
So two major factors to the performance: mmap and page fault. The
alternative patch did reduce the overhead of mmap to the same level as
the original patch.
The further perf profiling showed the cost of page fault is higher than
the original patch. But the profiling of page fault was interesting:
- 44.87% 0.01% usemem [kernel.kallsyms] [k]
do_translation_fault
- 44.86% do_translation_fault
- 44.83% do_page_fault
- 44.53% handle_mm_fault
9.04% __handle_mm_fault
Page fault consumed 40% of cpu time in handle_mm_fault, but
__handle_mm_fault just consumed 9%, I expected it should be the major
consumer.
So I annotated handle_mm_fault, then found the most time was consumed by
lru_gen_enter_fault() -> vma_has_recency() (my kernel has multi-gen LRU
enabled):
│ if (vma->vm_file && (vma->vm_file->f_mode & FMODE_NOREUSE))
│ ↓ cbz x1, b4
0.00 │ ldr w0, [x1, #12]
99.59 │ eor x0, x0, #0x800000
0.00 │ ubfx w0, w0, #23, #1
│ current->in_lru_fault = vma_has_recency(vma);
0.00 │ b4: ldrh w1, [x2, #1992]
0.01 │ bfi w1, w0, #5, #1
0.00 │ strh w1, [x2, #1992]
vma_has_recency() read vma->vm_file->f_mode if vma->vm_file is not NULL.
But that load took a long time. So I inspected struct file and saw:
struct file {
file_ref_t f_ref;
spinlock_t f_lock;
fmode_t f_mode;
const struct file_operations *f_op;
...
}
The f_mode is in the same cache line with f_ref (my kernel does NOT have
spin lock debug enabled). The test case mmap /dev/zero in a tight loop,
so the refcount is modified (fget/fput) very frequently, this resulted
in somehow false sharing.
So I tried the below patch on top of the alternative patch:
diff --git a/include/linux/mm_inline.h b/include/linux/mm_inline.h
index f9157a0c42a5..ba11dc0b1c7c 100644
--- a/include/linux/mm_inline.h
+++ b/include/linux/mm_inline.h
@@ -608,6 +608,9 @@ static inline bool vma_has_recency(struct
vm_area_struct *vma)
if (vma->vm_flags & (VM_SEQ_READ | VM_RAND_READ))
return false;
+ if (vma_is_anonymous(vma))
+ return true;
+
if (vma->vm_file && (vma->vm_file->f_mode & FMODE_NOREUSE))
return false;
This made the profiling of page fault look normal:
- 1.90% do_translation_fault
- 1.87% do_page_fault
- 1.49% handle_mm_fault
- 1.36% __handle_mm_fault
Please try this in your test.
But AFAICT I have never seen performance issue reported due to the false
sharing of refcount and other fields in struct file. This benchmark
stressed this quite badly.
>
>>
>> diff --git a/mm/vma.c b/mm/vma.c
>> index bb2119e5a0d0..1092222c40ae 100644
>> --- a/mm/vma.c
>> +++ b/mm/vma.c
>> @@ -1633,6 +1633,9 @@ static void unlink_file_vma_batch_process(struct
>> unlink_vma_file_batch *vb)
>> void unlink_file_vma_batch_add(struct unlink_vma_file_batch *vb,
>> struct vm_area_struct *vma)
>> {
>> + if (vma_is_anonymous(vma))
>> + return;
>> +
>> if (vma->vm_file == NULL)
>> return;
>>
>> @@ -1658,6 +1661,9 @@ void unlink_file_vma(struct vm_area_struct *vma)
>> {
>> struct file *file = vma->vm_file;
>>
>> + if (vma_is_anonymous(vma))
>> + return;
>> +
>> if (file) {
>> struct address_space *mapping = file->f_mapping;
>>
>> @@ -1672,6 +1678,9 @@ void vma_link_file(struct vm_area_struct *vma)
>> struct file *file = vma->vm_file;
>> struct address_space *mapping;
>>
>> + if (vma_is_anonymous(vma))
>> + return;
>> +
>> if (file) {
>> mapping = file->f_mapping;
>> i_mmap_lock_write(mapping);
>>
>>
>> Because /dev/zero's private mapping is an anonymous mapping with valid
>> vm_file, so we need to bail out early if the vma is anonymous even though it
>> has vm_file. IMHO, making /dev/zero private mapping a full anonymous mapping
>> looks more clean.
>>
> [1]
> =========================================================================================
> compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
> gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
>
> commit:
> a68d3cbfad ("memstick: core: fix kernel-doc notation")
> 52ec85cb99 <--- your patch
>
>
> a68d3cbfade64392 52ec85cb99e9b31dc304eae965a
> ---------------- ---------------------------
> %stddev %change %stddev
> \ | \
> 5.262e+09 ± 3% -45.0% 2.896e+09 ± 6% cpuidle..time
> 7924008 ± 3% -79.3% 1643339 ± 11% cpuidle..usage
> 1871164 ± 4% -22.4% 1452554 ± 12% numa-numastat.node3.local_node
> 1952164 ± 3% -20.1% 1560294 ± 12% numa-numastat.node3.numa_hit
> 399.52 -68.2% 126.86 uptime.boot
> 14507 -15.7% 12232 uptime.idle
> 6.99 ± 3% +147.9% 17.34 ± 4% vmstat.cpu.id
> 1.71 +473.6% 9.79 ± 2% vmstat.cpu.us
> 34204 ± 5% -72.9% 9272 ± 7% vmstat.system.cs
> 266575 -21.2% 210191 vmstat.system.in
> 3408 ± 5% -99.8% 8.38 ± 48% perf-c2c.DRAM.local
> 18076 ± 3% -99.8% 32.25 ± 27% perf-c2c.DRAM.remote
> 8082 ± 5% -99.8% 15.50 ± 64% perf-c2c.HITM.local
> 6544 ± 6% -99.8% 13.62 ± 51% perf-c2c.HITM.remote
> 14627 ± 4% -99.8% 29.12 ± 53% perf-c2c.HITM.total
> 6.49 ± 3% +8.8 15.24 ± 5% mpstat.cpu.all.idle%
> 0.63 -0.3 0.32 ± 4% mpstat.cpu.all.irq%
> 0.03 ± 2% +0.2 0.26 ± 2% mpstat.cpu.all.soft%
> 91.17 -17.0 74.15 mpstat.cpu.all.sys%
> 1.68 ± 2% +8.3 10.03 ± 2% mpstat.cpu.all.usr%
> 337.33 -97.4% 8.88 ± 75% mpstat.max_utilization.seconds
> 352.76 -77.3% 79.95 ± 2% time.elapsed_time
> 352.76 -77.3% 79.95 ± 2% time.elapsed_time.max
> 225965 ± 7% -16.0% 189844 ± 6% time.involuntary_context_switches
> 9.592e+08 ± 4% +11.9% 1.074e+09 time.minor_page_faults
> 20852 -8.8% 19012 time.percent_of_cpu_this_job_got
> 72302 -81.4% 13425 ± 3% time.system_time
> 1260 ± 3% +41.0% 1777 time.user_time
> 5393707 ± 5% -98.4% 86880 ± 17% time.voluntary_context_switches
> 1609925 -50.3% 800493 meminfo.Active
> 1609925 -50.3% 800493 meminfo.Active(anon)
> 160837 ± 33% -63.9% 58119 ± 13% meminfo.AnonHugePages
> 4435665 -18.5% 3614714 meminfo.Cached
> 1775547 -43.8% 998415 meminfo.Committed_AS
> 148539 -43.7% 83699 ± 4% meminfo.Mapped
> 4245538 ± 4% -20.9% 3356561 meminfo.PageTables
> 14166291 ± 4% -9.6% 12806082 meminfo.SUnreclaim
> 929777 -88.2% 109274 ± 3% meminfo.Shmem
> 14315492 ± 4% -9.6% 12947821 meminfo.Slab
> 64129 ± 4% +418.9% 332751 ± 3% vm-scalability.median
> 45.40 ± 5% +1961.8 2007 ± 8% vm-scalability.stddev%
> 14364828 ± 4% +410.6% 73349239 ± 3% vm-scalability.throughput
> 352.76 -77.3% 79.95 ± 2% vm-scalability.time.elapsed_time
> 352.76 -77.3% 79.95 ± 2% vm-scalability.time.elapsed_time.max
> 225965 ± 7% -16.0% 189844 ± 6% vm-scalability.time.involuntary_context_switches
> 9.592e+08 ± 4% +11.9% 1.074e+09 vm-scalability.time.minor_page_faults
> 20852 -8.8% 19012 vm-scalability.time.percent_of_cpu_this_job_got
> 72302 -81.4% 13425 ± 3% vm-scalability.time.system_time
> 1260 ± 3% +41.0% 1777 vm-scalability.time.user_time
> 5393707 ± 5% -98.4% 86880 ± 17% vm-scalability.time.voluntary_context_switches
> 4.316e+09 ± 4% +11.9% 4.832e+09 vm-scalability.workload
> 265763 ± 4% -20.5% 211398 ± 4% numa-vmstat.node0.nr_page_table_pages
> 31364 ±106% -85.0% 4690 ±169% numa-vmstat.node0.nr_shmem
> 12205 ± 67% -74.1% 3161 ±199% numa-vmstat.node1.nr_mapped
> 265546 ± 4% -21.8% 207742 ± 4% numa-vmstat.node1.nr_page_table_pages
> 44052 ± 71% -86.0% 6163 ±161% numa-vmstat.node1.nr_shmem
> 885590 ± 4% -9.9% 797649 ± 4% numa-vmstat.node1.nr_slab_unreclaimable
> 264589 ± 4% -21.2% 208598 ± 4% numa-vmstat.node2.nr_page_table_pages
> 881598 ± 4% -10.0% 793829 ± 4% numa-vmstat.node2.nr_slab_unreclaimable
> 192683 ± 30% -61.0% 75078 ± 70% numa-vmstat.node3.nr_active_anon
> 286819 ±108% -93.0% 19993 ± 39% numa-vmstat.node3.nr_file_pages
> 13124 ± 49% -92.3% 1006 ± 57% numa-vmstat.node3.nr_mapped
> 264499 ± 4% -22.1% 206135 ± 2% numa-vmstat.node3.nr_page_table_pages
> 139810 ± 14% -90.5% 13229 ± 89% numa-vmstat.node3.nr_shmem
> 880199 ± 4% -11.8% 776210 ± 5% numa-vmstat.node3.nr_slab_unreclaimable
> 192683 ± 30% -61.0% 75077 ± 70% numa-vmstat.node3.nr_zone_active_anon
> 1951359 ± 3% -20.1% 1558936 ± 12% numa-vmstat.node3.numa_hit
> 1870359 ± 4% -22.4% 1451195 ± 12% numa-vmstat.node3.numa_local
> 402515 -50.3% 200150 proc-vmstat.nr_active_anon
> 170568 +1.9% 173746 proc-vmstat.nr_anon_pages
> 4257257 +0.9% 4296664 proc-vmstat.nr_dirty_background_threshold
> 8524925 +0.9% 8603835 proc-vmstat.nr_dirty_threshold
> 1109246 -18.5% 903959 proc-vmstat.nr_file_pages
> 42815276 +0.9% 43210344 proc-vmstat.nr_free_pages
> 37525 -43.6% 21164 ± 4% proc-vmstat.nr_mapped
> 1059932 ± 4% -21.1% 836810 proc-vmstat.nr_page_table_pages
> 232507 -88.2% 27341 ± 3% proc-vmstat.nr_shmem
> 37297 -5.0% 35436 proc-vmstat.nr_slab_reclaimable
> 3537843 ± 4% -9.8% 3192506 proc-vmstat.nr_slab_unreclaimable
> 402515 -50.3% 200150 proc-vmstat.nr_zone_active_anon
> 61931 ± 8% -83.8% 10023 ± 45% proc-vmstat.numa_hint_faults
> 15755 ± 21% -87.1% 2039 ± 97% proc-vmstat.numa_hint_faults_local
> 6916516 ± 3% -7.1% 6425430 proc-vmstat.numa_hit
> 6568542 ± 3% -7.5% 6077764 proc-vmstat.numa_local
> 293942 ± 3% -69.6% 89435 ± 49% proc-vmstat.numa_pte_updates
> 9.608e+08 ± 4% +11.8% 1.074e+09 proc-vmstat.pgfault
> 55981 ± 2% -63.1% 20641 ± 2% proc-vmstat.pgreuse
> 1063552 ± 4% -20.3% 847673 ± 4% numa-meminfo.node0.PageTables
> 3565610 ± 4% -8.0% 3279375 ± 3% numa-meminfo.node0.SUnreclaim
> 125455 ±106% -85.2% 18620 ±168% numa-meminfo.node0.Shmem
> 3592377 ± 4% -7.1% 3336072 ± 4% numa-meminfo.node0.Slab
> 48482 ± 67% -74.3% 12475 ±199% numa-meminfo.node1.Mapped
> 1062709 ± 4% -21.7% 831966 ± 4% numa-meminfo.node1.PageTables
> 3543793 ± 4% -10.0% 3189589 ± 4% numa-meminfo.node1.SUnreclaim
> 176171 ± 71% -86.0% 24677 ±161% numa-meminfo.node1.Shmem
> 3593431 ± 4% -10.4% 3220352 ± 4% numa-meminfo.node1.Slab
> 1058901 ± 4% -21.3% 833124 ± 4% numa-meminfo.node2.PageTables
> 3527862 ± 4% -10.2% 3168666 ± 5% numa-meminfo.node2.SUnreclaim
> 3565750 ± 4% -10.3% 3200248 ± 5% numa-meminfo.node2.Slab
> 770405 ± 30% -61.0% 300435 ± 70% numa-meminfo.node3.Active
> 770405 ± 30% -61.0% 300435 ± 70% numa-meminfo.node3.Active(anon)
> 1146977 ±108% -93.0% 80110 ± 40% numa-meminfo.node3.FilePages
> 52663 ± 47% -91.6% 4397 ± 56% numa-meminfo.node3.Mapped
> 6368902 ± 20% -21.2% 5021246 ± 2% numa-meminfo.node3.MemUsed
> 1058539 ± 4% -22.2% 823061 ± 3% numa-meminfo.node3.PageTables
> 3522496 ± 4% -12.1% 3096728 ± 6% numa-meminfo.node3.SUnreclaim
> 558943 ± 14% -90.5% 53054 ± 89% numa-meminfo.node3.Shmem
> 3557392 ± 4% -12.3% 3119454 ± 6% numa-meminfo.node3.Slab
> 0.82 ± 4% -39.7% 0.50 ± 12% perf-stat.i.MPKI
> 2.714e+10 ± 2% +185.7% 7.755e+10 ± 6% perf-stat.i.branch-instructions
> 0.11 ± 3% +0.1 0.20 ± 5% perf-stat.i.branch-miss-rate%
> 24932893 +156.6% 63980942 ± 5% perf-stat.i.branch-misses
> 64.93 -10.1 54.87 ± 2% perf-stat.i.cache-miss-rate%
> 34508 ± 4% -61.4% 13315 ± 10% perf-stat.i.context-switches
> 7.67 -63.7% 2.79 ± 6% perf-stat.i.cpi
> 224605 +10.8% 248972 ± 4% perf-stat.i.cpu-clock
> 696.35 ± 2% -57.4% 296.79 ± 3% perf-stat.i.cpu-migrations
> 1.102e+11 +128.5% 2.518e+11 ± 6% perf-stat.i.instructions
> 0.14 +198.2% 0.42 ± 5% perf-stat.i.ipc
> 24.25 ± 3% +375.8% 115.36 ± 3% perf-stat.i.metric.K/sec
> 2722043 ± 3% +439.7% 14690226 ± 6% perf-stat.i.minor-faults
> 2722043 ± 3% +439.7% 14690226 ± 6% perf-stat.i.page-faults
> 224605 +10.8% 248972 ± 4% perf-stat.i.task-clock
> 0.81 ± 3% -52.5% 0.39 ± 14% perf-stat.overall.MPKI
> 0.09 -0.0 0.08 ± 2% perf-stat.overall.branch-miss-rate%
> 64.81 -6.4 58.40 perf-stat.overall.cache-miss-rate%
> 7.24 -56.3% 3.17 ± 3% perf-stat.overall.cpi
> 0.14 +129.0% 0.32 ± 3% perf-stat.overall.ipc
> 9012 ± 2% -57.5% 3827 perf-stat.overall.path-length
> 2.701e+10 ± 2% +159.6% 7.012e+10 ± 2% perf-stat.ps.branch-instructions
> 24708939 +119.2% 54173035 perf-stat.ps.branch-misses
> 34266 ± 5% -73.9% 8949 ± 7% perf-stat.ps.context-switches
> 7.941e+11 -9.1% 7.219e+11 perf-stat.ps.cpu-cycles
> 693.54 ± 2% -68.6% 217.73 ± 5% perf-stat.ps.cpu-migrations
> 1.097e+11 +108.1% 2.282e+11 ± 2% perf-stat.ps.instructions
> 2710577 ± 3% +388.7% 13246535 ± 2% perf-stat.ps.minor-faults
> 2710577 ± 3% +388.7% 13246536 ± 2% perf-stat.ps.page-faults
> 3.886e+13 ± 2% -52.4% 1.849e+13 perf-stat.total.instructions
> 64052898 ± 5% -96.2% 2460331 ±166% sched_debug.cfs_rq:/.avg_vruntime.avg
> 95701822 ± 7% -85.1% 14268127 ±116% sched_debug.cfs_rq:/.avg_vruntime.max
> 43098762 ± 6% -96.0% 1715136 ±173% sched_debug.cfs_rq:/.avg_vruntime.min
> 9223270 ± 9% -84.2% 1457904 ±122% sched_debug.cfs_rq:/.avg_vruntime.stddev
> 0.78 ± 2% -77.0% 0.18 ±130% sched_debug.cfs_rq:/.h_nr_running.avg
> 43049468 ± 22% -89.3% 4590302 ±180% sched_debug.cfs_rq:/.left_deadline.max
> 3836405 ± 37% -85.6% 550773 ±176% sched_debug.cfs_rq:/.left_deadline.stddev
> 43049467 ± 22% -89.3% 4590279 ±180% sched_debug.cfs_rq:/.left_vruntime.max
> 3836405 ± 37% -85.6% 550772 ±176% sched_debug.cfs_rq:/.left_vruntime.stddev
> 64052901 ± 5% -96.2% 2460341 ±166% sched_debug.cfs_rq:/.min_vruntime.avg
> 95701822 ± 7% -85.1% 14268127 ±116% sched_debug.cfs_rq:/.min_vruntime.max
> 43098762 ± 6% -96.0% 1715136 ±173% sched_debug.cfs_rq:/.min_vruntime.min
> 9223270 ± 9% -84.2% 1457902 ±122% sched_debug.cfs_rq:/.min_vruntime.stddev
> 0.77 ± 2% -77.4% 0.17 ±128% sched_debug.cfs_rq:/.nr_running.avg
> 1.61 ± 24% +396.0% 7.96 ± 62% sched_debug.cfs_rq:/.removed.runnable_avg.avg
> 86.69 +424.4% 454.62 ± 24% sched_debug.cfs_rq:/.removed.runnable_avg.max
> 11.14 ± 13% +409.8% 56.79 ± 35% sched_debug.cfs_rq:/.removed.runnable_avg.stddev
> 1.61 ± 24% +396.0% 7.96 ± 62% sched_debug.cfs_rq:/.removed.util_avg.avg
> 86.69 +424.4% 454.62 ± 24% sched_debug.cfs_rq:/.removed.util_avg.max
> 11.14 ± 13% +409.8% 56.79 ± 35% sched_debug.cfs_rq:/.removed.util_avg.stddev
> 43049467 ± 22% -89.3% 4590282 ±180% sched_debug.cfs_rq:/.right_vruntime.max
> 3836405 ± 37% -85.6% 550772 ±176% sched_debug.cfs_rq:/.right_vruntime.stddev
> 286633 ± 43% +262.3% 1038592 ± 36% sched_debug.cfs_rq:/.runnable_avg.avg
> 34728895 ± 30% +349.2% 1.56e+08 ± 26% sched_debug.cfs_rq:/.runnable_avg.max
> 2845573 ± 30% +325.9% 12119045 ± 26% sched_debug.cfs_rq:/.runnable_avg.stddev
> 769.03 -69.9% 231.86 ± 84% sched_debug.cfs_rq:/.util_avg.avg
> 1621 ± 5% -31.5% 1111 ± 8% sched_debug.cfs_rq:/.util_avg.max
> 724.17 ± 2% -89.6% 75.66 ±147% sched_debug.cfs_rq:/.util_est.avg
> 1360 ± 15% -39.2% 826.88 ± 37% sched_debug.cfs_rq:/.util_est.max
> 766944 ± 3% +18.1% 905901 sched_debug.cpu.avg_idle.avg
> 321459 ± 2% -35.6% 207172 ± 10% sched_debug.cpu.avg_idle.stddev
> 195573 -72.7% 53401 ± 24% sched_debug.cpu.clock.avg
> 195596 -72.7% 53442 ± 24% sched_debug.cpu.clock.max
> 195548 -72.7% 53352 ± 24% sched_debug.cpu.clock.min
> 194424 -72.6% 53229 ± 24% sched_debug.cpu.clock_task.avg
> 194608 -72.6% 53383 ± 24% sched_debug.cpu.clock_task.max
> 181834 -77.5% 40964 ± 31% sched_debug.cpu.clock_task.min
> 4241 ± 2% -80.6% 821.65 ±142% sched_debug.cpu.curr->pid.avg
> 9799 ± 2% -55.4% 4365 ± 17% sched_debug.cpu.curr->pid.max
> 1365 ± 10% -48.0% 709.44 ± 5% sched_debug.cpu.curr->pid.stddev
> 537665 ± 4% +31.2% 705318 ± 14% sched_debug.cpu.max_idle_balance_cost.max
> 3119 ± 56% +579.1% 21184 ± 39% sched_debug.cpu.max_idle_balance_cost.stddev
> 0.78 ± 2% -76.3% 0.18 ±135% sched_debug.cpu.nr_running.avg
> 25773 ± 5% -96.1% 1007 ± 41% sched_debug.cpu.nr_switches.avg
> 48669 ± 10% -76.5% 11448 ± 13% sched_debug.cpu.nr_switches.max
> 19006 ± 7% -98.6% 258.81 ± 64% sched_debug.cpu.nr_switches.min
> 4142 ± 8% -66.3% 1396 ± 17% sched_debug.cpu.nr_switches.stddev
> 0.07 ± 23% -92.9% 0.01 ± 41% sched_debug.cpu.nr_uninterruptible.avg
> 240.19 ± 16% -82.1% 42.94 ± 41% sched_debug.cpu.nr_uninterruptible.max
> -77.92 -88.1% -9.25 sched_debug.cpu.nr_uninterruptible.min
> 37.87 ± 5% -85.8% 5.36 ± 13% sched_debug.cpu.nr_uninterruptible.stddev
> 195549 -72.7% 53356 ± 24% sched_debug.cpu_clk
> 194699 -73.0% 52506 ± 25% sched_debug.ktime
> 0.00 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.avg
> 0.17 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.max
> 0.01 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.stddev
> 196368 -72.4% 54191 ± 24% sched_debug.sched_clk
> 0.17 ±142% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
> 0.19 ± 34% -51.3% 0.09 ± 37% perf-sched.sch_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 0.14 ± 55% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
> 0.14 ± 73% -82.5% 0.03 ±168% perf-sched.sch_delay.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
> 0.11 ± 59% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 0.04 ±132% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
> 0.02 ± 31% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
> 0.00 ±223% +51950.0% 0.26 ±212% perf-sched.sch_delay.avg.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit
> 0.25 ± 59% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
> 0.12 ±145% -99.1% 0.00 ±141% perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
> 0.25 ± 41% -81.6% 0.05 ± 69% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
> 0.11 ± 59% -87.1% 0.01 ±198% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
> 0.40 ± 50% -97.8% 0.01 ± 30% perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 2.25 ±138% -99.6% 0.01 ± 7% perf-sched.sch_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
> 0.32 ±104% -97.3% 0.01 ± 38% perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
> 0.01 ± 12% -34.9% 0.01 ± 18% perf-sched.sch_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 0.01 ± 20% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
> 0.19 ±185% -95.6% 0.01 ± 44% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 0.07 ± 20% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 0.26 ± 17% -98.8% 0.00 ± 10% perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
> 0.03 ± 51% -69.7% 0.01 ± 67% perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
> 0.01 ± 55% +721.9% 0.10 ± 29% perf-sched.sch_delay.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 0.01 ±128% -83.6% 0.00 ± 20% perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
> 0.06 ± 31% +1921.5% 1.23 ±165% perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 1.00 ±151% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
> 25.45 ± 94% -98.6% 0.36 ± 61% perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 4.56 ± 67% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
> 3.55 ± 97% -98.9% 0.04 ±189% perf-sched.sch_delay.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
> 3.16 ± 78% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 0.30 ±159% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
> 0.03 ± 86% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
> 0.00 ±223% +3.2e+06% 15.79 ±259% perf-sched.sch_delay.max.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit
> 3.09 ± 45% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
> 3.51 ± 21% -86.1% 0.49 ± 72% perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
> 3.59 ± 11% -92.0% 0.29 ±165% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
> 1.60 ± 69% -95.7% 0.07 ±243% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
> 0.81 ± 43% -98.5% 0.01 ± 43% perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 1.02 ± 88% -98.1% 0.02 ± 47% perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
> 9.68 ± 32% -92.2% 0.76 ± 72% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 12.26 ±109% -92.9% 0.87 ±101% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
> 0.03 ±106% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
> 37.84 ± 47% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 4.68 ± 36% -99.8% 0.01 ± 65% perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
> 0.36 ±186% -96.3% 0.01 ± 90% perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
> 97903 ± 4% -38.3% 60433 ± 29% perf-sched.total_wait_and_delay.count.ms
> 3.97 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 302.41 ± 5% -27.4% 219.54 ± 14% perf-sched.wait_and_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 1.48 ± 6% -90.9% 0.14 ± 79% perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 327.16 ± 9% -46.6% 174.81 ± 24% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 369.37 ± 2% -75.3% 91.05 ± 35% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 0.96 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 187.66 +120.6% 413.97 ± 14% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 1831 ± 9% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 6.17 ± 45% -79.7% 1.25 ±142% perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 40.50 ± 8% +245.7% 140.00 ± 23% perf-sched.wait_and_delay.count.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 13.17 ± 2% +624.4% 95.38 ± 19% perf-sched.wait_and_delay.count.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 73021 ± 3% -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 11323 ± 3% -75.9% 2725 ± 28% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 1887 ± 45% -96.1% 73.88 ± 78% perf-sched.wait_and_delay.count.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 1238 -34.5% 811.25 ± 13% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 35.19 ± 57% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 20.79 ± 19% -95.9% 0.84 ± 93% perf-sched.wait_and_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 1240 ± 20% -14.4% 1062 ± 10% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 500.34 +31.2% 656.38 ± 39% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 58.83 ± 39% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 1237 ± 34% +151.7% 3114 ± 25% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 49.27 ±119% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
> 58.17 ±187% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
> 3.78 ± 5% -97.6% 0.09 ± 37% perf-sched.wait_time.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 2.99 ± 4% +15.4% 3.45 ± 10% perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
> 3.92 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
> 4.71 ± 8% -99.5% 0.02 ±170% perf-sched.wait_time.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
> 1.67 ± 20% -92.7% 0.12 ± 30% perf-sched.wait_time.avg.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
> 2.10 ± 27% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 0.01 ± 44% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
> 1.67 ± 21% -94.3% 0.10 ± 35% perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 0.04 ±133% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
> 2.30 ± 14% -95.5% 0.10 ± 42% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
> 2.00 ± 74% +2917.4% 60.44 ± 33% perf-sched.wait_time.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
> 29.19 ± 5% -38.5% 17.96 ± 28% perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
> 0.37 ± 30% +5524.5% 20.95 ± 30% perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 302.40 ± 5% -27.4% 219.53 ± 14% perf-sched.wait_time.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 1.40 ± 6% -92.7% 0.10 ± 18% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 0.72 ±220% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
> 326.84 ± 9% -46.6% 174.54 ± 24% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 369.18 ± 2% -75.3% 91.04 ± 35% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 0.89 ± 6% -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 187.58 +120.6% 413.77 ± 14% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 2.36 ± 29% +1759.6% 43.80 ± 33% perf-sched.wait_time.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 0.01 ±156% -97.9% 0.00 ±264% perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
> 340.69 ±135% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
> 535.09 ±128% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
> 22.04 ± 32% -98.4% 0.36 ± 61% perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 13.57 ± 17% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
> 13.54 ± 10% -99.7% 0.04 ±189% perf-sched.wait_time.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
> 10.17 ± 19% -95.2% 0.49 ± 56% perf-sched.wait_time.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
> 11.35 ± 25% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 0.01 ± 32% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
> 10.62 ± 9% -96.5% 0.38 ± 72% perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 0.20 ±199% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
> 14.42 ± 22% -96.6% 0.49 ± 72% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
> 4.00 ± 74% +19182.5% 772.23 ± 40% perf-sched.wait_time.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
> 10.75 ± 98% +6512.2% 710.88 ± 56% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 15.80 ± 8% -95.2% 0.76 ± 72% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 11.64 ± 61% -98.9% 0.13 ±132% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
> 2.94 ±213% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
> 1240 ± 20% -14.3% 1062 ± 10% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 500.11 +31.2% 656.37 ± 39% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 32.65 ± 33% -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 1237 ± 34% +151.6% 3113 ± 25% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 95.59 -95.6 0.00 perf-profile.calltrace.cycles-pp.__mmap
> 95.54 -95.5 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
> 95.54 -95.5 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap
> 94.54 -94.5 0.00 perf-profile.calltrace.cycles-pp.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
> 94.46 -94.0 0.41 ±138% perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
> 94.14 -93.7 0.40 ±136% perf-profile.calltrace.cycles-pp.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff
> 93.79 -93.5 0.31 ±134% perf-profile.calltrace.cycles-pp.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff
> 93.40 -93.4 0.00 perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 93.33 -93.3 0.00 perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma
> 93.44 -93.3 0.14 ±264% perf-profile.calltrace.cycles-pp.down_write.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap
> 94.45 -93.0 1.42 ± 60% perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 94.25 -92.9 1.33 ± 61% perf-profile.calltrace.cycles-pp.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 92.89 -92.9 0.00 perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file
> 0.00 +1.1 1.09 ± 33% perf-profile.calltrace.cycles-pp.dup_mmap.dup_mm.copy_process.kernel_clone.__do_sys_clone
> 0.00 +1.4 1.37 ± 49% perf-profile.calltrace.cycles-pp.setlocale
> 0.00 +1.6 1.64 ± 47% perf-profile.calltrace.cycles-pp.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry
> 0.00 +1.6 1.64 ± 47% perf-profile.calltrace.cycles-pp.handle_softirqs.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt
> 0.00 +1.6 1.65 ± 43% perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +1.8 1.76 ± 44% perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +1.9 1.93 ± 26% perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64
> 0.00 +2.2 2.16 ± 44% perf-profile.calltrace.cycles-pp.do_pte_missing.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
> 0.00 +2.2 2.23 ± 33% perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +2.4 2.37 ± 36% perf-profile.calltrace.cycles-pp.zap_present_ptes.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
> 0.00 +2.5 2.48 ± 32% perf-profile.calltrace.cycles-pp.get_cpu_sleep_time_us.get_idle_time.uptime_proc_show.seq_read_iter.vfs_read
> 0.00 +2.5 2.50 ± 45% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
> 0.00 +2.5 2.54 ± 47% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group
> 0.00 +2.5 2.54 ± 47% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
> 0.00 +2.6 2.62 ± 35% perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
> 0.00 +2.6 2.62 ± 35% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
> 0.00 +2.6 2.62 ± 35% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe._Fork
> 0.00 +2.6 2.62 ± 35% perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
> 0.00 +2.7 2.68 ± 35% perf-profile.calltrace.cycles-pp.get_idle_time.uptime_proc_show.seq_read_iter.vfs_read.ksys_read
> 0.00 +2.8 2.77 ± 33% perf-profile.calltrace.cycles-pp.uptime_proc_show.seq_read_iter.vfs_read.ksys_read.do_syscall_64
> 0.00 +2.8 2.82 ± 32% perf-profile.calltrace.cycles-pp._Fork
> 0.00 +2.8 2.84 ± 45% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
> 0.00 +2.8 2.84 ± 45% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
> 0.00 +2.9 2.89 ± 39% perf-profile.calltrace.cycles-pp.event_function_call.perf_event_release_kernel.perf_release.__fput.task_work_run
> 0.00 +2.9 2.89 ± 39% perf-profile.calltrace.cycles-pp.smp_call_function_single.event_function_call.perf_event_release_kernel.perf_release.__fput
> 0.00 +3.1 3.10 ± 64% perf-profile.calltrace.cycles-pp.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +3.1 3.10 ± 64% perf-profile.calltrace.cycles-pp.seq_read_iter.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64
> 0.00 +3.1 3.13 ± 33% perf-profile.calltrace.cycles-pp.asm_exc_page_fault
> 0.00 +3.2 3.18 ± 37% perf-profile.calltrace.cycles-pp.seq_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +3.2 3.20 ± 28% perf-profile.calltrace.cycles-pp.mutex_unlock.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release
> 0.00 +3.2 3.24 ± 39% perf-profile.calltrace.cycles-pp.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +3.2 3.24 ± 36% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +3.2 3.24 ± 36% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64
> 0.00 +3.2 3.24 ± 36% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +3.2 3.24 ± 36% perf-profile.calltrace.cycles-pp.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +3.8 3.85 ± 39% perf-profile.calltrace.cycles-pp.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> 0.00 +3.8 3.85 ± 39% perf-profile.calltrace.cycles-pp.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> 0.00 +3.8 3.85 ± 39% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> 0.00 +3.8 3.85 ± 39% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.execve
> 0.00 +3.8 3.85 ± 39% perf-profile.calltrace.cycles-pp.execve
> 0.00 +4.0 4.04 ± 43% perf-profile.calltrace.cycles-pp.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +4.0 4.04 ± 43% perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64
> 0.00 +4.1 4.10 ± 30% perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.exit_mmap.__mmput.exit_mm
> 0.00 +4.2 4.18 ± 31% perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap.__mmput
> 0.00 +4.2 4.18 ± 31% perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap
> 0.00 +4.2 4.20 ± 28% perf-profile.calltrace.cycles-pp.unmap_vmas.exit_mmap.__mmput.exit_mm.do_exit
> 0.00 +4.2 4.25 ± 65% perf-profile.calltrace.cycles-pp.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64
> 0.00 +4.3 4.27 ± 26% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
> 0.00 +4.3 4.30 ± 22% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.get_signal
> 0.00 +4.3 4.30 ± 22% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
> 0.00 +4.5 4.46 ± 59% perf-profile.calltrace.cycles-pp.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +4.6 4.57 ± 58% perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen
> 0.00 +4.7 4.68 ± 55% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn
> 0.00 +4.7 4.68 ± 55% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn.perf_mmap__push
> 0.00 +4.7 4.68 ± 55% perf-profile.calltrace.cycles-pp.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record
> 0.00 +4.7 4.68 ± 55% perf-profile.calltrace.cycles-pp.write.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist
> 0.00 +4.7 4.68 ± 55% perf-profile.calltrace.cycles-pp.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record
> 0.00 +4.9 4.90 ± 57% perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
> 0.00 +4.9 4.92 ± 26% perf-profile.calltrace.cycles-pp.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release.__fput
> 0.00 +5.0 4.99 ±100% perf-profile.calltrace.cycles-pp.__intel_pmu_enable_all.perf_rotate_context.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt
> 0.00 +5.0 4.99 ±100% perf-profile.calltrace.cycles-pp.perf_rotate_context.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt
> 0.00 +5.1 5.08 ±102% perf-profile.calltrace.cycles-pp.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt
> 0.00 +5.1 5.14 ± 28% perf-profile.calltrace.cycles-pp.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin
> 0.00 +5.1 5.14 ± 28% perf-profile.calltrace.cycles-pp.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin.handle_internal_command
> 0.00 +5.4 5.43 ± 25% perf-profile.calltrace.cycles-pp._free_event.perf_event_release_kernel.perf_release.__fput.task_work_run
> 0.00 +5.8 5.82 ± 94% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry
> 0.00 +5.8 5.82 ± 94% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt
> 0.00 +6.1 6.07 ± 90% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
> 0.00 +6.6 6.62 ± 24% perf-profile.calltrace.cycles-pp.__cmd_record.cmd_record.run_builtin.handle_internal_command.main
> 0.00 +6.6 6.62 ± 24% perf-profile.calltrace.cycles-pp.cmd_record.run_builtin.handle_internal_command.main
> 0.00 +6.8 6.76 ± 18% perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit
> 0.00 +7.6 7.56 ± 76% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter
> 0.00 +8.0 8.03 ± 27% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
> 0.00 +8.0 8.03 ± 27% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
> 0.00 +8.0 8.05 ± 68% perf-profile.calltrace.cycles-pp.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter
> 0.00 +8.1 8.13 ± 28% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
> 0.00 +8.1 8.13 ± 28% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
> 0.00 +8.1 8.13 ± 28% perf-profile.calltrace.cycles-pp.read
> 0.00 +9.1 9.05 ± 35% perf-profile.calltrace.cycles-pp.handle_internal_command.main
> 0.00 +9.1 9.05 ± 35% perf-profile.calltrace.cycles-pp.main
> 0.00 +9.1 9.05 ± 35% perf-profile.calltrace.cycles-pp.run_builtin.handle_internal_command.main
> 0.00 +9.3 9.26 ± 30% perf-profile.calltrace.cycles-pp.perf_event_release_kernel.perf_release.__fput.task_work_run.do_exit
> 0.00 +9.3 9.26 ± 30% perf-profile.calltrace.cycles-pp.perf_release.__fput.task_work_run.do_exit.do_group_exit
> 0.00 +10.1 10.14 ± 28% perf-profile.calltrace.cycles-pp.__fput.task_work_run.do_exit.do_group_exit.get_signal
> 0.00 +10.2 10.23 ± 27% perf-profile.calltrace.cycles-pp.task_work_run.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
> 0.00 +11.0 10.98 ± 55% perf-profile.calltrace.cycles-pp.asm_sysvec_reschedule_ipi.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
> 0.00 +20.6 20.64 ± 30% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +20.6 20.64 ± 30% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
> 1.21 ± 3% +36.6 37.80 ± 12% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
> 1.21 ± 3% +36.6 37.80 ± 12% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.common_startup_64
> 1.22 ± 3% +36.8 38.00 ± 13% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.common_startup_64
> 1.22 ± 3% +36.9 38.10 ± 13% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.common_startup_64
> 1.22 ± 3% +36.9 38.10 ± 13% perf-profile.calltrace.cycles-pp.start_secondary.common_startup_64
> 1.21 ± 3% +37.2 38.43 ± 11% perf-profile.calltrace.cycles-pp.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
> 1.21 ± 3% +37.2 38.43 ± 11% perf-profile.calltrace.cycles-pp.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
> 1.21 ± 3% +37.3 38.54 ± 12% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
> 1.22 ± 3% +37.6 38.84 ± 12% perf-profile.calltrace.cycles-pp.common_startup_64
> 2.19 ± 3% +53.9 56.10 ± 19% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
> 95.60 -95.2 0.41 ±138% perf-profile.children.cycles-pp.__mmap
> 94.14 -93.7 0.49 ±130% perf-profile.children.cycles-pp.__mmap_new_vma
> 93.79 -93.5 0.31 ±134% perf-profile.children.cycles-pp.vma_link_file
> 93.40 -93.4 0.00 perf-profile.children.cycles-pp.rwsem_down_write_slowpath
> 93.33 -93.3 0.00 perf-profile.children.cycles-pp.rwsem_optimistic_spin
> 94.55 -93.1 1.42 ± 60% perf-profile.children.cycles-pp.ksys_mmap_pgoff
> 92.91 -92.9 0.00 perf-profile.children.cycles-pp.osq_lock
> 93.44 -92.7 0.75 ±109% perf-profile.children.cycles-pp.down_write
> 94.46 -92.6 1.84 ± 34% perf-profile.children.cycles-pp.vm_mmap_pgoff
> 94.45 -92.6 1.84 ± 34% perf-profile.children.cycles-pp.do_mmap
> 94.25 -92.6 1.66 ± 37% perf-profile.children.cycles-pp.__mmap_region
> 95.58 -44.8 50.78 ± 11% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
> 95.58 -44.8 50.78 ± 11% perf-profile.children.cycles-pp.do_syscall_64
> 0.00 +1.1 1.09 ± 33% perf-profile.children.cycles-pp.dup_mmap
> 0.00 +1.4 1.37 ± 49% perf-profile.children.cycles-pp.setlocale
> 0.00 +1.9 1.93 ± 26% perf-profile.children.cycles-pp.dup_mm
> 0.03 ± 70% +2.0 1.99 ± 36% perf-profile.children.cycles-pp.handle_softirqs
> 0.00 +2.0 1.99 ± 36% perf-profile.children.cycles-pp.__irq_exit_rcu
> 0.00 +2.0 2.02 ± 38% perf-profile.children.cycles-pp.folios_put_refs
> 0.00 +2.1 2.06 ± 52% perf-profile.children.cycles-pp._raw_spin_lock
> 0.00 +2.2 2.16 ± 44% perf-profile.children.cycles-pp.do_pte_missing
> 0.00 +2.2 2.21 ± 68% perf-profile.children.cycles-pp.link_path_walk
> 0.00 +2.2 2.23 ± 33% perf-profile.children.cycles-pp.copy_process
> 0.00 +2.3 2.30 ± 40% perf-profile.children.cycles-pp.__tlb_batch_free_encoded_pages
> 0.00 +2.3 2.30 ± 40% perf-profile.children.cycles-pp.free_pages_and_swap_cache
> 0.00 +2.3 2.34 ± 46% perf-profile.children.cycles-pp.walk_component
> 0.00 +2.4 2.37 ± 36% perf-profile.children.cycles-pp.zap_present_ptes
> 0.00 +2.5 2.48 ± 32% perf-profile.children.cycles-pp.get_cpu_sleep_time_us
> 0.00 +2.6 2.62 ± 35% perf-profile.children.cycles-pp.__do_sys_clone
> 0.00 +2.6 2.62 ± 35% perf-profile.children.cycles-pp.kernel_clone
> 0.00 +2.7 2.68 ± 35% perf-profile.children.cycles-pp.get_idle_time
> 0.00 +2.8 2.77 ± 33% perf-profile.children.cycles-pp.uptime_proc_show
> 0.00 +2.9 2.91 ± 32% perf-profile.children.cycles-pp._Fork
> 0.00 +3.1 3.10 ± 64% perf-profile.children.cycles-pp.proc_reg_read_iter
> 0.00 +3.2 3.24 ± 39% perf-profile.children.cycles-pp.bprm_execve
> 0.00 +3.2 3.24 ± 36% perf-profile.children.cycles-pp.__x64_sys_exit_group
> 0.00 +3.2 3.24 ± 36% perf-profile.children.cycles-pp.x64_sys_call
> 0.00 +3.8 3.85 ± 39% perf-profile.children.cycles-pp.__x64_sys_execve
> 0.00 +3.8 3.85 ± 39% perf-profile.children.cycles-pp.do_execveat_common
> 0.00 +3.8 3.85 ± 39% perf-profile.children.cycles-pp.execve
> 0.00 +4.0 3.99 ± 38% perf-profile.children.cycles-pp.mutex_unlock
> 0.00 +4.2 4.19 ± 31% perf-profile.children.cycles-pp.zap_pte_range
> 0.00 +4.2 4.25 ± 65% perf-profile.children.cycles-pp.generic_perform_write
> 0.00 +4.3 4.29 ± 29% perf-profile.children.cycles-pp.unmap_page_range
> 0.00 +4.3 4.29 ± 29% perf-profile.children.cycles-pp.zap_pmd_range
> 0.00 +4.3 4.31 ± 51% perf-profile.children.cycles-pp.do_filp_open
> 0.00 +4.3 4.31 ± 51% perf-profile.children.cycles-pp.path_openat
> 0.19 ± 23% +4.4 4.60 ± 26% perf-profile.children.cycles-pp.__handle_mm_fault
> 0.00 +4.5 4.46 ± 59% perf-profile.children.cycles-pp.shmem_file_write_iter
> 0.00 +4.5 4.55 ± 24% perf-profile.children.cycles-pp.event_function_call
> 0.00 +4.5 4.55 ± 24% perf-profile.children.cycles-pp.smp_call_function_single
> 0.00 +4.6 4.58 ± 30% perf-profile.children.cycles-pp.unmap_vmas
> 0.51 ± 6% +4.6 5.14 ± 24% perf-profile.children.cycles-pp.handle_mm_fault
> 0.00 +4.7 4.68 ± 55% perf-profile.children.cycles-pp.record__pushfn
> 0.00 +4.7 4.68 ± 55% perf-profile.children.cycles-pp.writen
> 0.00 +4.8 4.80 ± 48% perf-profile.children.cycles-pp.do_sys_openat2
> 0.77 ± 3% +4.8 5.59 ± 21% perf-profile.children.cycles-pp.exc_page_fault
> 0.76 ± 3% +4.8 5.59 ± 21% perf-profile.children.cycles-pp.do_user_addr_fault
> 0.00 +4.9 4.90 ± 57% perf-profile.children.cycles-pp.ksys_write
> 0.00 +4.9 4.90 ± 57% perf-profile.children.cycles-pp.vfs_write
> 0.00 +4.9 4.90 ± 48% perf-profile.children.cycles-pp.__x64_sys_openat
> 0.00 +4.9 4.92 ± 26% perf-profile.children.cycles-pp.sw_perf_event_destroy
> 0.00 +5.0 4.99 ±100% perf-profile.children.cycles-pp.perf_rotate_context
> 0.00 +5.0 5.01 ± 54% perf-profile.children.cycles-pp.write
> 0.00 +5.1 5.09 ±102% perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
> 0.00 +5.4 5.43 ± 25% perf-profile.children.cycles-pp._free_event
> 1.18 +5.6 6.78 ± 20% perf-profile.children.cycles-pp.asm_exc_page_fault
> 0.46 +5.6 6.07 ± 90% perf-profile.children.cycles-pp.__hrtimer_run_queues
> 0.00 +5.7 5.75 ± 39% perf-profile.children.cycles-pp.perf_mmap__push
> 0.00 +5.7 5.75 ± 39% perf-profile.children.cycles-pp.record__mmap_read_evlist
> 0.53 +5.8 6.28 ± 89% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
> 0.53 +5.8 6.28 ± 89% perf-profile.children.cycles-pp.hrtimer_interrupt
> 0.00 +6.6 6.65 ± 77% perf-profile.children.cycles-pp.__intel_pmu_enable_all
> 0.00 +6.8 6.85 ± 20% perf-profile.children.cycles-pp.exit_mm
> 0.58 ± 2% +7.6 8.14 ± 75% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
> 0.00 +7.7 7.67 ± 23% perf-profile.children.cycles-pp.exit_mmap
> 0.00 +7.7 7.67 ± 30% perf-profile.children.cycles-pp.seq_read_iter
> 0.00 +7.7 7.72 ± 80% perf-profile.children.cycles-pp.asm_sysvec_reschedule_ipi
> 0.00 +7.8 7.75 ± 23% perf-profile.children.cycles-pp.__mmput
> 0.00 +8.0 8.03 ± 27% perf-profile.children.cycles-pp.ksys_read
> 0.00 +8.0 8.03 ± 27% perf-profile.children.cycles-pp.vfs_read
> 0.00 +8.1 8.13 ± 28% perf-profile.children.cycles-pp.read
> 0.02 ±141% +9.0 9.05 ± 35% perf-profile.children.cycles-pp.__cmd_record
> 0.02 ±141% +9.0 9.05 ± 35% perf-profile.children.cycles-pp.cmd_record
> 0.02 ±141% +9.0 9.05 ± 35% perf-profile.children.cycles-pp.handle_internal_command
> 0.02 ±141% +9.0 9.05 ± 35% perf-profile.children.cycles-pp.main
> 0.02 ±141% +9.0 9.05 ± 35% perf-profile.children.cycles-pp.run_builtin
> 0.00 +9.3 9.26 ± 30% perf-profile.children.cycles-pp.perf_event_release_kernel
> 0.00 +9.3 9.26 ± 30% perf-profile.children.cycles-pp.perf_release
> 1.02 ± 4% +9.3 10.33 ± 27% perf-profile.children.cycles-pp.task_work_run
> 0.00 +11.0 11.05 ± 28% perf-profile.children.cycles-pp.__fput
> 0.00 +15.8 15.85 ± 25% perf-profile.children.cycles-pp.arch_do_signal_or_restart
> 0.00 +15.8 15.85 ± 25% perf-profile.children.cycles-pp.get_signal
> 0.00 +19.1 19.09 ± 19% perf-profile.children.cycles-pp.do_exit
> 0.00 +19.1 19.09 ± 19% perf-profile.children.cycles-pp.do_group_exit
> 1.70 ± 2% +30.7 32.41 ± 21% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
> 1.22 ± 3% +36.9 38.10 ± 13% perf-profile.children.cycles-pp.start_secondary
> 1.21 ± 3% +37.2 38.43 ± 11% perf-profile.children.cycles-pp.acpi_idle_do_entry
> 1.21 ± 3% +37.2 38.43 ± 11% perf-profile.children.cycles-pp.acpi_idle_enter
> 1.21 ± 3% +37.2 38.43 ± 11% perf-profile.children.cycles-pp.acpi_safe_halt
> 1.22 ± 3% +37.3 38.54 ± 12% perf-profile.children.cycles-pp.cpuidle_idle_call
> 1.21 ± 3% +37.3 38.54 ± 12% perf-profile.children.cycles-pp.cpuidle_enter
> 1.21 ± 3% +37.3 38.54 ± 12% perf-profile.children.cycles-pp.cpuidle_enter_state
> 1.22 ± 3% +37.6 38.84 ± 12% perf-profile.children.cycles-pp.common_startup_64
> 1.22 ± 3% +37.6 38.84 ± 12% perf-profile.children.cycles-pp.cpu_startup_entry
> 1.22 ± 3% +37.6 38.84 ± 12% perf-profile.children.cycles-pp.do_idle
> 92.37 -92.4 0.00 perf-profile.self.cycles-pp.osq_lock
> 0.00 +2.1 2.06 ± 52% perf-profile.self.cycles-pp._raw_spin_lock
> 0.00 +2.6 2.61 ± 36% perf-profile.self.cycles-pp.smp_call_function_single
> 0.00 +3.7 3.68 ± 37% perf-profile.self.cycles-pp.mutex_unlock
> 0.00 +6.6 6.65 ± 77% perf-profile.self.cycles-pp.__intel_pmu_enable_all
> 1.19 ± 3% +29.2 30.38 ± 15% perf-profile.self.cycles-pp.acpi_safe_halt
>
>
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-02-07 18:10 ` Yang Shi
@ 2025-02-13 2:04 ` Oliver Sang
2025-02-14 22:53 ` Yang Shi
0 siblings, 1 reply; 35+ messages in thread
From: Oliver Sang @ 2025-02-13 2:04 UTC (permalink / raw)
To: Yang Shi
Cc: oe-lkp, lkp, linux-kernel, arnd, gregkh, Liam.Howlett,
lorenzo.stoakes, vbabka, jannh, willy, liushixin2, akpm,
linux-mm, oliver.sang
hi, Yang Shi,
On Fri, Feb 07, 2025 at 10:10:37AM -0800, Yang Shi wrote:
>
> On 2/6/25 12:02 AM, Oliver Sang wrote:
[...]
>
> > since we applied your "/dev/zero: make private mapping full anonymous mapping"
> > patch upon a68d3cbfad like below:
> >
> > * 7143ee2391f1e /dev/zero: make private mapping full anonymous mapping
> > * a68d3cbfade64 memstick: core: fix kernel-doc notation
> >
> > so I applied below patch also upon a68d3cbfad.
> >
> > we saw big improvement but not that big.
> >
> > =========================================================================================
> > compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
> > gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
> >
> > commit:
> > a68d3cbfad ("memstick: core: fix kernel-doc notation")
> > 52ec85cb99 <--- your patch
> >
> >
> > a68d3cbfade64392 52ec85cb99e9b31dc304eae965a
> > ---------------- ---------------------------
> > %stddev %change %stddev
> > \ | \
> > 14364828 ± 4% +410.6% 73349239 ± 3% vm-scalability.throughput
> >
> > full comparison as below [1] just FYI.
>
> Thanks for the update. I stared at the profiling report for a whole day, but
> I didn't figure out where that 400% lost. I just saw the number of page
> faults was fewer. And it seems like the reduction of page faults match the
> 400% loss. So I did more trace and profiling.
>
> The test case did the below stuff in a tight loop:
> mmap 40K memory from /dev/zero (read only)
> read the area
>
> So two major factors to the performance: mmap and page fault. The
> alternative patch did reduce the overhead of mmap to the same level as the
> original patch.
>
> The further perf profiling showed the cost of page fault is higher than the
> original patch. But the profiling of page fault was interesting:
>
> - 44.87% 0.01% usemem [kernel.kallsyms] [k]
> do_translation_fault
> - 44.86% do_translation_fault
> - 44.83% do_page_fault
> - 44.53% handle_mm_fault
> 9.04% __handle_mm_fault
>
> Page fault consumed 40% of cpu time in handle_mm_fault, but
> __handle_mm_fault just consumed 9%, I expected it should be the major
> consumer.
>
> So I annotated handle_mm_fault, then found the most time was consumed by
> lru_gen_enter_fault() -> vma_has_recency() (my kernel has multi-gen LRU
> enabled):
>
> │ if (vma->vm_file && (vma->vm_file->f_mode & FMODE_NOREUSE))
> │ ↓ cbz x1, b4
> 0.00 │ ldr w0, [x1, #12]
> 99.59 │ eor x0, x0, #0x800000
> 0.00 │ ubfx w0, w0, #23, #1
> │ current->in_lru_fault = vma_has_recency(vma);
> 0.00 │ b4: ldrh w1, [x2, #1992]
> 0.01 │ bfi w1, w0, #5, #1
> 0.00 │ strh w1, [x2, #1992]
>
>
> vma_has_recency() read vma->vm_file->f_mode if vma->vm_file is not NULL. But
> that load took a long time. So I inspected struct file and saw:
>
> struct file {
> file_ref_t f_ref;
> spinlock_t f_lock;
> fmode_t f_mode;
> const struct file_operations *f_op;
> ...
> }
>
> The f_mode is in the same cache line with f_ref (my kernel does NOT have
> spin lock debug enabled). The test case mmap /dev/zero in a tight loop, so
> the refcount is modified (fget/fput) very frequently, this resulted in
> somehow false sharing.
>
> So I tried the below patch on top of the alternative patch:
>
> diff --git a/include/linux/mm_inline.h b/include/linux/mm_inline.h
> index f9157a0c42a5..ba11dc0b1c7c 100644
> --- a/include/linux/mm_inline.h
> +++ b/include/linux/mm_inline.h
> @@ -608,6 +608,9 @@ static inline bool vma_has_recency(struct vm_area_struct
> *vma)
> if (vma->vm_flags & (VM_SEQ_READ | VM_RAND_READ))
> return false;
>
> + if (vma_is_anonymous(vma))
> + return true;
> +
> if (vma->vm_file && (vma->vm_file->f_mode & FMODE_NOREUSE))
> return false;
>
> This made the profiling of page fault look normal:
>
> - 1.90% do_translation_fault
> - 1.87% do_page_fault
> - 1.49% handle_mm_fault
> - 1.36% __handle_mm_fault
>
> Please try this in your test.
>
> But AFAICT I have never seen performance issue reported due to the false
> sharing of refcount and other fields in struct file. This benchmark stressed
> this quite badly.
I applied your above patch upon alternative patch last time, then saw more
improvement (+445.2% vs a68d3cbfad), but still not that big as in our original
report.
=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
commit:
a68d3cbfad ("memstick: core: fix kernel-doc notation")
52ec85cb99 <--- a68d3cbfad + alternative
d4a204fefe <--- a68d3cbfad + alternative + new patch in vma_has_recency()
a68d3cbfade64392 52ec85cb99e9b31dc304eae965a d4a204fefec91546a317e52ae19
---------------- --------------------------- ---------------------------
%stddev %change %stddev %change %stddev
\ | \ | \
14364828 ± 4% +410.6% 73349239 ± 3% +445.2% 78318730 ± 4% vm-scalability.throughput
full comparison is as below:
=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
commit:
a68d3cbfad ("memstick: core: fix kernel-doc notation")
52ec85cb99 <--- a68d3cbfad + alternative
d4a204fefe <--- a68d3cbfad + alternative + new patch in vma_has_recency()
a68d3cbfade64392 52ec85cb99e9b31dc304eae965a d4a204fefec91546a317e52ae19
---------------- --------------------------- ---------------------------
%stddev %change %stddev %change %stddev
\ | \ | \
5.262e+09 ± 3% -45.0% 2.896e+09 ± 6% +10.0% 5.791e+09 ±126% cpuidle..time
7924008 ± 3% -79.3% 1643339 ± 11% -77.4% 1791703 ± 12% cpuidle..usage
1871164 ± 4% -22.4% 1452554 ± 12% -20.9% 1479724 ± 13% numa-numastat.node3.local_node
1952164 ± 3% -20.1% 1560294 ± 12% -19.1% 1580192 ± 12% numa-numastat.node3.numa_hit
399.52 -68.2% 126.86 -65.9% 136.26 ± 23% uptime.boot
14507 -15.7% 12232 +5.2% 15256 ± 48% uptime.idle
6.99 ± 3% +147.9% 17.34 ± 4% +249.9% 24.47 ± 62% vmstat.cpu.id
1.71 +473.6% 9.79 ± 2% +437.6% 9.18 ± 19% vmstat.cpu.us
34204 ± 5% -72.9% 9272 ± 7% -73.5% 9074 ± 16% vmstat.system.cs
266575 -21.2% 210191 -26.9% 194776 ± 20% vmstat.system.in
3408 ± 5% -99.8% 8.38 ± 48% -99.6% 13.38 ± 68% perf-c2c.DRAM.local
18076 ± 3% -99.8% 32.25 ± 27% -99.7% 54.12 ± 35% perf-c2c.DRAM.remote
8082 ± 5% -99.8% 15.50 ± 64% -99.7% 26.38 ± 52% perf-c2c.HITM.local
6544 ± 6% -99.8% 13.62 ± 51% -99.7% 19.25 ± 43% perf-c2c.HITM.remote
14627 ± 4% -99.8% 29.12 ± 53% -99.7% 45.62 ± 43% perf-c2c.HITM.total
6.49 ± 3% +8.8 15.24 ± 5% +15.9 22.44 ± 71% mpstat.cpu.all.idle%
0.63 -0.3 0.32 ± 4% -0.3 0.31 ± 22% mpstat.cpu.all.irq%
0.03 ± 2% +0.2 0.26 ± 2% +0.2 0.25 ± 20% mpstat.cpu.all.soft%
91.17 -17.0 74.15 -23.6 67.58 ± 20% mpstat.cpu.all.sys%
1.68 ± 2% +8.3 10.03 ± 2% +7.7 9.42 ± 19% mpstat.cpu.all.usr%
337.33 -97.4% 8.88 ± 75% -98.2% 6.00 ± 88% mpstat.max_utilization.seconds
352.76 -77.3% 79.95 ± 2% -78.5% 75.89 ± 3% time.elapsed_time
352.76 -77.3% 79.95 ± 2% -78.5% 75.89 ± 3% time.elapsed_time.max
225965 ± 7% -16.0% 189844 ± 6% -20.6% 179334 ± 3% time.involuntary_context_switches
9.592e+08 ± 4% +11.9% 1.074e+09 +11.9% 1.074e+09 time.minor_page_faults
20852 -8.8% 19012 -9.8% 18815 time.percent_of_cpu_this_job_got
72302 -81.4% 13425 ± 3% -82.6% 12566 ± 4% time.system_time
1260 ± 3% +41.0% 1777 +36.2% 1716 time.user_time
5393707 ± 5% -98.4% 86880 ± 17% -98.2% 96659 ± 22% time.voluntary_context_switches
1609925 -50.3% 800493 -51.0% 788816 ± 2% meminfo.Active
1609925 -50.3% 800493 -51.0% 788816 ± 2% meminfo.Active(anon)
160837 ± 33% -63.9% 58119 ± 13% -65.9% 54899 ± 31% meminfo.AnonHugePages
4435665 -18.5% 3614714 -18.7% 3604829 meminfo.Cached
1775547 -43.8% 998415 -44.8% 980447 ± 3% meminfo.Committed_AS
148539 -43.7% 83699 ± 4% -46.1% 80050 ± 2% meminfo.Mapped
4245538 ± 4% -20.9% 3356561 -28.0% 3056817 ± 20% meminfo.PageTables
14166291 ± 4% -9.6% 12806082 -15.9% 11919101 ± 19% meminfo.SUnreclaim
929777 -88.2% 109274 ± 3% -89.4% 98935 ± 15% meminfo.Shmem
14315492 ± 4% -9.6% 12947821 -15.7% 12061412 ± 19% meminfo.Slab
25676018 ± 3% +10.9% 28487403 +16.3% 29863951 ± 8% meminfo.max_used_kB
64129 ± 4% +418.9% 332751 ± 3% +453.6% 355040 ± 4% vm-scalability.median
45.40 ± 5% +1961.8 2007 ± 8% +2094.7 2140 ± 11% vm-scalability.stddev%
14364828 ± 4% +410.6% 73349239 ± 3% +445.2% 78318730 ± 4% vm-scalability.throughput
352.76 -77.3% 79.95 ± 2% -78.5% 75.89 ± 3% vm-scalability.time.elapsed_time
352.76 -77.3% 79.95 ± 2% -78.5% 75.89 ± 3% vm-scalability.time.elapsed_time.max
225965 ± 7% -16.0% 189844 ± 6% -20.6% 179334 ± 3% vm-scalability.time.involuntary_context_switches
9.592e+08 ± 4% +11.9% 1.074e+09 +11.9% 1.074e+09 vm-scalability.time.minor_page_faults
20852 -8.8% 19012 -9.8% 18815 vm-scalability.time.percent_of_cpu_this_job_got
72302 -81.4% 13425 ± 3% -82.6% 12566 ± 4% vm-scalability.time.system_time
1260 ± 3% +41.0% 1777 +36.2% 1716 vm-scalability.time.user_time
5393707 ± 5% -98.4% 86880 ± 17% -98.2% 96659 ± 22% vm-scalability.time.voluntary_context_switches
4.316e+09 ± 4% +11.9% 4.832e+09 +11.9% 4.832e+09 vm-scalability.workload
265763 ± 4% -20.5% 211398 ± 4% -28.7% 189557 ± 22% numa-vmstat.node0.nr_page_table_pages
31364 ±106% -85.0% 4690 ±169% -66.5% 10503 ±106% numa-vmstat.node0.nr_shmem
891094 ± 4% -8.0% 819697 ± 3% -17.0% 739565 ± 21% numa-vmstat.node0.nr_slab_unreclaimable
12205 ± 67% -74.1% 3161 ±199% -30.0% 8543 ± 98% numa-vmstat.node1.nr_mapped
265546 ± 4% -21.8% 207742 ± 4% -27.1% 193704 ± 22% numa-vmstat.node1.nr_page_table_pages
44052 ± 71% -86.0% 6163 ±161% -92.9% 3126 ±239% numa-vmstat.node1.nr_shmem
885590 ± 4% -9.9% 797649 ± 4% -15.0% 752585 ± 21% numa-vmstat.node1.nr_slab_unreclaimable
264589 ± 4% -21.2% 208598 ± 4% -28.0% 190497 ± 20% numa-vmstat.node2.nr_page_table_pages
881598 ± 4% -10.0% 793829 ± 4% -15.3% 747142 ± 19% numa-vmstat.node2.nr_slab_unreclaimable
192683 ± 30% -61.0% 75078 ± 70% -90.4% 18510 ±122% numa-vmstat.node3.nr_active_anon
286819 ±108% -93.0% 19993 ± 39% -88.8% 32096 ± 44% numa-vmstat.node3.nr_file_pages
13124 ± 49% -92.3% 1006 ± 57% -96.1% 510.58 ± 55% numa-vmstat.node3.nr_mapped
264499 ± 4% -22.1% 206135 ± 2% -30.9% 182777 ± 21% numa-vmstat.node3.nr_page_table_pages
139810 ± 14% -90.5% 13229 ± 89% -99.4% 844.61 ± 73% numa-vmstat.node3.nr_shmem
880199 ± 4% -11.8% 776210 ± 5% -18.3% 718982 ± 21% numa-vmstat.node3.nr_slab_unreclaimable
192683 ± 30% -61.0% 75077 ± 70% -90.4% 18510 ±122% numa-vmstat.node3.nr_zone_active_anon
1951359 ± 3% -20.1% 1558936 ± 12% -19.1% 1578968 ± 12% numa-vmstat.node3.numa_hit
1870359 ± 4% -22.4% 1451195 ± 12% -21.0% 1478500 ± 13% numa-vmstat.node3.numa_local
402515 -50.3% 200150 -51.0% 197173 ± 2% proc-vmstat.nr_active_anon
170568 +1.9% 173746 +1.7% 173416 proc-vmstat.nr_anon_pages
4257257 +0.9% 4296664 +1.7% 4330365 proc-vmstat.nr_dirty_background_threshold
8524925 +0.9% 8603835 +1.7% 8671318 proc-vmstat.nr_dirty_threshold
1109246 -18.5% 903959 -18.7% 901412 proc-vmstat.nr_file_pages
42815276 +0.9% 43210344 +1.7% 43547728 proc-vmstat.nr_free_pages
37525 -43.6% 21164 ± 4% -46.1% 20229 ± 2% proc-vmstat.nr_mapped
1059932 ± 4% -21.1% 836810 -28.3% 760302 ± 20% proc-vmstat.nr_page_table_pages
232507 -88.2% 27341 ± 3% -89.4% 24701 ± 15% proc-vmstat.nr_shmem
37297 -5.0% 35436 -4.6% 35576 proc-vmstat.nr_slab_reclaimable
3537843 ± 4% -9.8% 3192506 -16.1% 2966663 ± 20% proc-vmstat.nr_slab_unreclaimable
402515 -50.3% 200150 -51.0% 197173 ± 2% proc-vmstat.nr_zone_active_anon
61931 ± 8% -83.8% 10023 ± 45% -76.8% 14345 ± 33% proc-vmstat.numa_hint_faults
15755 ± 21% -87.1% 2039 ± 97% -79.9% 3159 ± 84% proc-vmstat.numa_hint_faults_local
6916516 ± 3% -7.1% 6425430 -7.0% 6429349 proc-vmstat.numa_hit
6568542 ± 3% -7.5% 6077764 -7.4% 6081764 proc-vmstat.numa_local
293942 ± 3% -69.6% 89435 ± 49% -68.7% 92135 ± 33% proc-vmstat.numa_pte_updates
9.608e+08 ± 4% +11.8% 1.074e+09 +11.8% 1.074e+09 proc-vmstat.pgfault
55981 ± 2% -63.1% 20641 ± 2% -61.6% 21497 ± 15% proc-vmstat.pgreuse
1063552 ± 4% -20.3% 847673 ± 4% -28.4% 761616 ± 21% numa-meminfo.node0.PageTables
3565610 ± 4% -8.0% 3279375 ± 3% -16.8% 2967130 ± 20% numa-meminfo.node0.SUnreclaim
125455 ±106% -85.2% 18620 ±168% -66.2% 42381 ±106% numa-meminfo.node0.Shmem
3592377 ± 4% -7.1% 3336072 ± 4% -16.2% 3011209 ± 20% numa-meminfo.node0.Slab
48482 ± 67% -74.3% 12475 ±199% -30.6% 33629 ± 99% numa-meminfo.node1.Mapped
1062709 ± 4% -21.7% 831966 ± 4% -26.7% 778849 ± 22% numa-meminfo.node1.PageTables
3543793 ± 4% -10.0% 3189589 ± 4% -14.8% 3018852 ± 21% numa-meminfo.node1.SUnreclaim
176171 ± 71% -86.0% 24677 ±161% -92.9% 12510 ±239% numa-meminfo.node1.Shmem
3593431 ± 4% -10.4% 3220352 ± 4% -14.6% 3069779 ± 21% numa-meminfo.node1.Slab
1058901 ± 4% -21.3% 833124 ± 4% -27.7% 766065 ± 19% numa-meminfo.node2.PageTables
3527862 ± 4% -10.2% 3168666 ± 5% -15.0% 2999540 ± 19% numa-meminfo.node2.SUnreclaim
3565750 ± 4% -10.3% 3200248 ± 5% -15.2% 3022861 ± 19% numa-meminfo.node2.Slab
770405 ± 30% -61.0% 300435 ± 70% -90.4% 74044 ±122% numa-meminfo.node3.Active
770405 ± 30% -61.0% 300435 ± 70% -90.4% 74044 ±122% numa-meminfo.node3.Active(anon)
380096 ± 50% -32.8% 255397 ± 73% -78.2% 82996 ±115% numa-meminfo.node3.AnonPages.max
1146977 ±108% -93.0% 80110 ± 40% -88.8% 128436 ± 44% numa-meminfo.node3.FilePages
52663 ± 47% -91.6% 4397 ± 56% -96.0% 2104 ± 52% numa-meminfo.node3.Mapped
6368902 ± 20% -21.2% 5021246 ± 2% -27.8% 4597733 ± 18% numa-meminfo.node3.MemUsed
1058539 ± 4% -22.2% 823061 ± 3% -30.6% 734757 ± 20% numa-meminfo.node3.PageTables
3522496 ± 4% -12.1% 3096728 ± 6% -18.1% 2885117 ± 21% numa-meminfo.node3.SUnreclaim
558943 ± 14% -90.5% 53054 ± 89% -99.4% 3423 ± 71% numa-meminfo.node3.Shmem
3557392 ± 4% -12.3% 3119454 ± 6% -18.2% 2909118 ± 20% numa-meminfo.node3.Slab
0.82 ± 4% -39.7% 0.50 ± 12% -28.2% 0.59 ± 34% perf-stat.i.MPKI
2.714e+10 ± 2% +185.7% 7.755e+10 ± 6% +174.8% 7.457e+10 ± 27% perf-stat.i.branch-instructions
0.11 ± 3% +0.1 0.20 ± 5% +0.3 0.40 ±121% perf-stat.i.branch-miss-rate%
24932893 +156.6% 63980942 ± 5% +150.2% 62383567 ± 25% perf-stat.i.branch-misses
64.93 -10.1 54.87 ± 2% -13.6 51.34 ± 20% perf-stat.i.cache-miss-rate%
34508 ± 4% -61.4% 13315 ± 10% -64.1% 12391 ± 25% perf-stat.i.context-switches
7.67 -63.7% 2.79 ± 6% -67.4% 2.50 ± 14% perf-stat.i.cpi
224605 +10.8% 248972 ± 4% +11.8% 251127 ± 4% perf-stat.i.cpu-clock
696.35 ± 2% -57.4% 296.79 ± 3% -59.8% 279.73 ± 5% perf-stat.i.cpu-migrations
10834 ± 4% -12.5% 9483 ± 20% -20.2% 8648 ± 28% perf-stat.i.cycles-between-cache-misses
1.102e+11 +128.5% 2.518e+11 ± 6% +119.9% 2.423e+11 ± 27% perf-stat.i.instructions
0.14 +198.2% 0.42 ± 5% +239.7% 0.48 ± 21% perf-stat.i.ipc
24.25 ± 3% +375.8% 115.36 ± 3% +353.8% 110.03 ± 26% perf-stat.i.metric.K/sec
2722043 ± 3% +439.7% 14690226 ± 6% +418.1% 14103930 ± 27% perf-stat.i.minor-faults
2722043 ± 3% +439.7% 14690226 ± 6% +418.1% 14103929 ± 27% perf-stat.i.page-faults
224605 +10.8% 248972 ± 4% +11.8% 251127 ± 4% perf-stat.i.task-clock
0.81 ± 3% -52.5% 0.39 ± 14% -59.6% 0.33 ± 38% perf-stat.overall.MPKI
0.09 -0.0 0.08 ± 2% -0.0 0.07 ± 37% perf-stat.overall.branch-miss-rate%
64.81 -6.4 58.40 -13.3 51.49 ± 37% perf-stat.overall.cache-miss-rate%
7.24 -56.3% 3.17 ± 3% -63.8% 2.62 ± 38% perf-stat.overall.cpi
8933 ± 4% -6.0% 8401 ± 16% -21.3% 7029 ± 38% perf-stat.overall.cycles-between-cache-misses
0.14 +129.0% 0.32 ± 3% +112.0% 0.29 ± 38% perf-stat.overall.ipc
9012 ± 2% -57.5% 3827 -62.8% 3349 ± 37% perf-stat.overall.path-length
2.701e+10 ± 2% +159.6% 7.012e+10 ± 2% +117.1% 5.863e+10 ± 43% perf-stat.ps.branch-instructions
24708939 +119.2% 54173035 +81.0% 44726149 ± 43% perf-stat.ps.branch-misses
34266 ± 5% -73.9% 8949 ± 7% -77.8% 7599 ± 41% perf-stat.ps.context-switches
7.941e+11 -9.1% 7.219e+11 -27.9% 5.729e+11 ± 44% perf-stat.ps.cpu-cycles
693.54 ± 2% -68.6% 217.73 ± 5% -74.1% 179.66 ± 38% perf-stat.ps.cpu-migrations
1.097e+11 +108.1% 2.282e+11 ± 2% +73.9% 1.907e+11 ± 43% perf-stat.ps.instructions
2710577 ± 3% +388.7% 13246535 ± 2% +308.6% 11076222 ± 44% perf-stat.ps.minor-faults
2710577 ± 3% +388.7% 13246536 ± 2% +308.6% 11076222 ± 44% perf-stat.ps.page-faults
3.886e+13 ± 2% -52.4% 1.849e+13 -58.3% 1.619e+13 ± 37% perf-stat.total.instructions
64052898 ± 5% -96.2% 2460331 ±166% -93.1% 4432025 ±129% sched_debug.cfs_rq:/.avg_vruntime.avg
95701822 ± 7% -85.1% 14268127 ±116% -60.2% 38124846 ±118% sched_debug.cfs_rq:/.avg_vruntime.max
43098762 ± 6% -96.0% 1715136 ±173% -93.3% 2867368 ±131% sched_debug.cfs_rq:/.avg_vruntime.min
9223270 ± 9% -84.2% 1457904 ±122% -61.0% 3595639 ±113% sched_debug.cfs_rq:/.avg_vruntime.stddev
0.00 ± 22% -80.1% 0.00 ±185% -86.8% 0.00 ±173% sched_debug.cfs_rq:/.h_nr_delayed.avg
0.69 ± 8% -73.0% 0.19 ±185% -82.0% 0.12 ±173% sched_debug.cfs_rq:/.h_nr_delayed.max
0.05 ± 12% -76.3% 0.01 ±185% -84.2% 0.01 ±173% sched_debug.cfs_rq:/.h_nr_delayed.stddev
0.78 ± 2% -77.0% 0.18 ±130% -71.9% 0.22 ±107% sched_debug.cfs_rq:/.h_nr_running.avg
43049468 ± 22% -89.3% 4590302 ±180% -89.0% 4726833 ±129% sched_debug.cfs_rq:/.left_deadline.max
3836405 ± 37% -85.6% 550773 ±176% -77.5% 864733 ±132% sched_debug.cfs_rq:/.left_deadline.stddev
43049467 ± 22% -89.3% 4590279 ±180% -89.0% 4726820 ±129% sched_debug.cfs_rq:/.left_vruntime.max
3836405 ± 37% -85.6% 550772 ±176% -77.5% 862614 ±132% sched_debug.cfs_rq:/.left_vruntime.stddev
64052901 ± 5% -96.2% 2460341 ±166% -93.1% 4432036 ±129% sched_debug.cfs_rq:/.min_vruntime.avg
95701822 ± 7% -85.1% 14268127 ±116% -60.2% 38124846 ±118% sched_debug.cfs_rq:/.min_vruntime.max
43098762 ± 6% -96.0% 1715136 ±173% -93.3% 2867368 ±131% sched_debug.cfs_rq:/.min_vruntime.min
9223270 ± 9% -84.2% 1457902 ±122% -61.0% 3595638 ±113% sched_debug.cfs_rq:/.min_vruntime.stddev
0.77 ± 2% -77.4% 0.17 ±128% -72.3% 0.21 ±107% sched_debug.cfs_rq:/.nr_running.avg
1.61 ± 24% +396.0% 7.96 ± 62% +355.1% 7.31 ± 52% sched_debug.cfs_rq:/.removed.runnable_avg.avg
86.69 +424.4% 454.62 ± 24% +400.6% 433.98 ± 26% sched_debug.cfs_rq:/.removed.runnable_avg.max
11.14 ± 13% +409.8% 56.79 ± 35% +373.6% 52.77 ± 34% sched_debug.cfs_rq:/.removed.runnable_avg.stddev
1.61 ± 24% +396.0% 7.96 ± 62% +355.1% 7.31 ± 52% sched_debug.cfs_rq:/.removed.util_avg.avg
86.69 +424.4% 454.62 ± 24% +400.6% 433.98 ± 26% sched_debug.cfs_rq:/.removed.util_avg.max
11.14 ± 13% +409.8% 56.79 ± 35% +373.6% 52.77 ± 34% sched_debug.cfs_rq:/.removed.util_avg.stddev
43049467 ± 22% -89.3% 4590282 ±180% -89.0% 4726821 ±129% sched_debug.cfs_rq:/.right_vruntime.max
3836405 ± 37% -85.6% 550772 ±176% -77.5% 862614 ±132% sched_debug.cfs_rq:/.right_vruntime.stddev
286633 ± 43% +262.3% 1038592 ± 36% +188.3% 826260 ± 58% sched_debug.cfs_rq:/.runnable_avg.avg
34728895 ± 30% +349.2% 1.56e+08 ± 26% +293.3% 1.366e+08 ± 60% sched_debug.cfs_rq:/.runnable_avg.max
2845573 ± 30% +325.9% 12119045 ± 26% +251.3% 9995202 ± 55% sched_debug.cfs_rq:/.runnable_avg.stddev
769.03 -69.9% 231.86 ± 84% -66.3% 259.37 ± 72% sched_debug.cfs_rq:/.util_avg.avg
1621 ± 5% -31.5% 1111 ± 8% -35.4% 1048 ± 8% sched_debug.cfs_rq:/.util_avg.max
159.12 ± 8% +22.3% 194.66 ± 12% +35.0% 214.82 ± 14% sched_debug.cfs_rq:/.util_avg.stddev
724.17 ± 2% -89.6% 75.66 ±147% -88.3% 84.74 ±123% sched_debug.cfs_rq:/.util_est.avg
1360 ± 15% -39.2% 826.88 ± 37% -29.0% 965.90 ± 48% sched_debug.cfs_rq:/.util_est.max
766944 ± 3% +18.1% 905901 +21.7% 933047 ± 2% sched_debug.cpu.avg_idle.avg
1067639 ± 5% +30.0% 1387534 ± 16% +38.2% 1475131 ± 15% sched_debug.cpu.avg_idle.max
321459 ± 2% -35.6% 207172 ± 10% -33.5% 213764 ± 15% sched_debug.cpu.avg_idle.stddev
195573 -72.7% 53401 ± 24% -68.5% 61507 ± 35% sched_debug.cpu.clock.avg
195596 -72.7% 53442 ± 24% -68.5% 61565 ± 35% sched_debug.cpu.clock.max
195548 -72.7% 53352 ± 24% -68.6% 61431 ± 35% sched_debug.cpu.clock.min
194424 -72.6% 53229 ± 24% -68.5% 61304 ± 35% sched_debug.cpu.clock_task.avg
194608 -72.6% 53383 ± 24% -68.4% 61478 ± 34% sched_debug.cpu.clock_task.max
181834 -77.5% 40964 ± 31% -73.0% 49012 ± 43% sched_debug.cpu.clock_task.min
4241 ± 2% -80.6% 821.65 ±142% -77.1% 971.85 ±116% sched_debug.cpu.curr->pid.avg
9799 ± 2% -55.4% 4365 ± 17% -51.6% 4747 ± 22% sched_debug.cpu.curr->pid.max
1365 ± 10% -48.0% 709.44 ± 5% -39.9% 820.19 ± 24% sched_debug.cpu.curr->pid.stddev
537665 ± 4% +31.2% 705318 ± 14% +44.0% 774261 ± 15% sched_debug.cpu.max_idle_balance_cost.max
3119 ± 56% +579.1% 21184 ± 39% +1048.3% 35821 ± 65% sched_debug.cpu.max_idle_balance_cost.stddev
0.78 ± 2% -76.3% 0.18 ±135% -72.0% 0.22 ±114% sched_debug.cpu.nr_running.avg
25773 ± 5% -96.1% 1007 ± 41% -95.2% 1246 ± 53% sched_debug.cpu.nr_switches.avg
48669 ± 10% -76.5% 11448 ± 13% -66.5% 16288 ± 70% sched_debug.cpu.nr_switches.max
19006 ± 7% -98.6% 258.81 ± 64% -98.4% 311.75 ± 58% sched_debug.cpu.nr_switches.min
4142 ± 8% -66.3% 1396 ± 17% -58.3% 1726 ± 51% sched_debug.cpu.nr_switches.stddev
0.07 ± 23% -92.9% 0.01 ± 41% -94.3% 0.00 ± 46% sched_debug.cpu.nr_uninterruptible.avg
240.19 ± 16% -82.1% 42.94 ± 41% -84.0% 38.50 ± 19% sched_debug.cpu.nr_uninterruptible.max
-77.92 -88.1% -9.25 -84.9% -11.77 sched_debug.cpu.nr_uninterruptible.min
37.87 ± 5% -85.8% 5.36 ± 13% -85.3% 5.57 ± 5% sched_debug.cpu.nr_uninterruptible.stddev
195549 -72.7% 53356 ± 24% -68.6% 61438 ± 35% sched_debug.cpu_clk
194699 -73.0% 52506 ± 25% -68.9% 60588 ± 35% sched_debug.ktime
0.00 -100.0% 0.00 -62.5% 0.00 ±264% sched_debug.rt_rq:.rt_nr_running.avg
0.17 -100.0% 0.00 -62.5% 0.06 ±264% sched_debug.rt_rq:.rt_nr_running.max
0.01 -100.0% 0.00 -62.5% 0.00 ±264% sched_debug.rt_rq:.rt_nr_running.stddev
196368 -72.4% 54191 ± 24% -68.3% 62327 ± 34% sched_debug.sched_clk
0.17 ±142% -100.0% 0.00 -97.8% 0.00 ±264% perf-sched.sch_delay.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
0.19 ± 34% -51.3% 0.09 ± 37% -76.7% 0.04 ±110% perf-sched.sch_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
0.14 ± 55% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
0.14 ± 73% -82.5% 0.03 ±168% -64.1% 0.05 ±177% perf-sched.sch_delay.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
0.11 ± 59% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
0.04 ±132% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
0.02 ± 31% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
0.00 ±223% +51950.0% 0.26 ±212% +6325.0% 0.03 ±124% perf-sched.sch_delay.avg.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit
0.25 ± 59% -100.0% 0.00 -64.9% 0.09 ±253% perf-sched.sch_delay.avg.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
0.12 ±145% -99.1% 0.00 ±141% -99.5% 0.00 ±264% perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
0.04 ± 55% +99.5% 0.08 ±254% -92.0% 0.00 ±103% perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.25 ± 41% -81.6% 0.05 ± 69% -94.4% 0.01 ± 69% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
0.11 ± 59% -87.1% 0.01 ±198% -96.2% 0.00 ±128% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
0.40 ± 50% -97.8% 0.01 ± 30% -97.2% 0.01 ± 45% perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
2.25 ±138% -99.6% 0.01 ± 7% -63.9% 0.81 ±261% perf-sched.sch_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
0.32 ±104% -97.3% 0.01 ± 38% -97.7% 0.01 ± 61% perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
0.12 ± 21% -61.6% 0.04 ±233% -85.7% 0.02 ±190% perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.01 ± 12% -34.9% 0.01 ± 18% +722.2% 0.07 ±251% perf-sched.sch_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
0.01 ± 42% -41.4% 0.00 ± 72% -76.6% 0.00 ± 77% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown].[unknown]
0.01 ± 20% -100.0% 0.00 -96.4% 0.00 ±264% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
0.19 ±185% -95.6% 0.01 ± 44% +266.3% 0.70 ±261% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
0.07 ± 20% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
0.26 ± 17% -98.8% 0.00 ± 10% -98.9% 0.00 ± 39% perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.03 ± 51% -69.7% 0.01 ± 67% -83.7% 0.01 ± 15% perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
0.01 ± 55% +721.9% 0.10 ± 29% +1608.3% 0.20 ±227% perf-sched.sch_delay.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.01 ±128% -83.6% 0.00 ± 20% -86.2% 0.00 ± 43% perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
0.06 ± 31% +1921.5% 1.23 ±165% +13539.3% 8.30 ±201% perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
1.00 ±151% -100.0% 0.00 -99.6% 0.00 ±264% perf-sched.sch_delay.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
25.45 ± 94% -98.6% 0.36 ± 61% -99.4% 0.15 ±143% perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
4.56 ± 67% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
3.55 ± 97% -98.9% 0.04 ±189% -98.5% 0.05 ±177% perf-sched.sch_delay.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
2.13 ± 67% -77.2% 0.49 ± 56% -88.8% 0.24 ±147% perf-sched.sch_delay.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
3.16 ± 78% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
0.30 ±159% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
1.61 ±100% -76.7% 0.38 ± 72% -91.7% 0.13 ±145% perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
0.03 ± 86% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
0.00 ±223% +3.2e+06% 15.79 ±259% +44450.0% 0.22 ±132% perf-sched.sch_delay.max.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit
3.09 ± 45% -100.0% 0.00 -94.6% 0.17 ±259% perf-sched.sch_delay.max.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
3.51 ± 21% -86.1% 0.49 ± 72% -90.7% 0.33 ±127% perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
0.83 ±160% -99.7% 0.00 ±141% -99.9% 0.00 ±264% perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
0.09 ± 31% +179.7% 0.25 ±258% -91.5% 0.01 ±132% perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
3.59 ± 11% -92.0% 0.29 ±165% -99.2% 0.03 ±118% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
1.60 ± 69% -95.7% 0.07 ±243% -99.0% 0.02 ±210% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
0.81 ± 43% -98.5% 0.01 ± 43% -98.3% 0.01 ± 41% perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
1.02 ± 88% -98.1% 0.02 ± 47% -98.7% 0.01 ± 71% perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
9.68 ± 32% -92.2% 0.76 ± 72% -78.1% 2.12 ±187% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
0.01 ± 49% -51.9% 0.00 ± 72% -80.8% 0.00 ± 77% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown].[unknown]
12.26 ±109% -92.9% 0.87 ±101% -86.9% 1.61 ±225% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
5.60 ±139% -97.6% 0.13 ±132% -99.3% 0.04 ±255% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
0.03 ±106% -100.0% 0.00 -99.1% 0.00 ±264% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
2.11 ± 61% -85.5% 0.31 ± 85% -96.0% 0.08 ±124% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
37.84 ± 47% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
4.68 ± 36% -99.8% 0.01 ± 65% -99.8% 0.01 ± 77% perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
7.56 ± 74% -51.5% 3.67 ±147% -99.8% 0.02 ± 54% perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
0.36 ±186% -96.3% 0.01 ± 90% -97.9% 0.01 ± 59% perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
97903 ± 4% -38.3% 60433 ± 29% -71.4% 27976 ±109% perf-sched.total_wait_and_delay.count.ms
3.97 ± 6% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
302.41 ± 5% -27.4% 219.54 ± 14% -10.8% 269.81 ± 60% perf-sched.wait_and_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
1.48 ± 6% -90.9% 0.14 ± 79% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
327.16 ± 9% -46.6% 174.81 ± 24% -38.4% 201.64 ± 71% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
369.37 ± 2% -75.3% 91.05 ± 35% -77.7% 82.29 ±119% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
0.96 ± 6% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
187.66 +120.6% 413.97 ± 14% +116.9% 407.06 ± 43% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1831 ± 9% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
6.17 ± 45% -79.7% 1.25 ±142% -91.9% 0.50 ±264% perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
14.33 ± 5% +13.4% 16.25 ± 23% -58.1% 6.00 ± 66% perf-sched.wait_and_delay.count.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
810.00 ± 10% -38.0% 502.25 ± 92% -100.0% 0.00 perf-sched.wait_and_delay.count.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
40.50 ± 8% +245.7% 140.00 ± 23% +72.5% 69.88 ± 91% perf-sched.wait_and_delay.count.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
13.17 ± 2% +624.4% 95.38 ± 19% +347.2% 58.88 ± 78% perf-sched.wait_and_delay.count.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
73021 ± 3% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
11323 ± 3% -75.9% 2725 ± 28% -86.4% 1536 ± 34% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1887 ± 45% -96.1% 73.88 ± 78% -98.5% 28.75 ±120% perf-sched.wait_and_delay.count.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
1238 -34.5% 811.25 ± 13% -58.6% 512.62 ± 49% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
35.19 ± 57% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
20.79 ± 19% -95.9% 0.84 ± 93% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
1240 ± 20% -14.4% 1062 ± 10% -25.2% 928.21 ± 40% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
500.34 +31.2% 656.38 ± 39% -15.0% 425.46 ± 61% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
58.83 ± 39% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
1237 ± 34% +151.7% 3114 ± 25% +51.6% 1876 ± 64% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
49.27 ±119% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
58.17 ±187% -100.0% 0.00 -100.0% 0.00 ±264% perf-sched.wait_time.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
3.78 ± 5% -97.6% 0.09 ± 37% -98.8% 0.04 ±111% perf-sched.wait_time.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
2.99 ± 4% +15.4% 3.45 ± 10% +28.8% 3.85 ± 54% perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
3.92 ± 5% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
4.71 ± 8% -99.5% 0.02 ±170% -98.9% 0.05 ±177% perf-sched.wait_time.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
1.67 ± 20% -92.7% 0.12 ± 30% -96.8% 0.05 ±130% perf-sched.wait_time.avg.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
2.10 ± 27% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
0.01 ± 44% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
1.67 ± 21% -94.3% 0.10 ± 35% -97.0% 0.05 ±137% perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
0.04 ±133% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
67.14 ± 73% +75.6% 117.89 ±108% -92.8% 4.82 ±259% perf-sched.wait_time.avg.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
1.65 ± 67% -95.8% 0.07 ±128% -99.2% 0.01 ±175% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
2.30 ± 14% -95.5% 0.10 ± 42% -96.4% 0.08 ±108% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
2.00 ± 74% +2917.4% 60.44 ± 33% +1369.3% 29.43 ± 74% perf-sched.wait_time.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
29.19 ± 5% -38.5% 17.96 ± 28% -49.0% 14.89 ± 54% perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
0.37 ± 30% +5524.5% 20.95 ± 30% +2028.0% 7.93 ±117% perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
302.40 ± 5% -27.4% 219.53 ± 14% -10.8% 269.75 ± 60% perf-sched.wait_time.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
1.40 ± 6% -92.7% 0.10 ± 18% -95.4% 0.06 ±109% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
0.72 ±220% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
326.84 ± 9% -46.6% 174.54 ± 24% -38.6% 200.64 ± 72% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
369.18 ± 2% -75.3% 91.04 ± 35% -74.2% 95.16 ± 98% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
0.89 ± 6% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
187.58 +120.6% 413.77 ± 14% +116.9% 406.79 ± 43% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
2.36 ± 29% +1759.6% 43.80 ± 33% +3763.5% 90.99 ±115% perf-sched.wait_time.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.01 ±156% -97.9% 0.00 ±264% -98.9% 0.00 ±264% perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
750.01 -14.5% 641.50 ± 14% -41.1% 442.13 ± 58% perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
340.69 ±135% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
535.09 ±128% -100.0% 0.00 -100.0% 0.00 ±264% perf-sched.wait_time.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
22.04 ± 32% -98.4% 0.36 ± 61% -99.3% 0.15 ±143% perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
13.57 ± 17% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
13.54 ± 10% -99.7% 0.04 ±189% -99.6% 0.05 ±177% perf-sched.wait_time.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
10.17 ± 19% -95.2% 0.49 ± 56% -97.7% 0.24 ±147% perf-sched.wait_time.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
11.35 ± 25% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
0.01 ± 32% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
10.62 ± 9% -96.5% 0.38 ± 72% -98.7% 0.13 ±145% perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
0.20 ±199% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
1559 ± 64% -92.3% 120.30 ±109% -99.4% 9.63 ±259% perf-sched.wait_time.max.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
6.93 ± 53% -98.1% 0.13 ± 99% -99.8% 0.01 ±175% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
14.42 ± 22% -96.6% 0.49 ± 72% -97.7% 0.33 ±127% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
4.00 ± 74% +19182.5% 772.23 ± 40% +7266.0% 295.00 ± 92% perf-sched.wait_time.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
10.75 ± 98% +6512.2% 710.88 ± 56% +2526.4% 282.37 ±130% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
15.80 ± 8% -95.2% 0.76 ± 72% -86.6% 2.12 ±187% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
11.64 ± 61% -98.9% 0.13 ±132% -99.7% 0.04 ±255% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
2.94 ±213% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
175.70 ±210% -64.6% 62.26 ±263% -99.8% 0.31 ±116% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
1240 ± 20% -14.3% 1062 ± 10% -25.2% 928.20 ± 40% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
500.11 +31.2% 656.37 ± 39% -2.4% 487.96 ± 41% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
32.65 ± 33% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
1237 ± 34% +151.6% 3113 ± 25% +49.0% 1844 ± 63% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.36 ±190% -97.2% 0.01 ±127% -98.5% 0.01 ± 88% perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
95.59 -95.6 0.00 -95.6 0.00 perf-profile.calltrace.cycles-pp.__mmap
95.54 -95.5 0.00 -95.5 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
95.54 -95.5 0.00 -95.5 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap
94.54 -94.5 0.00 -94.5 0.00 perf-profile.calltrace.cycles-pp.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
94.46 -94.0 0.41 ±138% -93.9 0.57 ±103% perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
94.14 -93.7 0.40 ±136% -93.6 0.50 ± 79% perf-profile.calltrace.cycles-pp.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff
93.79 -93.5 0.31 ±134% -93.2 0.58 ±111% perf-profile.calltrace.cycles-pp.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff
93.40 -93.4 0.00 -93.4 0.00 perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma.__mmap_region
93.33 -93.3 0.00 -93.3 0.00 perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma
93.44 -93.3 0.14 ±264% -93.4 0.00 perf-profile.calltrace.cycles-pp.down_write.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap
94.45 -93.0 1.42 ± 60% -92.9 1.51 ± 51% perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
94.25 -92.9 1.33 ± 61% -92.8 1.43 ± 57% perf-profile.calltrace.cycles-pp.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
92.89 -92.9 0.00 -92.9 0.00 perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file
0.00 +0.3 0.29 ±129% +1.1 1.10 ± 27% perf-profile.calltrace.cycles-pp.do_open.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat
0.00 +0.3 0.32 ±129% +1.7 1.70 ± 39% perf-profile.calltrace.cycles-pp.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin.generic_perform_write.shmem_file_write_iter
0.00 +0.3 0.32 ±129% +1.7 1.74 ± 40% perf-profile.calltrace.cycles-pp.shmem_get_folio_gfp.shmem_write_begin.generic_perform_write.shmem_file_write_iter.vfs_write
0.00 +0.5 0.49 ± 78% +1.7 1.74 ± 40% perf-profile.calltrace.cycles-pp.shmem_write_begin.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
0.00 +1.1 1.09 ± 33% +0.4 0.44 ±177% perf-profile.calltrace.cycles-pp.dup_mmap.dup_mm.copy_process.kernel_clone.__do_sys_clone
0.00 +1.3 1.32 ± 54% +1.4 1.36 ± 33% perf-profile.calltrace.cycles-pp.filp_close.put_files_struct.do_exit.do_group_exit.get_signal
0.00 +1.3 1.32 ± 54% +1.4 1.36 ± 33% perf-profile.calltrace.cycles-pp.put_files_struct.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
0.00 +1.4 1.37 ± 49% +1.8 1.77 ± 50% perf-profile.calltrace.cycles-pp.setlocale
0.00 +1.4 1.39 ± 70% +1.8 1.80 ± 48% perf-profile.calltrace.cycles-pp.seq_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +1.4 1.39 ± 70% +1.8 1.80 ± 48% perf-profile.calltrace.cycles-pp.seq_read_iter.seq_read.vfs_read.ksys_read.do_syscall_64
0.00 +1.5 1.55 ± 63% +1.6 1.62 ± 37% perf-profile.calltrace.cycles-pp.do_read_fault.do_pte_missing.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
0.00 +1.6 1.60 ± 57% +1.6 1.63 ± 87% perf-profile.calltrace.cycles-pp.swevent_hlist_put_cpu.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release
0.00 +1.6 1.64 ± 47% +0.9 0.90 ±101% perf-profile.calltrace.cycles-pp.handle_softirqs.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt
0.00 +1.6 1.64 ± 47% +1.0 1.02 ± 83% perf-profile.calltrace.cycles-pp.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry
0.00 +1.6 1.65 ± 43% +1.1 1.15 ± 76% perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +1.8 1.76 ± 44% +1.1 1.15 ± 76% perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +1.9 1.93 ± 26% +1.1 1.11 ±127% perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64
0.00 +2.0 2.04 ± 66% +3.6 3.65 ± 42% perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64
0.00 +2.1 2.12 ± 58% +3.6 3.65 ± 42% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64
0.00 +2.1 2.12 ± 58% +3.6 3.65 ± 42% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.open64
0.00 +2.1 2.12 ± 58% +3.7 3.71 ± 40% perf-profile.calltrace.cycles-pp.open64
0.00 +2.2 2.16 ± 44% +1.6 1.62 ± 37% perf-profile.calltrace.cycles-pp.do_pte_missing.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
0.00 +2.2 2.20 ± 74% +3.6 3.65 ± 42% perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64
0.00 +2.2 2.23 ± 33% +1.4 1.40 ± 99% perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +2.3 2.34 ±103% +5.1 5.09 ± 64% perf-profile.calltrace.cycles-pp.__cmd_record.cmd_record.perf_c2c__record.run_builtin.handle_internal_command
0.00 +2.3 2.34 ±103% +5.1 5.09 ± 64% perf-profile.calltrace.cycles-pp.cmd_record.perf_c2c__record.run_builtin.handle_internal_command.main
0.00 +2.3 2.34 ±103% +5.1 5.09 ± 64% perf-profile.calltrace.cycles-pp.perf_c2c__record.run_builtin.handle_internal_command.main
0.00 +2.4 2.37 ± 36% +1.9 1.93 ± 35% perf-profile.calltrace.cycles-pp.zap_present_ptes.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
0.00 +2.5 2.48 ± 32% +2.4 2.45 ± 60% perf-profile.calltrace.cycles-pp.get_cpu_sleep_time_us.get_idle_time.uptime_proc_show.seq_read_iter.vfs_read
0.00 +2.5 2.50 ± 45% +1.2 1.21 ± 73% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
0.00 +2.5 2.54 ± 47% +1.3 1.28 ± 61% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group
0.00 +2.5 2.54 ± 47% +1.3 1.28 ± 61% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
0.00 +2.6 2.62 ± 35% +1.6 1.57 ± 91% perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
0.00 +2.6 2.62 ± 35% +1.6 1.57 ± 91% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
0.00 +2.6 2.62 ± 35% +1.6 1.57 ± 91% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe._Fork
0.00 +2.6 2.62 ± 35% +1.6 1.57 ± 91% perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
0.00 +2.7 2.67 ± 54% +2.6 2.59 ± 40% perf-profile.calltrace.cycles-pp.load_elf_binary.search_binary_handler.exec_binprm.bprm_execve.do_execveat_common
0.00 +2.7 2.68 ± 35% +3.0 3.02 ± 45% perf-profile.calltrace.cycles-pp.get_idle_time.uptime_proc_show.seq_read_iter.vfs_read.ksys_read
0.00 +2.8 2.77 ± 33% +4.2 4.17 ± 35% perf-profile.calltrace.cycles-pp.uptime_proc_show.seq_read_iter.vfs_read.ksys_read.do_syscall_64
0.00 +2.8 2.82 ± 32% +1.8 1.83 ± 85% perf-profile.calltrace.cycles-pp._Fork
0.00 +2.8 2.83 ± 48% +2.6 2.59 ± 40% perf-profile.calltrace.cycles-pp.search_binary_handler.exec_binprm.bprm_execve.do_execveat_common.__x64_sys_execve
0.00 +2.8 2.83 ± 48% +2.7 2.68 ± 42% perf-profile.calltrace.cycles-pp.exec_binprm.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64
0.00 +2.8 2.84 ± 45% +1.2 1.21 ± 73% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
0.00 +2.8 2.84 ± 45% +1.2 1.21 ± 73% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
0.00 +2.9 2.89 ± 39% +3.1 3.14 ± 39% perf-profile.calltrace.cycles-pp.event_function_call.perf_event_release_kernel.perf_release.__fput.task_work_run
0.00 +2.9 2.89 ± 39% +3.1 3.14 ± 39% perf-profile.calltrace.cycles-pp.smp_call_function_single.event_function_call.perf_event_release_kernel.perf_release.__fput
0.00 +3.1 3.10 ± 64% +0.9 0.91 ±264% perf-profile.calltrace.cycles-pp.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +3.1 3.10 ± 64% +0.9 0.91 ±264% perf-profile.calltrace.cycles-pp.seq_read_iter.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64
0.00 +3.1 3.13 ± 33% +1.7 1.68 ± 77% perf-profile.calltrace.cycles-pp.asm_exc_page_fault
0.00 +3.2 3.18 ± 37% +4.3 4.31 ± 34% perf-profile.calltrace.cycles-pp.seq_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +3.2 3.20 ± 28% +3.0 3.02 ± 73% perf-profile.calltrace.cycles-pp.mutex_unlock.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release
0.00 +3.2 3.24 ± 39% +2.8 2.85 ± 49% perf-profile.calltrace.cycles-pp.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +3.2 3.24 ± 36% +2.0 2.00 ± 56% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +3.2 3.24 ± 36% +2.0 2.00 ± 56% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64
0.00 +3.2 3.24 ± 36% +2.0 2.00 ± 56% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +3.2 3.24 ± 36% +2.0 2.00 ± 56% perf-profile.calltrace.cycles-pp.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +3.8 3.85 ± 39% +3.3 3.25 ± 47% perf-profile.calltrace.cycles-pp.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
0.00 +3.8 3.85 ± 39% +3.3 3.25 ± 47% perf-profile.calltrace.cycles-pp.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
0.00 +3.8 3.85 ± 39% +3.3 3.25 ± 47% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
0.00 +3.8 3.85 ± 39% +3.3 3.25 ± 47% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.execve
0.00 +3.8 3.85 ± 39% +3.3 3.29 ± 47% perf-profile.calltrace.cycles-pp.execve
0.00 +4.0 4.04 ± 43% +5.2 5.21 ± 49% perf-profile.calltrace.cycles-pp.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +4.0 4.04 ± 43% +5.2 5.21 ± 49% perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64
0.00 +4.1 4.10 ± 30% +2.6 2.56 ± 28% perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.exit_mmap.__mmput.exit_mm
0.00 +4.2 4.18 ± 31% +2.8 2.82 ± 21% perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap.__mmput
0.00 +4.2 4.18 ± 31% +2.8 2.82 ± 21% perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap
0.00 +4.2 4.20 ± 28% +2.7 2.68 ± 34% perf-profile.calltrace.cycles-pp.unmap_vmas.exit_mmap.__mmput.exit_mm.do_exit
0.00 +4.2 4.25 ± 65% +8.0 7.98 ± 43% perf-profile.calltrace.cycles-pp.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64
0.00 +4.3 4.27 ± 26% +3.2 3.23 ± 34% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
0.00 +4.3 4.30 ± 22% +3.9 3.95 ± 32% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.get_signal
0.00 +4.3 4.30 ± 22% +3.9 3.95 ± 32% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
0.00 +4.5 4.46 ± 59% +8.1 8.07 ± 42% perf-profile.calltrace.cycles-pp.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +4.6 4.57 ± 58% +8.1 8.07 ± 42% perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen
0.00 +4.7 4.68 ± 55% +8.1 8.12 ± 43% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn
0.00 +4.7 4.68 ± 55% +8.1 8.12 ± 43% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn.perf_mmap__push
0.00 +4.7 4.68 ± 55% +8.2 8.16 ± 44% perf-profile.calltrace.cycles-pp.write.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist
0.00 +4.7 4.68 ± 55% +8.4 8.39 ± 39% perf-profile.calltrace.cycles-pp.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record
0.00 +4.7 4.68 ± 55% +8.6 8.61 ± 38% perf-profile.calltrace.cycles-pp.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record
0.00 +4.9 4.90 ± 57% +10.3 10.28 ± 65% perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
0.00 +4.9 4.92 ± 26% +4.6 4.56 ± 47% perf-profile.calltrace.cycles-pp.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release.__fput
0.00 +5.0 4.99 ±100% +2.6 2.64 ±101% perf-profile.calltrace.cycles-pp.__intel_pmu_enable_all.perf_rotate_context.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt
0.00 +5.0 4.99 ±100% +2.6 2.64 ±101% perf-profile.calltrace.cycles-pp.perf_rotate_context.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt
0.00 +5.1 5.08 ±102% +2.6 2.64 ±101% perf-profile.calltrace.cycles-pp.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt
0.00 +5.1 5.14 ± 28% +6.0 6.01 ± 41% perf-profile.calltrace.cycles-pp.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin
0.00 +5.1 5.14 ± 28% +6.2 6.16 ± 39% perf-profile.calltrace.cycles-pp.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin.handle_internal_command
0.00 +5.4 5.43 ± 25% +5.0 4.97 ± 45% perf-profile.calltrace.cycles-pp._free_event.perf_event_release_kernel.perf_release.__fput.task_work_run
0.00 +5.8 5.82 ± 94% +4.2 4.21 ± 49% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt
0.00 +5.8 5.82 ± 94% +4.3 4.35 ± 53% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry
0.00 +6.1 6.07 ± 90% +4.3 4.32 ± 58% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
0.00 +6.6 6.62 ± 24% +7.0 6.99 ± 41% perf-profile.calltrace.cycles-pp.__cmd_record.cmd_record.run_builtin.handle_internal_command.main
0.00 +6.6 6.62 ± 24% +7.0 6.99 ± 41% perf-profile.calltrace.cycles-pp.cmd_record.run_builtin.handle_internal_command.main
0.00 +6.8 6.76 ± 18% +5.2 5.23 ± 25% perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit
0.00 +7.6 7.56 ± 76% +6.0 5.99 ± 38% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter
0.00 +8.0 8.03 ± 27% +7.4 7.37 ± 52% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
0.00 +8.0 8.03 ± 27% +7.4 7.37 ± 52% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
0.00 +8.0 8.05 ± 68% +6.3 6.27 ± 37% perf-profile.calltrace.cycles-pp.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter
0.00 +8.1 8.13 ± 28% +7.4 7.37 ± 52% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
0.00 +8.1 8.13 ± 28% +7.4 7.37 ± 52% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
0.00 +8.1 8.13 ± 28% +7.4 7.37 ± 52% perf-profile.calltrace.cycles-pp.read
0.00 +9.1 9.05 ± 35% +13.9 13.88 ± 19% perf-profile.calltrace.cycles-pp.handle_internal_command.main
0.00 +9.1 9.05 ± 35% +13.9 13.88 ± 19% perf-profile.calltrace.cycles-pp.main
0.00 +9.1 9.05 ± 35% +13.9 13.88 ± 19% perf-profile.calltrace.cycles-pp.run_builtin.handle_internal_command.main
0.00 +9.3 9.26 ± 30% +9.0 8.96 ± 31% perf-profile.calltrace.cycles-pp.perf_event_release_kernel.perf_release.__fput.task_work_run.do_exit
0.00 +9.3 9.26 ± 30% +9.0 8.96 ± 31% perf-profile.calltrace.cycles-pp.perf_release.__fput.task_work_run.do_exit.do_group_exit
0.00 +10.1 10.14 ± 28% +10.0 10.04 ± 34% perf-profile.calltrace.cycles-pp.__fput.task_work_run.do_exit.do_group_exit.get_signal
0.00 +10.2 10.23 ± 27% +10.7 10.65 ± 35% perf-profile.calltrace.cycles-pp.task_work_run.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
0.00 +11.0 10.98 ± 55% +13.0 13.00 ± 27% perf-profile.calltrace.cycles-pp.asm_sysvec_reschedule_ipi.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
0.00 +20.6 20.64 ± 30% +19.5 19.49 ± 43% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +20.6 20.64 ± 30% +19.5 19.49 ± 43% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
1.21 ± 3% +36.6 37.80 ± 12% +34.1 35.32 ± 11% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
1.21 ± 3% +36.6 37.80 ± 12% +34.4 35.62 ± 11% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.common_startup_64
1.22 ± 3% +36.8 38.00 ± 13% +34.8 36.05 ± 11% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.common_startup_64
1.22 ± 3% +36.9 38.10 ± 13% +34.8 36.05 ± 11% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.common_startup_64
1.22 ± 3% +36.9 38.10 ± 13% +34.8 36.05 ± 11% perf-profile.calltrace.cycles-pp.start_secondary.common_startup_64
1.21 ± 3% +37.2 38.43 ± 11% +34.2 35.40 ± 8% perf-profile.calltrace.cycles-pp.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
1.21 ± 3% +37.2 38.43 ± 11% +34.2 35.40 ± 8% perf-profile.calltrace.cycles-pp.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
1.21 ± 3% +37.3 38.54 ± 12% +34.7 35.87 ± 10% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
1.22 ± 3% +37.6 38.84 ± 12% +35.4 36.60 ± 11% perf-profile.calltrace.cycles-pp.common_startup_64
2.19 ± 3% +53.9 56.10 ± 19% +48.4 50.63 ± 13% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
95.60 -95.2 0.41 ±138% -94.9 0.72 ± 95% perf-profile.children.cycles-pp.__mmap
94.14 -93.7 0.49 ±130% -92.9 1.21 ± 33% perf-profile.children.cycles-pp.__mmap_new_vma
93.79 -93.5 0.31 ±134% -93.1 0.71 ± 78% perf-profile.children.cycles-pp.vma_link_file
93.40 -93.4 0.00 -93.4 0.00 perf-profile.children.cycles-pp.rwsem_down_write_slowpath
93.33 -93.3 0.00 -93.3 0.00 perf-profile.children.cycles-pp.rwsem_optimistic_spin
94.55 -93.1 1.42 ± 60% -93.0 1.55 ± 50% perf-profile.children.cycles-pp.ksys_mmap_pgoff
92.91 -92.9 0.00 -92.9 0.00 perf-profile.children.cycles-pp.osq_lock
93.44 -92.7 0.75 ±109% -93.4 0.06 ±264% perf-profile.children.cycles-pp.down_write
94.46 -92.6 1.84 ± 34% -92.0 2.48 ± 28% perf-profile.children.cycles-pp.vm_mmap_pgoff
94.45 -92.6 1.84 ± 34% -92.0 2.48 ± 28% perf-profile.children.cycles-pp.do_mmap
94.25 -92.6 1.66 ± 37% -91.9 2.40 ± 30% perf-profile.children.cycles-pp.__mmap_region
95.58 -44.8 50.78 ± 11% -42.8 52.76 ± 11% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
95.58 -44.8 50.78 ± 11% -42.8 52.76 ± 11% perf-profile.children.cycles-pp.do_syscall_64
0.00 +0.1 0.09 ±264% +1.0 0.96 ± 46% perf-profile.children.cycles-pp.kcpustat_cpu_fetch
0.25 ± 3% +0.2 0.45 ±133% +0.7 0.92 ± 41% perf-profile.children.cycles-pp.vma_interval_tree_insert
0.00 +0.3 0.29 ±129% +1.2 1.16 ± 26% perf-profile.children.cycles-pp.do_open
0.00 +0.3 0.32 ±129% +1.8 1.79 ± 43% perf-profile.children.cycles-pp.shmem_alloc_and_add_folio
0.00 +0.3 0.32 ±129% +1.8 1.83 ± 44% perf-profile.children.cycles-pp.shmem_get_folio_gfp
0.00 +0.5 0.49 ± 78% +1.8 1.83 ± 44% perf-profile.children.cycles-pp.shmem_write_begin
0.00 +1.1 1.09 ± 33% +0.5 0.48 ±160% perf-profile.children.cycles-pp.dup_mmap
0.00 +1.1 1.11 ±106% +1.6 1.60 ± 54% perf-profile.children.cycles-pp.__open64_nocancel
0.00 +1.1 1.15 ±102% +1.2 1.16 ± 86% perf-profile.children.cycles-pp.evlist_cpu_iterator__next
0.00 +1.3 1.32 ± 54% +1.4 1.36 ± 33% perf-profile.children.cycles-pp.filp_close
0.00 +1.3 1.32 ± 54% +1.5 1.47 ± 29% perf-profile.children.cycles-pp.put_files_struct
0.00 +1.4 1.37 ± 49% +1.8 1.77 ± 50% perf-profile.children.cycles-pp.setlocale
0.00 +1.4 1.39 ± 70% +1.8 1.80 ± 48% perf-profile.children.cycles-pp.seq_read
0.00 +1.5 1.55 ± 63% +1.7 1.75 ± 30% perf-profile.children.cycles-pp.do_read_fault
0.00 +1.7 1.66 ± 76% +0.9 0.91 ± 44% perf-profile.children.cycles-pp.event_function
0.00 +1.7 1.66 ± 76% +0.9 0.91 ± 44% perf-profile.children.cycles-pp.remote_function
0.00 +1.7 1.70 ± 71% +1.5 1.53 ± 73% perf-profile.children.cycles-pp.lookup_fast
0.00 +1.7 1.73 ± 53% +1.4 1.40 ± 77% perf-profile.children.cycles-pp.swevent_hlist_put_cpu
0.04 ± 44% +1.8 1.83 ± 96% +2.4 2.47 ± 44% perf-profile.children.cycles-pp.__schedule
0.00 +1.9 1.93 ± 26% +1.1 1.15 ±120% perf-profile.children.cycles-pp.dup_mm
0.03 ± 70% +2.0 1.99 ± 36% +1.2 1.23 ± 81% perf-profile.children.cycles-pp.handle_softirqs
0.00 +2.0 1.99 ± 36% +1.1 1.13 ± 67% perf-profile.children.cycles-pp.__irq_exit_rcu
0.00 +2.0 2.02 ± 38% +1.3 1.33 ± 57% perf-profile.children.cycles-pp.folios_put_refs
0.00 +2.1 2.06 ± 52% +1.4 1.38 ± 77% perf-profile.children.cycles-pp._raw_spin_lock
0.00 +2.1 2.12 ± 58% +3.7 3.71 ± 40% perf-profile.children.cycles-pp.open64
0.00 +2.2 2.16 ± 44% +1.7 1.75 ± 30% perf-profile.children.cycles-pp.do_pte_missing
0.00 +2.2 2.21 ± 68% +2.2 2.18 ± 58% perf-profile.children.cycles-pp.link_path_walk
0.00 +2.2 2.23 ± 33% +1.4 1.40 ± 99% perf-profile.children.cycles-pp.copy_process
0.00 +2.3 2.30 ± 40% +1.8 1.78 ± 48% perf-profile.children.cycles-pp.__tlb_batch_free_encoded_pages
0.00 +2.3 2.30 ± 40% +1.8 1.78 ± 48% perf-profile.children.cycles-pp.free_pages_and_swap_cache
0.00 +2.3 2.34 ±103% +5.1 5.09 ± 64% perf-profile.children.cycles-pp.perf_c2c__record
0.00 +2.3 2.34 ± 46% +1.5 1.52 ± 99% perf-profile.children.cycles-pp.walk_component
0.00 +2.4 2.37 ± 36% +2.0 2.04 ± 32% perf-profile.children.cycles-pp.zap_present_ptes
0.00 +2.5 2.48 ± 32% +2.5 2.51 ± 55% perf-profile.children.cycles-pp.get_cpu_sleep_time_us
0.00 +2.5 2.50 ± 73% +1.6 1.56 ± 76% perf-profile.children.cycles-pp.__evlist__enable
0.00 +2.6 2.62 ± 35% +1.6 1.57 ± 91% perf-profile.children.cycles-pp.__do_sys_clone
0.00 +2.6 2.62 ± 35% +1.6 1.57 ± 91% perf-profile.children.cycles-pp.kernel_clone
0.00 +2.7 2.67 ± 54% +2.6 2.59 ± 40% perf-profile.children.cycles-pp.load_elf_binary
0.00 +2.7 2.68 ± 35% +3.0 3.02 ± 45% perf-profile.children.cycles-pp.get_idle_time
0.00 +2.8 2.77 ± 33% +4.2 4.17 ± 35% perf-profile.children.cycles-pp.uptime_proc_show
0.00 +2.8 2.83 ± 48% +2.6 2.59 ± 40% perf-profile.children.cycles-pp.search_binary_handler
0.00 +2.8 2.83 ± 48% +2.7 2.68 ± 42% perf-profile.children.cycles-pp.exec_binprm
0.00 +2.9 2.91 ± 32% +1.8 1.83 ± 85% perf-profile.children.cycles-pp._Fork
0.00 +3.1 3.10 ± 64% +0.9 0.95 ±252% perf-profile.children.cycles-pp.proc_reg_read_iter
0.00 +3.2 3.24 ± 39% +2.8 2.85 ± 49% perf-profile.children.cycles-pp.bprm_execve
0.00 +3.2 3.24 ± 36% +2.0 2.00 ± 56% perf-profile.children.cycles-pp.__x64_sys_exit_group
0.00 +3.2 3.24 ± 36% +2.1 2.09 ± 53% perf-profile.children.cycles-pp.x64_sys_call
0.00 +3.8 3.85 ± 39% +3.3 3.29 ± 47% perf-profile.children.cycles-pp.execve
0.00 +3.8 3.85 ± 39% +3.3 3.34 ± 49% perf-profile.children.cycles-pp.__x64_sys_execve
0.00 +3.8 3.85 ± 39% +3.3 3.34 ± 49% perf-profile.children.cycles-pp.do_execveat_common
0.00 +4.0 3.99 ± 38% +4.1 4.06 ± 54% perf-profile.children.cycles-pp.mutex_unlock
0.00 +4.2 4.19 ± 31% +3.0 3.02 ± 20% perf-profile.children.cycles-pp.zap_pte_range
0.00 +4.2 4.25 ± 65% +8.0 7.98 ± 43% perf-profile.children.cycles-pp.generic_perform_write
0.00 +4.3 4.29 ± 29% +3.0 3.02 ± 20% perf-profile.children.cycles-pp.unmap_page_range
0.00 +4.3 4.29 ± 29% +3.0 3.02 ± 20% perf-profile.children.cycles-pp.zap_pmd_range
0.00 +4.3 4.31 ± 51% +5.3 5.31 ± 46% perf-profile.children.cycles-pp.do_filp_open
0.00 +4.3 4.31 ± 51% +5.3 5.31 ± 46% perf-profile.children.cycles-pp.path_openat
0.19 ± 23% +4.4 4.60 ± 26% +3.4 3.54 ± 27% perf-profile.children.cycles-pp.__handle_mm_fault
0.00 +4.5 4.46 ± 59% +8.1 8.07 ± 42% perf-profile.children.cycles-pp.shmem_file_write_iter
0.00 +4.5 4.55 ± 24% +4.0 3.97 ± 39% perf-profile.children.cycles-pp.smp_call_function_single
0.00 +4.5 4.55 ± 24% +4.1 4.06 ± 38% perf-profile.children.cycles-pp.event_function_call
0.00 +4.6 4.58 ± 30% +3.2 3.19 ± 24% perf-profile.children.cycles-pp.unmap_vmas
0.51 ± 6% +4.6 5.14 ± 24% +3.6 4.06 ± 30% perf-profile.children.cycles-pp.handle_mm_fault
0.00 +4.7 4.68 ± 55% +8.4 8.41 ± 39% perf-profile.children.cycles-pp.writen
0.00 +4.7 4.68 ± 55% +8.5 8.49 ± 39% perf-profile.children.cycles-pp.record__pushfn
0.00 +4.8 4.80 ± 48% +6.1 6.15 ± 34% perf-profile.children.cycles-pp.do_sys_openat2
0.77 ± 3% +4.8 5.59 ± 21% +4.3 5.07 ± 29% perf-profile.children.cycles-pp.exc_page_fault
0.76 ± 3% +4.8 5.59 ± 21% +4.3 5.07 ± 29% perf-profile.children.cycles-pp.do_user_addr_fault
0.00 +4.9 4.90 ± 57% +10.3 10.28 ± 65% perf-profile.children.cycles-pp.vfs_write
0.00 +4.9 4.90 ± 57% +10.4 10.41 ± 63% perf-profile.children.cycles-pp.ksys_write
0.00 +4.9 4.90 ± 48% +6.1 6.15 ± 34% perf-profile.children.cycles-pp.__x64_sys_openat
0.00 +4.9 4.92 ± 26% +4.7 4.66 ± 47% perf-profile.children.cycles-pp.sw_perf_event_destroy
0.00 +5.0 4.99 ±100% +2.6 2.64 ±101% perf-profile.children.cycles-pp.perf_rotate_context
0.00 +5.0 5.01 ± 54% +10.9 10.87 ± 59% perf-profile.children.cycles-pp.write
0.00 +5.1 5.09 ±102% +2.7 2.74 ± 94% perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
0.00 +5.4 5.43 ± 25% +5.0 4.97 ± 45% perf-profile.children.cycles-pp._free_event
1.18 +5.6 6.78 ± 20% +5.5 6.71 ± 24% perf-profile.children.cycles-pp.asm_exc_page_fault
0.46 +5.6 6.07 ± 90% +4.1 4.54 ± 53% perf-profile.children.cycles-pp.__hrtimer_run_queues
0.00 +5.7 5.75 ± 39% +10.2 10.22 ± 24% perf-profile.children.cycles-pp.perf_mmap__push
0.00 +5.7 5.75 ± 39% +10.4 10.38 ± 23% perf-profile.children.cycles-pp.record__mmap_read_evlist
0.53 +5.8 6.28 ± 89% +4.4 4.91 ± 50% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
0.53 +5.8 6.28 ± 89% +4.4 4.91 ± 50% perf-profile.children.cycles-pp.hrtimer_interrupt
0.00 +6.6 6.65 ± 77% +3.3 3.32 ± 91% perf-profile.children.cycles-pp.__intel_pmu_enable_all
0.00 +6.8 6.85 ± 20% +5.2 5.23 ± 25% perf-profile.children.cycles-pp.exit_mm
0.58 ± 2% +7.6 8.14 ± 75% +6.0 6.55 ± 38% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
0.00 +7.7 7.67 ± 23% +6.1 6.14 ± 15% perf-profile.children.cycles-pp.exit_mmap
0.00 +7.7 7.67 ± 30% +7.0 7.05 ± 50% perf-profile.children.cycles-pp.seq_read_iter
0.00 +7.7 7.72 ± 80% +8.2 8.15 ± 51% perf-profile.children.cycles-pp.asm_sysvec_reschedule_ipi
0.00 +7.8 7.75 ± 23% +6.1 6.14 ± 15% perf-profile.children.cycles-pp.__mmput
0.00 +8.0 8.03 ± 27% +7.4 7.37 ± 52% perf-profile.children.cycles-pp.ksys_read
0.00 +8.0 8.03 ± 27% +7.4 7.37 ± 52% perf-profile.children.cycles-pp.vfs_read
0.00 +8.1 8.13 ± 28% +7.4 7.37 ± 52% perf-profile.children.cycles-pp.read
0.02 ±141% +9.0 9.05 ± 35% +13.9 13.88 ± 19% perf-profile.children.cycles-pp.__cmd_record
0.02 ±141% +9.0 9.05 ± 35% +13.9 13.88 ± 19% perf-profile.children.cycles-pp.cmd_record
0.02 ±141% +9.0 9.05 ± 35% +13.9 13.88 ± 19% perf-profile.children.cycles-pp.handle_internal_command
0.02 ±141% +9.0 9.05 ± 35% +13.9 13.88 ± 19% perf-profile.children.cycles-pp.main
0.02 ±141% +9.0 9.05 ± 35% +13.9 13.88 ± 19% perf-profile.children.cycles-pp.run_builtin
0.00 +9.3 9.26 ± 30% +9.0 8.96 ± 31% perf-profile.children.cycles-pp.perf_event_release_kernel
0.00 +9.3 9.26 ± 30% +9.0 8.96 ± 31% perf-profile.children.cycles-pp.perf_release
1.02 ± 4% +9.3 10.33 ± 27% +9.8 10.80 ± 35% perf-profile.children.cycles-pp.task_work_run
0.00 +11.0 11.05 ± 28% +10.4 10.37 ± 32% perf-profile.children.cycles-pp.__fput
0.00 +15.8 15.85 ± 25% +16.1 16.11 ± 29% perf-profile.children.cycles-pp.get_signal
0.00 +15.8 15.85 ± 25% +16.2 16.17 ± 29% perf-profile.children.cycles-pp.arch_do_signal_or_restart
0.00 +19.1 19.09 ± 19% +18.1 18.06 ± 29% perf-profile.children.cycles-pp.do_exit
0.00 +19.1 19.09 ± 19% +18.1 18.06 ± 29% perf-profile.children.cycles-pp.do_group_exit
1.70 ± 2% +30.7 32.41 ± 21% +27.2 28.87 ± 12% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
1.22 ± 3% +36.9 38.10 ± 13% +34.8 36.05 ± 11% perf-profile.children.cycles-pp.start_secondary
1.21 ± 3% +37.2 38.43 ± 11% +34.2 35.40 ± 8% perf-profile.children.cycles-pp.acpi_idle_do_entry
1.21 ± 3% +37.2 38.43 ± 11% +34.2 35.40 ± 8% perf-profile.children.cycles-pp.acpi_idle_enter
1.21 ± 3% +37.2 38.43 ± 11% +34.2 35.40 ± 8% perf-profile.children.cycles-pp.acpi_safe_halt
1.22 ± 3% +37.3 38.54 ± 12% +35.0 36.18 ± 10% perf-profile.children.cycles-pp.cpuidle_idle_call
1.21 ± 3% +37.3 38.54 ± 12% +34.7 35.87 ± 10% perf-profile.children.cycles-pp.cpuidle_enter
1.21 ± 3% +37.3 38.54 ± 12% +34.7 35.87 ± 10% perf-profile.children.cycles-pp.cpuidle_enter_state
1.22 ± 3% +37.6 38.84 ± 12% +35.4 36.60 ± 11% perf-profile.children.cycles-pp.common_startup_64
1.22 ± 3% +37.6 38.84 ± 12% +35.4 36.60 ± 11% perf-profile.children.cycles-pp.cpu_startup_entry
1.22 ± 3% +37.6 38.84 ± 12% +35.4 36.60 ± 11% perf-profile.children.cycles-pp.do_idle
92.37 -92.4 0.00 -92.4 0.00 perf-profile.self.cycles-pp.osq_lock
0.00 +0.1 0.09 ±264% +0.8 0.84 ± 51% perf-profile.self.cycles-pp.kcpustat_cpu_fetch
0.00 +2.1 2.06 ± 52% +1.4 1.38 ± 77% perf-profile.self.cycles-pp._raw_spin_lock
0.00 +2.6 2.61 ± 36% +2.8 2.75 ± 48% perf-profile.self.cycles-pp.smp_call_function_single
0.00 +3.7 3.68 ± 37% +3.7 3.70 ± 64% perf-profile.self.cycles-pp.mutex_unlock
0.00 +6.6 6.65 ± 77% +3.3 3.32 ± 91% perf-profile.self.cycles-pp.__intel_pmu_enable_all
1.19 ± 3% +29.2 30.38 ± 15% +27.9 29.13 ± 13% perf-profile.self.cycles-pp.acpi_safe_halt
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-02-13 2:04 ` Oliver Sang
@ 2025-02-14 22:53 ` Yang Shi
2025-02-18 6:30 ` Oliver Sang
0 siblings, 1 reply; 35+ messages in thread
From: Yang Shi @ 2025-02-14 22:53 UTC (permalink / raw)
To: Oliver Sang
Cc: oe-lkp, lkp, linux-kernel, arnd, gregkh, Liam.Howlett,
lorenzo.stoakes, vbabka, jannh, willy, liushixin2, akpm,
linux-mm
On 2/12/25 6:04 PM, Oliver Sang wrote:
> hi, Yang Shi,
>
> On Fri, Feb 07, 2025 at 10:10:37AM -0800, Yang Shi wrote:
>> On 2/6/25 12:02 AM, Oliver Sang wrote:
> [...]
>
>>> since we applied your "/dev/zero: make private mapping full anonymous mapping"
>>> patch upon a68d3cbfad like below:
>>>
>>> * 7143ee2391f1e /dev/zero: make private mapping full anonymous mapping
>>> * a68d3cbfade64 memstick: core: fix kernel-doc notation
>>>
>>> so I applied below patch also upon a68d3cbfad.
>>>
>>> we saw big improvement but not that big.
>>>
>>> =========================================================================================
>>> compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
>>> gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
>>>
>>> commit:
>>> a68d3cbfad ("memstick: core: fix kernel-doc notation")
>>> 52ec85cb99 <--- your patch
>>>
>>>
>>> a68d3cbfade64392 52ec85cb99e9b31dc304eae965a
>>> ---------------- ---------------------------
>>> %stddev %change %stddev
>>> \ | \
>>> 14364828 ± 4% +410.6% 73349239 ± 3% vm-scalability.throughput
>>>
>>> full comparison as below [1] just FYI.
>> Thanks for the update. I stared at the profiling report for a whole day, but
>> I didn't figure out where that 400% lost. I just saw the number of page
>> faults was fewer. And it seems like the reduction of page faults match the
>> 400% loss. So I did more trace and profiling.
>>
>> The test case did the below stuff in a tight loop:
>> mmap 40K memory from /dev/zero (read only)
>> read the area
>>
>> So two major factors to the performance: mmap and page fault. The
>> alternative patch did reduce the overhead of mmap to the same level as the
>> original patch.
>>
>> The further perf profiling showed the cost of page fault is higher than the
>> original patch. But the profiling of page fault was interesting:
>>
>> - 44.87% 0.01% usemem [kernel.kallsyms] [k]
>> do_translation_fault
>> - 44.86% do_translation_fault
>> - 44.83% do_page_fault
>> - 44.53% handle_mm_fault
>> 9.04% __handle_mm_fault
>>
>> Page fault consumed 40% of cpu time in handle_mm_fault, but
>> __handle_mm_fault just consumed 9%, I expected it should be the major
>> consumer.
>>
>> So I annotated handle_mm_fault, then found the most time was consumed by
>> lru_gen_enter_fault() -> vma_has_recency() (my kernel has multi-gen LRU
>> enabled):
>>
>> │ if (vma->vm_file && (vma->vm_file->f_mode & FMODE_NOREUSE))
>> │ ↓ cbz x1, b4
>> 0.00 │ ldr w0, [x1, #12]
>> 99.59 │ eor x0, x0, #0x800000
>> 0.00 │ ubfx w0, w0, #23, #1
>> │ current->in_lru_fault = vma_has_recency(vma);
>> 0.00 │ b4: ldrh w1, [x2, #1992]
>> 0.01 │ bfi w1, w0, #5, #1
>> 0.00 │ strh w1, [x2, #1992]
>>
>>
>> vma_has_recency() read vma->vm_file->f_mode if vma->vm_file is not NULL. But
>> that load took a long time. So I inspected struct file and saw:
>>
>> struct file {
>> file_ref_t f_ref;
>> spinlock_t f_lock;
>> fmode_t f_mode;
>> const struct file_operations *f_op;
>> ...
>> }
>>
>> The f_mode is in the same cache line with f_ref (my kernel does NOT have
>> spin lock debug enabled). The test case mmap /dev/zero in a tight loop, so
>> the refcount is modified (fget/fput) very frequently, this resulted in
>> somehow false sharing.
>>
>> So I tried the below patch on top of the alternative patch:
>>
>> diff --git a/include/linux/mm_inline.h b/include/linux/mm_inline.h
>> index f9157a0c42a5..ba11dc0b1c7c 100644
>> --- a/include/linux/mm_inline.h
>> +++ b/include/linux/mm_inline.h
>> @@ -608,6 +608,9 @@ static inline bool vma_has_recency(struct vm_area_struct
>> *vma)
>> if (vma->vm_flags & (VM_SEQ_READ | VM_RAND_READ))
>> return false;
>>
>> + if (vma_is_anonymous(vma))
>> + return true;
>> +
>> if (vma->vm_file && (vma->vm_file->f_mode & FMODE_NOREUSE))
>> return false;
>>
>> This made the profiling of page fault look normal:
>>
>> - 1.90% do_translation_fault
>> - 1.87% do_page_fault
>> - 1.49% handle_mm_fault
>> - 1.36% __handle_mm_fault
>>
>> Please try this in your test.
>>
>> But AFAICT I have never seen performance issue reported due to the false
>> sharing of refcount and other fields in struct file. This benchmark stressed
>> this quite badly.
> I applied your above patch upon alternative patch last time, then saw more
> improvement (+445.2% vs a68d3cbfad), but still not that big as in our original
> report.
Thanks for the update. It looks like the problem is still in page
faults. I did my test on arm64 machine. I also noticed struct file has
"__randomize_layout", so it may have different layout on x86 than arm64?
The page fault handler may also access other fields of struct file that
may cause false sharing, for example, accessing f_mapping to read gfp
flags. This may not be a problem on my machine, but may be more costly
on yours depending on the real layout of struct file on the machines,
Can you please try the below patch on top of the current patches? Thank
you so much for your patience.
diff --git a/mm/memory.c b/mm/memory.c
index 539c0f7c6d54..1fa9dbce0f66 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -3214,6 +3214,9 @@ static gfp_t __get_fault_gfp_mask(struct
vm_area_struct *vma)
{
struct file *vm_file = vma->vm_file;
+ if (vma_is_anonymous(vma))
+ return GFP_KERNEL;
+
if (vm_file)
return mapping_gfp_mask(vm_file->f_mapping) | __GFP_FS
| __GFP_IO;
>
> =========================================================================================
> compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
> gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
>
> commit:
> a68d3cbfad ("memstick: core: fix kernel-doc notation")
> 52ec85cb99 <--- a68d3cbfad + alternative
> d4a204fefe <--- a68d3cbfad + alternative + new patch in vma_has_recency()
>
> a68d3cbfade64392 52ec85cb99e9b31dc304eae965a d4a204fefec91546a317e52ae19
> ---------------- --------------------------- ---------------------------
> %stddev %change %stddev %change %stddev
> \ | \ | \
> 14364828 ± 4% +410.6% 73349239 ± 3% +445.2% 78318730 ± 4% vm-scalability.throughput
>
>
> full comparison is as below:
>
> =========================================================================================
> compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
> gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
>
> commit:
> a68d3cbfad ("memstick: core: fix kernel-doc notation")
> 52ec85cb99 <--- a68d3cbfad + alternative
> d4a204fefe <--- a68d3cbfad + alternative + new patch in vma_has_recency()
>
> a68d3cbfade64392 52ec85cb99e9b31dc304eae965a d4a204fefec91546a317e52ae19
> ---------------- --------------------------- ---------------------------
> %stddev %change %stddev %change %stddev
> \ | \ | \
> 5.262e+09 ± 3% -45.0% 2.896e+09 ± 6% +10.0% 5.791e+09 ±126% cpuidle..time
> 7924008 ± 3% -79.3% 1643339 ± 11% -77.4% 1791703 ± 12% cpuidle..usage
> 1871164 ± 4% -22.4% 1452554 ± 12% -20.9% 1479724 ± 13% numa-numastat.node3.local_node
> 1952164 ± 3% -20.1% 1560294 ± 12% -19.1% 1580192 ± 12% numa-numastat.node3.numa_hit
> 399.52 -68.2% 126.86 -65.9% 136.26 ± 23% uptime.boot
> 14507 -15.7% 12232 +5.2% 15256 ± 48% uptime.idle
> 6.99 ± 3% +147.9% 17.34 ± 4% +249.9% 24.47 ± 62% vmstat.cpu.id
> 1.71 +473.6% 9.79 ± 2% +437.6% 9.18 ± 19% vmstat.cpu.us
> 34204 ± 5% -72.9% 9272 ± 7% -73.5% 9074 ± 16% vmstat.system.cs
> 266575 -21.2% 210191 -26.9% 194776 ± 20% vmstat.system.in
> 3408 ± 5% -99.8% 8.38 ± 48% -99.6% 13.38 ± 68% perf-c2c.DRAM.local
> 18076 ± 3% -99.8% 32.25 ± 27% -99.7% 54.12 ± 35% perf-c2c.DRAM.remote
> 8082 ± 5% -99.8% 15.50 ± 64% -99.7% 26.38 ± 52% perf-c2c.HITM.local
> 6544 ± 6% -99.8% 13.62 ± 51% -99.7% 19.25 ± 43% perf-c2c.HITM.remote
> 14627 ± 4% -99.8% 29.12 ± 53% -99.7% 45.62 ± 43% perf-c2c.HITM.total
> 6.49 ± 3% +8.8 15.24 ± 5% +15.9 22.44 ± 71% mpstat.cpu.all.idle%
> 0.63 -0.3 0.32 ± 4% -0.3 0.31 ± 22% mpstat.cpu.all.irq%
> 0.03 ± 2% +0.2 0.26 ± 2% +0.2 0.25 ± 20% mpstat.cpu.all.soft%
> 91.17 -17.0 74.15 -23.6 67.58 ± 20% mpstat.cpu.all.sys%
> 1.68 ± 2% +8.3 10.03 ± 2% +7.7 9.42 ± 19% mpstat.cpu.all.usr%
> 337.33 -97.4% 8.88 ± 75% -98.2% 6.00 ± 88% mpstat.max_utilization.seconds
> 352.76 -77.3% 79.95 ± 2% -78.5% 75.89 ± 3% time.elapsed_time
> 352.76 -77.3% 79.95 ± 2% -78.5% 75.89 ± 3% time.elapsed_time.max
> 225965 ± 7% -16.0% 189844 ± 6% -20.6% 179334 ± 3% time.involuntary_context_switches
> 9.592e+08 ± 4% +11.9% 1.074e+09 +11.9% 1.074e+09 time.minor_page_faults
> 20852 -8.8% 19012 -9.8% 18815 time.percent_of_cpu_this_job_got
> 72302 -81.4% 13425 ± 3% -82.6% 12566 ± 4% time.system_time
> 1260 ± 3% +41.0% 1777 +36.2% 1716 time.user_time
> 5393707 ± 5% -98.4% 86880 ± 17% -98.2% 96659 ± 22% time.voluntary_context_switches
> 1609925 -50.3% 800493 -51.0% 788816 ± 2% meminfo.Active
> 1609925 -50.3% 800493 -51.0% 788816 ± 2% meminfo.Active(anon)
> 160837 ± 33% -63.9% 58119 ± 13% -65.9% 54899 ± 31% meminfo.AnonHugePages
> 4435665 -18.5% 3614714 -18.7% 3604829 meminfo.Cached
> 1775547 -43.8% 998415 -44.8% 980447 ± 3% meminfo.Committed_AS
> 148539 -43.7% 83699 ± 4% -46.1% 80050 ± 2% meminfo.Mapped
> 4245538 ± 4% -20.9% 3356561 -28.0% 3056817 ± 20% meminfo.PageTables
> 14166291 ± 4% -9.6% 12806082 -15.9% 11919101 ± 19% meminfo.SUnreclaim
> 929777 -88.2% 109274 ± 3% -89.4% 98935 ± 15% meminfo.Shmem
> 14315492 ± 4% -9.6% 12947821 -15.7% 12061412 ± 19% meminfo.Slab
> 25676018 ± 3% +10.9% 28487403 +16.3% 29863951 ± 8% meminfo.max_used_kB
> 64129 ± 4% +418.9% 332751 ± 3% +453.6% 355040 ± 4% vm-scalability.median
> 45.40 ± 5% +1961.8 2007 ± 8% +2094.7 2140 ± 11% vm-scalability.stddev%
> 14364828 ± 4% +410.6% 73349239 ± 3% +445.2% 78318730 ± 4% vm-scalability.throughput
> 352.76 -77.3% 79.95 ± 2% -78.5% 75.89 ± 3% vm-scalability.time.elapsed_time
> 352.76 -77.3% 79.95 ± 2% -78.5% 75.89 ± 3% vm-scalability.time.elapsed_time.max
> 225965 ± 7% -16.0% 189844 ± 6% -20.6% 179334 ± 3% vm-scalability.time.involuntary_context_switches
> 9.592e+08 ± 4% +11.9% 1.074e+09 +11.9% 1.074e+09 vm-scalability.time.minor_page_faults
> 20852 -8.8% 19012 -9.8% 18815 vm-scalability.time.percent_of_cpu_this_job_got
> 72302 -81.4% 13425 ± 3% -82.6% 12566 ± 4% vm-scalability.time.system_time
> 1260 ± 3% +41.0% 1777 +36.2% 1716 vm-scalability.time.user_time
> 5393707 ± 5% -98.4% 86880 ± 17% -98.2% 96659 ± 22% vm-scalability.time.voluntary_context_switches
> 4.316e+09 ± 4% +11.9% 4.832e+09 +11.9% 4.832e+09 vm-scalability.workload
> 265763 ± 4% -20.5% 211398 ± 4% -28.7% 189557 ± 22% numa-vmstat.node0.nr_page_table_pages
> 31364 ±106% -85.0% 4690 ±169% -66.5% 10503 ±106% numa-vmstat.node0.nr_shmem
> 891094 ± 4% -8.0% 819697 ± 3% -17.0% 739565 ± 21% numa-vmstat.node0.nr_slab_unreclaimable
> 12205 ± 67% -74.1% 3161 ±199% -30.0% 8543 ± 98% numa-vmstat.node1.nr_mapped
> 265546 ± 4% -21.8% 207742 ± 4% -27.1% 193704 ± 22% numa-vmstat.node1.nr_page_table_pages
> 44052 ± 71% -86.0% 6163 ±161% -92.9% 3126 ±239% numa-vmstat.node1.nr_shmem
> 885590 ± 4% -9.9% 797649 ± 4% -15.0% 752585 ± 21% numa-vmstat.node1.nr_slab_unreclaimable
> 264589 ± 4% -21.2% 208598 ± 4% -28.0% 190497 ± 20% numa-vmstat.node2.nr_page_table_pages
> 881598 ± 4% -10.0% 793829 ± 4% -15.3% 747142 ± 19% numa-vmstat.node2.nr_slab_unreclaimable
> 192683 ± 30% -61.0% 75078 ± 70% -90.4% 18510 ±122% numa-vmstat.node3.nr_active_anon
> 286819 ±108% -93.0% 19993 ± 39% -88.8% 32096 ± 44% numa-vmstat.node3.nr_file_pages
> 13124 ± 49% -92.3% 1006 ± 57% -96.1% 510.58 ± 55% numa-vmstat.node3.nr_mapped
> 264499 ± 4% -22.1% 206135 ± 2% -30.9% 182777 ± 21% numa-vmstat.node3.nr_page_table_pages
> 139810 ± 14% -90.5% 13229 ± 89% -99.4% 844.61 ± 73% numa-vmstat.node3.nr_shmem
> 880199 ± 4% -11.8% 776210 ± 5% -18.3% 718982 ± 21% numa-vmstat.node3.nr_slab_unreclaimable
> 192683 ± 30% -61.0% 75077 ± 70% -90.4% 18510 ±122% numa-vmstat.node3.nr_zone_active_anon
> 1951359 ± 3% -20.1% 1558936 ± 12% -19.1% 1578968 ± 12% numa-vmstat.node3.numa_hit
> 1870359 ± 4% -22.4% 1451195 ± 12% -21.0% 1478500 ± 13% numa-vmstat.node3.numa_local
> 402515 -50.3% 200150 -51.0% 197173 ± 2% proc-vmstat.nr_active_anon
> 170568 +1.9% 173746 +1.7% 173416 proc-vmstat.nr_anon_pages
> 4257257 +0.9% 4296664 +1.7% 4330365 proc-vmstat.nr_dirty_background_threshold
> 8524925 +0.9% 8603835 +1.7% 8671318 proc-vmstat.nr_dirty_threshold
> 1109246 -18.5% 903959 -18.7% 901412 proc-vmstat.nr_file_pages
> 42815276 +0.9% 43210344 +1.7% 43547728 proc-vmstat.nr_free_pages
> 37525 -43.6% 21164 ± 4% -46.1% 20229 ± 2% proc-vmstat.nr_mapped
> 1059932 ± 4% -21.1% 836810 -28.3% 760302 ± 20% proc-vmstat.nr_page_table_pages
> 232507 -88.2% 27341 ± 3% -89.4% 24701 ± 15% proc-vmstat.nr_shmem
> 37297 -5.0% 35436 -4.6% 35576 proc-vmstat.nr_slab_reclaimable
> 3537843 ± 4% -9.8% 3192506 -16.1% 2966663 ± 20% proc-vmstat.nr_slab_unreclaimable
> 402515 -50.3% 200150 -51.0% 197173 ± 2% proc-vmstat.nr_zone_active_anon
> 61931 ± 8% -83.8% 10023 ± 45% -76.8% 14345 ± 33% proc-vmstat.numa_hint_faults
> 15755 ± 21% -87.1% 2039 ± 97% -79.9% 3159 ± 84% proc-vmstat.numa_hint_faults_local
> 6916516 ± 3% -7.1% 6425430 -7.0% 6429349 proc-vmstat.numa_hit
> 6568542 ± 3% -7.5% 6077764 -7.4% 6081764 proc-vmstat.numa_local
> 293942 ± 3% -69.6% 89435 ± 49% -68.7% 92135 ± 33% proc-vmstat.numa_pte_updates
> 9.608e+08 ± 4% +11.8% 1.074e+09 +11.8% 1.074e+09 proc-vmstat.pgfault
> 55981 ± 2% -63.1% 20641 ± 2% -61.6% 21497 ± 15% proc-vmstat.pgreuse
> 1063552 ± 4% -20.3% 847673 ± 4% -28.4% 761616 ± 21% numa-meminfo.node0.PageTables
> 3565610 ± 4% -8.0% 3279375 ± 3% -16.8% 2967130 ± 20% numa-meminfo.node0.SUnreclaim
> 125455 ±106% -85.2% 18620 ±168% -66.2% 42381 ±106% numa-meminfo.node0.Shmem
> 3592377 ± 4% -7.1% 3336072 ± 4% -16.2% 3011209 ± 20% numa-meminfo.node0.Slab
> 48482 ± 67% -74.3% 12475 ±199% -30.6% 33629 ± 99% numa-meminfo.node1.Mapped
> 1062709 ± 4% -21.7% 831966 ± 4% -26.7% 778849 ± 22% numa-meminfo.node1.PageTables
> 3543793 ± 4% -10.0% 3189589 ± 4% -14.8% 3018852 ± 21% numa-meminfo.node1.SUnreclaim
> 176171 ± 71% -86.0% 24677 ±161% -92.9% 12510 ±239% numa-meminfo.node1.Shmem
> 3593431 ± 4% -10.4% 3220352 ± 4% -14.6% 3069779 ± 21% numa-meminfo.node1.Slab
> 1058901 ± 4% -21.3% 833124 ± 4% -27.7% 766065 ± 19% numa-meminfo.node2.PageTables
> 3527862 ± 4% -10.2% 3168666 ± 5% -15.0% 2999540 ± 19% numa-meminfo.node2.SUnreclaim
> 3565750 ± 4% -10.3% 3200248 ± 5% -15.2% 3022861 ± 19% numa-meminfo.node2.Slab
> 770405 ± 30% -61.0% 300435 ± 70% -90.4% 74044 ±122% numa-meminfo.node3.Active
> 770405 ± 30% -61.0% 300435 ± 70% -90.4% 74044 ±122% numa-meminfo.node3.Active(anon)
> 380096 ± 50% -32.8% 255397 ± 73% -78.2% 82996 ±115% numa-meminfo.node3.AnonPages.max
> 1146977 ±108% -93.0% 80110 ± 40% -88.8% 128436 ± 44% numa-meminfo.node3.FilePages
> 52663 ± 47% -91.6% 4397 ± 56% -96.0% 2104 ± 52% numa-meminfo.node3.Mapped
> 6368902 ± 20% -21.2% 5021246 ± 2% -27.8% 4597733 ± 18% numa-meminfo.node3.MemUsed
> 1058539 ± 4% -22.2% 823061 ± 3% -30.6% 734757 ± 20% numa-meminfo.node3.PageTables
> 3522496 ± 4% -12.1% 3096728 ± 6% -18.1% 2885117 ± 21% numa-meminfo.node3.SUnreclaim
> 558943 ± 14% -90.5% 53054 ± 89% -99.4% 3423 ± 71% numa-meminfo.node3.Shmem
> 3557392 ± 4% -12.3% 3119454 ± 6% -18.2% 2909118 ± 20% numa-meminfo.node3.Slab
> 0.82 ± 4% -39.7% 0.50 ± 12% -28.2% 0.59 ± 34% perf-stat.i.MPKI
> 2.714e+10 ± 2% +185.7% 7.755e+10 ± 6% +174.8% 7.457e+10 ± 27% perf-stat.i.branch-instructions
> 0.11 ± 3% +0.1 0.20 ± 5% +0.3 0.40 ±121% perf-stat.i.branch-miss-rate%
> 24932893 +156.6% 63980942 ± 5% +150.2% 62383567 ± 25% perf-stat.i.branch-misses
> 64.93 -10.1 54.87 ± 2% -13.6 51.34 ± 20% perf-stat.i.cache-miss-rate%
> 34508 ± 4% -61.4% 13315 ± 10% -64.1% 12391 ± 25% perf-stat.i.context-switches
> 7.67 -63.7% 2.79 ± 6% -67.4% 2.50 ± 14% perf-stat.i.cpi
> 224605 +10.8% 248972 ± 4% +11.8% 251127 ± 4% perf-stat.i.cpu-clock
> 696.35 ± 2% -57.4% 296.79 ± 3% -59.8% 279.73 ± 5% perf-stat.i.cpu-migrations
> 10834 ± 4% -12.5% 9483 ± 20% -20.2% 8648 ± 28% perf-stat.i.cycles-between-cache-misses
> 1.102e+11 +128.5% 2.518e+11 ± 6% +119.9% 2.423e+11 ± 27% perf-stat.i.instructions
> 0.14 +198.2% 0.42 ± 5% +239.7% 0.48 ± 21% perf-stat.i.ipc
> 24.25 ± 3% +375.8% 115.36 ± 3% +353.8% 110.03 ± 26% perf-stat.i.metric.K/sec
> 2722043 ± 3% +439.7% 14690226 ± 6% +418.1% 14103930 ± 27% perf-stat.i.minor-faults
> 2722043 ± 3% +439.7% 14690226 ± 6% +418.1% 14103929 ± 27% perf-stat.i.page-faults
> 224605 +10.8% 248972 ± 4% +11.8% 251127 ± 4% perf-stat.i.task-clock
> 0.81 ± 3% -52.5% 0.39 ± 14% -59.6% 0.33 ± 38% perf-stat.overall.MPKI
> 0.09 -0.0 0.08 ± 2% -0.0 0.07 ± 37% perf-stat.overall.branch-miss-rate%
> 64.81 -6.4 58.40 -13.3 51.49 ± 37% perf-stat.overall.cache-miss-rate%
> 7.24 -56.3% 3.17 ± 3% -63.8% 2.62 ± 38% perf-stat.overall.cpi
> 8933 ± 4% -6.0% 8401 ± 16% -21.3% 7029 ± 38% perf-stat.overall.cycles-between-cache-misses
> 0.14 +129.0% 0.32 ± 3% +112.0% 0.29 ± 38% perf-stat.overall.ipc
> 9012 ± 2% -57.5% 3827 -62.8% 3349 ± 37% perf-stat.overall.path-length
> 2.701e+10 ± 2% +159.6% 7.012e+10 ± 2% +117.1% 5.863e+10 ± 43% perf-stat.ps.branch-instructions
> 24708939 +119.2% 54173035 +81.0% 44726149 ± 43% perf-stat.ps.branch-misses
> 34266 ± 5% -73.9% 8949 ± 7% -77.8% 7599 ± 41% perf-stat.ps.context-switches
> 7.941e+11 -9.1% 7.219e+11 -27.9% 5.729e+11 ± 44% perf-stat.ps.cpu-cycles
> 693.54 ± 2% -68.6% 217.73 ± 5% -74.1% 179.66 ± 38% perf-stat.ps.cpu-migrations
> 1.097e+11 +108.1% 2.282e+11 ± 2% +73.9% 1.907e+11 ± 43% perf-stat.ps.instructions
> 2710577 ± 3% +388.7% 13246535 ± 2% +308.6% 11076222 ± 44% perf-stat.ps.minor-faults
> 2710577 ± 3% +388.7% 13246536 ± 2% +308.6% 11076222 ± 44% perf-stat.ps.page-faults
> 3.886e+13 ± 2% -52.4% 1.849e+13 -58.3% 1.619e+13 ± 37% perf-stat.total.instructions
> 64052898 ± 5% -96.2% 2460331 ±166% -93.1% 4432025 ±129% sched_debug.cfs_rq:/.avg_vruntime.avg
> 95701822 ± 7% -85.1% 14268127 ±116% -60.2% 38124846 ±118% sched_debug.cfs_rq:/.avg_vruntime.max
> 43098762 ± 6% -96.0% 1715136 ±173% -93.3% 2867368 ±131% sched_debug.cfs_rq:/.avg_vruntime.min
> 9223270 ± 9% -84.2% 1457904 ±122% -61.0% 3595639 ±113% sched_debug.cfs_rq:/.avg_vruntime.stddev
> 0.00 ± 22% -80.1% 0.00 ±185% -86.8% 0.00 ±173% sched_debug.cfs_rq:/.h_nr_delayed.avg
> 0.69 ± 8% -73.0% 0.19 ±185% -82.0% 0.12 ±173% sched_debug.cfs_rq:/.h_nr_delayed.max
> 0.05 ± 12% -76.3% 0.01 ±185% -84.2% 0.01 ±173% sched_debug.cfs_rq:/.h_nr_delayed.stddev
> 0.78 ± 2% -77.0% 0.18 ±130% -71.9% 0.22 ±107% sched_debug.cfs_rq:/.h_nr_running.avg
> 43049468 ± 22% -89.3% 4590302 ±180% -89.0% 4726833 ±129% sched_debug.cfs_rq:/.left_deadline.max
> 3836405 ± 37% -85.6% 550773 ±176% -77.5% 864733 ±132% sched_debug.cfs_rq:/.left_deadline.stddev
> 43049467 ± 22% -89.3% 4590279 ±180% -89.0% 4726820 ±129% sched_debug.cfs_rq:/.left_vruntime.max
> 3836405 ± 37% -85.6% 550772 ±176% -77.5% 862614 ±132% sched_debug.cfs_rq:/.left_vruntime.stddev
> 64052901 ± 5% -96.2% 2460341 ±166% -93.1% 4432036 ±129% sched_debug.cfs_rq:/.min_vruntime.avg
> 95701822 ± 7% -85.1% 14268127 ±116% -60.2% 38124846 ±118% sched_debug.cfs_rq:/.min_vruntime.max
> 43098762 ± 6% -96.0% 1715136 ±173% -93.3% 2867368 ±131% sched_debug.cfs_rq:/.min_vruntime.min
> 9223270 ± 9% -84.2% 1457902 ±122% -61.0% 3595638 ±113% sched_debug.cfs_rq:/.min_vruntime.stddev
> 0.77 ± 2% -77.4% 0.17 ±128% -72.3% 0.21 ±107% sched_debug.cfs_rq:/.nr_running.avg
> 1.61 ± 24% +396.0% 7.96 ± 62% +355.1% 7.31 ± 52% sched_debug.cfs_rq:/.removed.runnable_avg.avg
> 86.69 +424.4% 454.62 ± 24% +400.6% 433.98 ± 26% sched_debug.cfs_rq:/.removed.runnable_avg.max
> 11.14 ± 13% +409.8% 56.79 ± 35% +373.6% 52.77 ± 34% sched_debug.cfs_rq:/.removed.runnable_avg.stddev
> 1.61 ± 24% +396.0% 7.96 ± 62% +355.1% 7.31 ± 52% sched_debug.cfs_rq:/.removed.util_avg.avg
> 86.69 +424.4% 454.62 ± 24% +400.6% 433.98 ± 26% sched_debug.cfs_rq:/.removed.util_avg.max
> 11.14 ± 13% +409.8% 56.79 ± 35% +373.6% 52.77 ± 34% sched_debug.cfs_rq:/.removed.util_avg.stddev
> 43049467 ± 22% -89.3% 4590282 ±180% -89.0% 4726821 ±129% sched_debug.cfs_rq:/.right_vruntime.max
> 3836405 ± 37% -85.6% 550772 ±176% -77.5% 862614 ±132% sched_debug.cfs_rq:/.right_vruntime.stddev
> 286633 ± 43% +262.3% 1038592 ± 36% +188.3% 826260 ± 58% sched_debug.cfs_rq:/.runnable_avg.avg
> 34728895 ± 30% +349.2% 1.56e+08 ± 26% +293.3% 1.366e+08 ± 60% sched_debug.cfs_rq:/.runnable_avg.max
> 2845573 ± 30% +325.9% 12119045 ± 26% +251.3% 9995202 ± 55% sched_debug.cfs_rq:/.runnable_avg.stddev
> 769.03 -69.9% 231.86 ± 84% -66.3% 259.37 ± 72% sched_debug.cfs_rq:/.util_avg.avg
> 1621 ± 5% -31.5% 1111 ± 8% -35.4% 1048 ± 8% sched_debug.cfs_rq:/.util_avg.max
> 159.12 ± 8% +22.3% 194.66 ± 12% +35.0% 214.82 ± 14% sched_debug.cfs_rq:/.util_avg.stddev
> 724.17 ± 2% -89.6% 75.66 ±147% -88.3% 84.74 ±123% sched_debug.cfs_rq:/.util_est.avg
> 1360 ± 15% -39.2% 826.88 ± 37% -29.0% 965.90 ± 48% sched_debug.cfs_rq:/.util_est.max
> 766944 ± 3% +18.1% 905901 +21.7% 933047 ± 2% sched_debug.cpu.avg_idle.avg
> 1067639 ± 5% +30.0% 1387534 ± 16% +38.2% 1475131 ± 15% sched_debug.cpu.avg_idle.max
> 321459 ± 2% -35.6% 207172 ± 10% -33.5% 213764 ± 15% sched_debug.cpu.avg_idle.stddev
> 195573 -72.7% 53401 ± 24% -68.5% 61507 ± 35% sched_debug.cpu.clock.avg
> 195596 -72.7% 53442 ± 24% -68.5% 61565 ± 35% sched_debug.cpu.clock.max
> 195548 -72.7% 53352 ± 24% -68.6% 61431 ± 35% sched_debug.cpu.clock.min
> 194424 -72.6% 53229 ± 24% -68.5% 61304 ± 35% sched_debug.cpu.clock_task.avg
> 194608 -72.6% 53383 ± 24% -68.4% 61478 ± 34% sched_debug.cpu.clock_task.max
> 181834 -77.5% 40964 ± 31% -73.0% 49012 ± 43% sched_debug.cpu.clock_task.min
> 4241 ± 2% -80.6% 821.65 ±142% -77.1% 971.85 ±116% sched_debug.cpu.curr->pid.avg
> 9799 ± 2% -55.4% 4365 ± 17% -51.6% 4747 ± 22% sched_debug.cpu.curr->pid.max
> 1365 ± 10% -48.0% 709.44 ± 5% -39.9% 820.19 ± 24% sched_debug.cpu.curr->pid.stddev
> 537665 ± 4% +31.2% 705318 ± 14% +44.0% 774261 ± 15% sched_debug.cpu.max_idle_balance_cost.max
> 3119 ± 56% +579.1% 21184 ± 39% +1048.3% 35821 ± 65% sched_debug.cpu.max_idle_balance_cost.stddev
> 0.78 ± 2% -76.3% 0.18 ±135% -72.0% 0.22 ±114% sched_debug.cpu.nr_running.avg
> 25773 ± 5% -96.1% 1007 ± 41% -95.2% 1246 ± 53% sched_debug.cpu.nr_switches.avg
> 48669 ± 10% -76.5% 11448 ± 13% -66.5% 16288 ± 70% sched_debug.cpu.nr_switches.max
> 19006 ± 7% -98.6% 258.81 ± 64% -98.4% 311.75 ± 58% sched_debug.cpu.nr_switches.min
> 4142 ± 8% -66.3% 1396 ± 17% -58.3% 1726 ± 51% sched_debug.cpu.nr_switches.stddev
> 0.07 ± 23% -92.9% 0.01 ± 41% -94.3% 0.00 ± 46% sched_debug.cpu.nr_uninterruptible.avg
> 240.19 ± 16% -82.1% 42.94 ± 41% -84.0% 38.50 ± 19% sched_debug.cpu.nr_uninterruptible.max
> -77.92 -88.1% -9.25 -84.9% -11.77 sched_debug.cpu.nr_uninterruptible.min
> 37.87 ± 5% -85.8% 5.36 ± 13% -85.3% 5.57 ± 5% sched_debug.cpu.nr_uninterruptible.stddev
> 195549 -72.7% 53356 ± 24% -68.6% 61438 ± 35% sched_debug.cpu_clk
> 194699 -73.0% 52506 ± 25% -68.9% 60588 ± 35% sched_debug.ktime
> 0.00 -100.0% 0.00 -62.5% 0.00 ±264% sched_debug.rt_rq:.rt_nr_running.avg
> 0.17 -100.0% 0.00 -62.5% 0.06 ±264% sched_debug.rt_rq:.rt_nr_running.max
> 0.01 -100.0% 0.00 -62.5% 0.00 ±264% sched_debug.rt_rq:.rt_nr_running.stddev
> 196368 -72.4% 54191 ± 24% -68.3% 62327 ± 34% sched_debug.sched_clk
> 0.17 ±142% -100.0% 0.00 -97.8% 0.00 ±264% perf-sched.sch_delay.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
> 0.19 ± 34% -51.3% 0.09 ± 37% -76.7% 0.04 ±110% perf-sched.sch_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 0.14 ± 55% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
> 0.14 ± 73% -82.5% 0.03 ±168% -64.1% 0.05 ±177% perf-sched.sch_delay.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
> 0.11 ± 59% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 0.04 ±132% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
> 0.02 ± 31% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
> 0.00 ±223% +51950.0% 0.26 ±212% +6325.0% 0.03 ±124% perf-sched.sch_delay.avg.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit
> 0.25 ± 59% -100.0% 0.00 -64.9% 0.09 ±253% perf-sched.sch_delay.avg.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
> 0.12 ±145% -99.1% 0.00 ±141% -99.5% 0.00 ±264% perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
> 0.04 ± 55% +99.5% 0.08 ±254% -92.0% 0.00 ±103% perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 0.25 ± 41% -81.6% 0.05 ± 69% -94.4% 0.01 ± 69% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
> 0.11 ± 59% -87.1% 0.01 ±198% -96.2% 0.00 ±128% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
> 0.40 ± 50% -97.8% 0.01 ± 30% -97.2% 0.01 ± 45% perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 2.25 ±138% -99.6% 0.01 ± 7% -63.9% 0.81 ±261% perf-sched.sch_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
> 0.32 ±104% -97.3% 0.01 ± 38% -97.7% 0.01 ± 61% perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
> 0.12 ± 21% -61.6% 0.04 ±233% -85.7% 0.02 ±190% perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.01 ± 12% -34.9% 0.01 ± 18% +722.2% 0.07 ±251% perf-sched.sch_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 0.01 ± 42% -41.4% 0.00 ± 72% -76.6% 0.00 ± 77% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown].[unknown]
> 0.01 ± 20% -100.0% 0.00 -96.4% 0.00 ±264% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
> 0.19 ±185% -95.6% 0.01 ± 44% +266.3% 0.70 ±261% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 0.07 ± 20% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 0.26 ± 17% -98.8% 0.00 ± 10% -98.9% 0.00 ± 39% perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
> 0.03 ± 51% -69.7% 0.01 ± 67% -83.7% 0.01 ± 15% perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
> 0.01 ± 55% +721.9% 0.10 ± 29% +1608.3% 0.20 ±227% perf-sched.sch_delay.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 0.01 ±128% -83.6% 0.00 ± 20% -86.2% 0.00 ± 43% perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
> 0.06 ± 31% +1921.5% 1.23 ±165% +13539.3% 8.30 ±201% perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 1.00 ±151% -100.0% 0.00 -99.6% 0.00 ±264% perf-sched.sch_delay.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
> 25.45 ± 94% -98.6% 0.36 ± 61% -99.4% 0.15 ±143% perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 4.56 ± 67% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
> 3.55 ± 97% -98.9% 0.04 ±189% -98.5% 0.05 ±177% perf-sched.sch_delay.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
> 2.13 ± 67% -77.2% 0.49 ± 56% -88.8% 0.24 ±147% perf-sched.sch_delay.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
> 3.16 ± 78% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 0.30 ±159% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
> 1.61 ±100% -76.7% 0.38 ± 72% -91.7% 0.13 ±145% perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 0.03 ± 86% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
> 0.00 ±223% +3.2e+06% 15.79 ±259% +44450.0% 0.22 ±132% perf-sched.sch_delay.max.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit
> 3.09 ± 45% -100.0% 0.00 -94.6% 0.17 ±259% perf-sched.sch_delay.max.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
> 3.51 ± 21% -86.1% 0.49 ± 72% -90.7% 0.33 ±127% perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
> 0.83 ±160% -99.7% 0.00 ±141% -99.9% 0.00 ±264% perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
> 0.09 ± 31% +179.7% 0.25 ±258% -91.5% 0.01 ±132% perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 3.59 ± 11% -92.0% 0.29 ±165% -99.2% 0.03 ±118% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
> 1.60 ± 69% -95.7% 0.07 ±243% -99.0% 0.02 ±210% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
> 0.81 ± 43% -98.5% 0.01 ± 43% -98.3% 0.01 ± 41% perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 1.02 ± 88% -98.1% 0.02 ± 47% -98.7% 0.01 ± 71% perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
> 9.68 ± 32% -92.2% 0.76 ± 72% -78.1% 2.12 ±187% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 0.01 ± 49% -51.9% 0.00 ± 72% -80.8% 0.00 ± 77% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown].[unknown]
> 12.26 ±109% -92.9% 0.87 ±101% -86.9% 1.61 ±225% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
> 5.60 ±139% -97.6% 0.13 ±132% -99.3% 0.04 ±255% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
> 0.03 ±106% -100.0% 0.00 -99.1% 0.00 ±264% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
> 2.11 ± 61% -85.5% 0.31 ± 85% -96.0% 0.08 ±124% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
> 37.84 ± 47% -100.0% 0.00 -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 4.68 ± 36% -99.8% 0.01 ± 65% -99.8% 0.01 ± 77% perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
> 7.56 ± 74% -51.5% 3.67 ±147% -99.8% 0.02 ± 54% perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
> 0.36 ±186% -96.3% 0.01 ± 90% -97.9% 0.01 ± 59% perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
> 97903 ± 4% -38.3% 60433 ± 29% -71.4% 27976 ±109% perf-sched.total_wait_and_delay.count.ms
> 3.97 ± 6% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 302.41 ± 5% -27.4% 219.54 ± 14% -10.8% 269.81 ± 60% perf-sched.wait_and_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 1.48 ± 6% -90.9% 0.14 ± 79% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 327.16 ± 9% -46.6% 174.81 ± 24% -38.4% 201.64 ± 71% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 369.37 ± 2% -75.3% 91.05 ± 35% -77.7% 82.29 ±119% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 0.96 ± 6% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 187.66 +120.6% 413.97 ± 14% +116.9% 407.06 ± 43% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 1831 ± 9% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 6.17 ± 45% -79.7% 1.25 ±142% -91.9% 0.50 ±264% perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 14.33 ± 5% +13.4% 16.25 ± 23% -58.1% 6.00 ± 66% perf-sched.wait_and_delay.count.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 810.00 ± 10% -38.0% 502.25 ± 92% -100.0% 0.00 perf-sched.wait_and_delay.count.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 40.50 ± 8% +245.7% 140.00 ± 23% +72.5% 69.88 ± 91% perf-sched.wait_and_delay.count.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 13.17 ± 2% +624.4% 95.38 ± 19% +347.2% 58.88 ± 78% perf-sched.wait_and_delay.count.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 73021 ± 3% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 11323 ± 3% -75.9% 2725 ± 28% -86.4% 1536 ± 34% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 1887 ± 45% -96.1% 73.88 ± 78% -98.5% 28.75 ±120% perf-sched.wait_and_delay.count.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 1238 -34.5% 811.25 ± 13% -58.6% 512.62 ± 49% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 35.19 ± 57% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 20.79 ± 19% -95.9% 0.84 ± 93% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 1240 ± 20% -14.4% 1062 ± 10% -25.2% 928.21 ± 40% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 500.34 +31.2% 656.38 ± 39% -15.0% 425.46 ± 61% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 58.83 ± 39% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 1237 ± 34% +151.7% 3114 ± 25% +51.6% 1876 ± 64% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 49.27 ±119% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
> 58.17 ±187% -100.0% 0.00 -100.0% 0.00 ±264% perf-sched.wait_time.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
> 3.78 ± 5% -97.6% 0.09 ± 37% -98.8% 0.04 ±111% perf-sched.wait_time.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 2.99 ± 4% +15.4% 3.45 ± 10% +28.8% 3.85 ± 54% perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
> 3.92 ± 5% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
> 4.71 ± 8% -99.5% 0.02 ±170% -98.9% 0.05 ±177% perf-sched.wait_time.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
> 1.67 ± 20% -92.7% 0.12 ± 30% -96.8% 0.05 ±130% perf-sched.wait_time.avg.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
> 2.10 ± 27% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 0.01 ± 44% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
> 1.67 ± 21% -94.3% 0.10 ± 35% -97.0% 0.05 ±137% perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 0.04 ±133% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
> 67.14 ± 73% +75.6% 117.89 ±108% -92.8% 4.82 ±259% perf-sched.wait_time.avg.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
> 1.65 ± 67% -95.8% 0.07 ±128% -99.2% 0.01 ±175% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
> 2.30 ± 14% -95.5% 0.10 ± 42% -96.4% 0.08 ±108% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
> 2.00 ± 74% +2917.4% 60.44 ± 33% +1369.3% 29.43 ± 74% perf-sched.wait_time.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
> 29.19 ± 5% -38.5% 17.96 ± 28% -49.0% 14.89 ± 54% perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
> 0.37 ± 30% +5524.5% 20.95 ± 30% +2028.0% 7.93 ±117% perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 302.40 ± 5% -27.4% 219.53 ± 14% -10.8% 269.75 ± 60% perf-sched.wait_time.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 1.40 ± 6% -92.7% 0.10 ± 18% -95.4% 0.06 ±109% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 0.72 ±220% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
> 326.84 ± 9% -46.6% 174.54 ± 24% -38.6% 200.64 ± 72% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 369.18 ± 2% -75.3% 91.04 ± 35% -74.2% 95.16 ± 98% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 0.89 ± 6% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 187.58 +120.6% 413.77 ± 14% +116.9% 406.79 ± 43% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 2.36 ± 29% +1759.6% 43.80 ± 33% +3763.5% 90.99 ±115% perf-sched.wait_time.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 0.01 ±156% -97.9% 0.00 ±264% -98.9% 0.00 ±264% perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
> 750.01 -14.5% 641.50 ± 14% -41.1% 442.13 ± 58% perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 340.69 ±135% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
> 535.09 ±128% -100.0% 0.00 -100.0% 0.00 ±264% perf-sched.wait_time.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
> 22.04 ± 32% -98.4% 0.36 ± 61% -99.3% 0.15 ±143% perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 13.57 ± 17% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
> 13.54 ± 10% -99.7% 0.04 ±189% -99.6% 0.05 ±177% perf-sched.wait_time.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
> 10.17 ± 19% -95.2% 0.49 ± 56% -97.7% 0.24 ±147% perf-sched.wait_time.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
> 11.35 ± 25% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 0.01 ± 32% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
> 10.62 ± 9% -96.5% 0.38 ± 72% -98.7% 0.13 ±145% perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 0.20 ±199% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
> 1559 ± 64% -92.3% 120.30 ±109% -99.4% 9.63 ±259% perf-sched.wait_time.max.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
> 6.93 ± 53% -98.1% 0.13 ± 99% -99.8% 0.01 ±175% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
> 14.42 ± 22% -96.6% 0.49 ± 72% -97.7% 0.33 ±127% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
> 4.00 ± 74% +19182.5% 772.23 ± 40% +7266.0% 295.00 ± 92% perf-sched.wait_time.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
> 10.75 ± 98% +6512.2% 710.88 ± 56% +2526.4% 282.37 ±130% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 15.80 ± 8% -95.2% 0.76 ± 72% -86.6% 2.12 ±187% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 11.64 ± 61% -98.9% 0.13 ±132% -99.7% 0.04 ±255% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
> 2.94 ±213% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
> 175.70 ±210% -64.6% 62.26 ±263% -99.8% 0.31 ±116% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
> 1240 ± 20% -14.3% 1062 ± 10% -25.2% 928.20 ± 40% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 500.11 +31.2% 656.37 ± 39% -2.4% 487.96 ± 41% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 32.65 ± 33% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 1237 ± 34% +151.6% 3113 ± 25% +49.0% 1844 ± 63% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 0.36 ±190% -97.2% 0.01 ±127% -98.5% 0.01 ± 88% perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
> 95.59 -95.6 0.00 -95.6 0.00 perf-profile.calltrace.cycles-pp.__mmap
> 95.54 -95.5 0.00 -95.5 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
> 95.54 -95.5 0.00 -95.5 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap
> 94.54 -94.5 0.00 -94.5 0.00 perf-profile.calltrace.cycles-pp.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
> 94.46 -94.0 0.41 ±138% -93.9 0.57 ±103% perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
> 94.14 -93.7 0.40 ±136% -93.6 0.50 ± 79% perf-profile.calltrace.cycles-pp.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff
> 93.79 -93.5 0.31 ±134% -93.2 0.58 ±111% perf-profile.calltrace.cycles-pp.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff
> 93.40 -93.4 0.00 -93.4 0.00 perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 93.33 -93.3 0.00 -93.3 0.00 perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma
> 93.44 -93.3 0.14 ±264% -93.4 0.00 perf-profile.calltrace.cycles-pp.down_write.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap
> 94.45 -93.0 1.42 ± 60% -92.9 1.51 ± 51% perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 94.25 -92.9 1.33 ± 61% -92.8 1.43 ± 57% perf-profile.calltrace.cycles-pp.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 92.89 -92.9 0.00 -92.9 0.00 perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file
> 0.00 +0.3 0.29 ±129% +1.1 1.10 ± 27% perf-profile.calltrace.cycles-pp.do_open.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat
> 0.00 +0.3 0.32 ±129% +1.7 1.70 ± 39% perf-profile.calltrace.cycles-pp.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin.generic_perform_write.shmem_file_write_iter
> 0.00 +0.3 0.32 ±129% +1.7 1.74 ± 40% perf-profile.calltrace.cycles-pp.shmem_get_folio_gfp.shmem_write_begin.generic_perform_write.shmem_file_write_iter.vfs_write
> 0.00 +0.5 0.49 ± 78% +1.7 1.74 ± 40% perf-profile.calltrace.cycles-pp.shmem_write_begin.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
> 0.00 +1.1 1.09 ± 33% +0.4 0.44 ±177% perf-profile.calltrace.cycles-pp.dup_mmap.dup_mm.copy_process.kernel_clone.__do_sys_clone
> 0.00 +1.3 1.32 ± 54% +1.4 1.36 ± 33% perf-profile.calltrace.cycles-pp.filp_close.put_files_struct.do_exit.do_group_exit.get_signal
> 0.00 +1.3 1.32 ± 54% +1.4 1.36 ± 33% perf-profile.calltrace.cycles-pp.put_files_struct.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
> 0.00 +1.4 1.37 ± 49% +1.8 1.77 ± 50% perf-profile.calltrace.cycles-pp.setlocale
> 0.00 +1.4 1.39 ± 70% +1.8 1.80 ± 48% perf-profile.calltrace.cycles-pp.seq_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +1.4 1.39 ± 70% +1.8 1.80 ± 48% perf-profile.calltrace.cycles-pp.seq_read_iter.seq_read.vfs_read.ksys_read.do_syscall_64
> 0.00 +1.5 1.55 ± 63% +1.6 1.62 ± 37% perf-profile.calltrace.cycles-pp.do_read_fault.do_pte_missing.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
> 0.00 +1.6 1.60 ± 57% +1.6 1.63 ± 87% perf-profile.calltrace.cycles-pp.swevent_hlist_put_cpu.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release
> 0.00 +1.6 1.64 ± 47% +0.9 0.90 ±101% perf-profile.calltrace.cycles-pp.handle_softirqs.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt
> 0.00 +1.6 1.64 ± 47% +1.0 1.02 ± 83% perf-profile.calltrace.cycles-pp.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry
> 0.00 +1.6 1.65 ± 43% +1.1 1.15 ± 76% perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +1.8 1.76 ± 44% +1.1 1.15 ± 76% perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +1.9 1.93 ± 26% +1.1 1.11 ±127% perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64
> 0.00 +2.0 2.04 ± 66% +3.6 3.65 ± 42% perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64
> 0.00 +2.1 2.12 ± 58% +3.6 3.65 ± 42% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64
> 0.00 +2.1 2.12 ± 58% +3.6 3.65 ± 42% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.open64
> 0.00 +2.1 2.12 ± 58% +3.7 3.71 ± 40% perf-profile.calltrace.cycles-pp.open64
> 0.00 +2.2 2.16 ± 44% +1.6 1.62 ± 37% perf-profile.calltrace.cycles-pp.do_pte_missing.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
> 0.00 +2.2 2.20 ± 74% +3.6 3.65 ± 42% perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64
> 0.00 +2.2 2.23 ± 33% +1.4 1.40 ± 99% perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +2.3 2.34 ±103% +5.1 5.09 ± 64% perf-profile.calltrace.cycles-pp.__cmd_record.cmd_record.perf_c2c__record.run_builtin.handle_internal_command
> 0.00 +2.3 2.34 ±103% +5.1 5.09 ± 64% perf-profile.calltrace.cycles-pp.cmd_record.perf_c2c__record.run_builtin.handle_internal_command.main
> 0.00 +2.3 2.34 ±103% +5.1 5.09 ± 64% perf-profile.calltrace.cycles-pp.perf_c2c__record.run_builtin.handle_internal_command.main
> 0.00 +2.4 2.37 ± 36% +1.9 1.93 ± 35% perf-profile.calltrace.cycles-pp.zap_present_ptes.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
> 0.00 +2.5 2.48 ± 32% +2.4 2.45 ± 60% perf-profile.calltrace.cycles-pp.get_cpu_sleep_time_us.get_idle_time.uptime_proc_show.seq_read_iter.vfs_read
> 0.00 +2.5 2.50 ± 45% +1.2 1.21 ± 73% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
> 0.00 +2.5 2.54 ± 47% +1.3 1.28 ± 61% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group
> 0.00 +2.5 2.54 ± 47% +1.3 1.28 ± 61% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
> 0.00 +2.6 2.62 ± 35% +1.6 1.57 ± 91% perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
> 0.00 +2.6 2.62 ± 35% +1.6 1.57 ± 91% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
> 0.00 +2.6 2.62 ± 35% +1.6 1.57 ± 91% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe._Fork
> 0.00 +2.6 2.62 ± 35% +1.6 1.57 ± 91% perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
> 0.00 +2.7 2.67 ± 54% +2.6 2.59 ± 40% perf-profile.calltrace.cycles-pp.load_elf_binary.search_binary_handler.exec_binprm.bprm_execve.do_execveat_common
> 0.00 +2.7 2.68 ± 35% +3.0 3.02 ± 45% perf-profile.calltrace.cycles-pp.get_idle_time.uptime_proc_show.seq_read_iter.vfs_read.ksys_read
> 0.00 +2.8 2.77 ± 33% +4.2 4.17 ± 35% perf-profile.calltrace.cycles-pp.uptime_proc_show.seq_read_iter.vfs_read.ksys_read.do_syscall_64
> 0.00 +2.8 2.82 ± 32% +1.8 1.83 ± 85% perf-profile.calltrace.cycles-pp._Fork
> 0.00 +2.8 2.83 ± 48% +2.6 2.59 ± 40% perf-profile.calltrace.cycles-pp.search_binary_handler.exec_binprm.bprm_execve.do_execveat_common.__x64_sys_execve
> 0.00 +2.8 2.83 ± 48% +2.7 2.68 ± 42% perf-profile.calltrace.cycles-pp.exec_binprm.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64
> 0.00 +2.8 2.84 ± 45% +1.2 1.21 ± 73% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
> 0.00 +2.8 2.84 ± 45% +1.2 1.21 ± 73% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
> 0.00 +2.9 2.89 ± 39% +3.1 3.14 ± 39% perf-profile.calltrace.cycles-pp.event_function_call.perf_event_release_kernel.perf_release.__fput.task_work_run
> 0.00 +2.9 2.89 ± 39% +3.1 3.14 ± 39% perf-profile.calltrace.cycles-pp.smp_call_function_single.event_function_call.perf_event_release_kernel.perf_release.__fput
> 0.00 +3.1 3.10 ± 64% +0.9 0.91 ±264% perf-profile.calltrace.cycles-pp.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +3.1 3.10 ± 64% +0.9 0.91 ±264% perf-profile.calltrace.cycles-pp.seq_read_iter.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64
> 0.00 +3.1 3.13 ± 33% +1.7 1.68 ± 77% perf-profile.calltrace.cycles-pp.asm_exc_page_fault
> 0.00 +3.2 3.18 ± 37% +4.3 4.31 ± 34% perf-profile.calltrace.cycles-pp.seq_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +3.2 3.20 ± 28% +3.0 3.02 ± 73% perf-profile.calltrace.cycles-pp.mutex_unlock.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release
> 0.00 +3.2 3.24 ± 39% +2.8 2.85 ± 49% perf-profile.calltrace.cycles-pp.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +3.2 3.24 ± 36% +2.0 2.00 ± 56% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +3.2 3.24 ± 36% +2.0 2.00 ± 56% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64
> 0.00 +3.2 3.24 ± 36% +2.0 2.00 ± 56% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +3.2 3.24 ± 36% +2.0 2.00 ± 56% perf-profile.calltrace.cycles-pp.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +3.8 3.85 ± 39% +3.3 3.25 ± 47% perf-profile.calltrace.cycles-pp.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> 0.00 +3.8 3.85 ± 39% +3.3 3.25 ± 47% perf-profile.calltrace.cycles-pp.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> 0.00 +3.8 3.85 ± 39% +3.3 3.25 ± 47% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> 0.00 +3.8 3.85 ± 39% +3.3 3.25 ± 47% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.execve
> 0.00 +3.8 3.85 ± 39% +3.3 3.29 ± 47% perf-profile.calltrace.cycles-pp.execve
> 0.00 +4.0 4.04 ± 43% +5.2 5.21 ± 49% perf-profile.calltrace.cycles-pp.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +4.0 4.04 ± 43% +5.2 5.21 ± 49% perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64
> 0.00 +4.1 4.10 ± 30% +2.6 2.56 ± 28% perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.exit_mmap.__mmput.exit_mm
> 0.00 +4.2 4.18 ± 31% +2.8 2.82 ± 21% perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap.__mmput
> 0.00 +4.2 4.18 ± 31% +2.8 2.82 ± 21% perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap
> 0.00 +4.2 4.20 ± 28% +2.7 2.68 ± 34% perf-profile.calltrace.cycles-pp.unmap_vmas.exit_mmap.__mmput.exit_mm.do_exit
> 0.00 +4.2 4.25 ± 65% +8.0 7.98 ± 43% perf-profile.calltrace.cycles-pp.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64
> 0.00 +4.3 4.27 ± 26% +3.2 3.23 ± 34% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
> 0.00 +4.3 4.30 ± 22% +3.9 3.95 ± 32% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.get_signal
> 0.00 +4.3 4.30 ± 22% +3.9 3.95 ± 32% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
> 0.00 +4.5 4.46 ± 59% +8.1 8.07 ± 42% perf-profile.calltrace.cycles-pp.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +4.6 4.57 ± 58% +8.1 8.07 ± 42% perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen
> 0.00 +4.7 4.68 ± 55% +8.1 8.12 ± 43% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn
> 0.00 +4.7 4.68 ± 55% +8.1 8.12 ± 43% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn.perf_mmap__push
> 0.00 +4.7 4.68 ± 55% +8.2 8.16 ± 44% perf-profile.calltrace.cycles-pp.write.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist
> 0.00 +4.7 4.68 ± 55% +8.4 8.39 ± 39% perf-profile.calltrace.cycles-pp.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record
> 0.00 +4.7 4.68 ± 55% +8.6 8.61 ± 38% perf-profile.calltrace.cycles-pp.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record
> 0.00 +4.9 4.90 ± 57% +10.3 10.28 ± 65% perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
> 0.00 +4.9 4.92 ± 26% +4.6 4.56 ± 47% perf-profile.calltrace.cycles-pp.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release.__fput
> 0.00 +5.0 4.99 ±100% +2.6 2.64 ±101% perf-profile.calltrace.cycles-pp.__intel_pmu_enable_all.perf_rotate_context.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt
> 0.00 +5.0 4.99 ±100% +2.6 2.64 ±101% perf-profile.calltrace.cycles-pp.perf_rotate_context.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt
> 0.00 +5.1 5.08 ±102% +2.6 2.64 ±101% perf-profile.calltrace.cycles-pp.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt
> 0.00 +5.1 5.14 ± 28% +6.0 6.01 ± 41% perf-profile.calltrace.cycles-pp.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin
> 0.00 +5.1 5.14 ± 28% +6.2 6.16 ± 39% perf-profile.calltrace.cycles-pp.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin.handle_internal_command
> 0.00 +5.4 5.43 ± 25% +5.0 4.97 ± 45% perf-profile.calltrace.cycles-pp._free_event.perf_event_release_kernel.perf_release.__fput.task_work_run
> 0.00 +5.8 5.82 ± 94% +4.2 4.21 ± 49% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt
> 0.00 +5.8 5.82 ± 94% +4.3 4.35 ± 53% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry
> 0.00 +6.1 6.07 ± 90% +4.3 4.32 ± 58% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
> 0.00 +6.6 6.62 ± 24% +7.0 6.99 ± 41% perf-profile.calltrace.cycles-pp.__cmd_record.cmd_record.run_builtin.handle_internal_command.main
> 0.00 +6.6 6.62 ± 24% +7.0 6.99 ± 41% perf-profile.calltrace.cycles-pp.cmd_record.run_builtin.handle_internal_command.main
> 0.00 +6.8 6.76 ± 18% +5.2 5.23 ± 25% perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit
> 0.00 +7.6 7.56 ± 76% +6.0 5.99 ± 38% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter
> 0.00 +8.0 8.03 ± 27% +7.4 7.37 ± 52% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
> 0.00 +8.0 8.03 ± 27% +7.4 7.37 ± 52% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
> 0.00 +8.0 8.05 ± 68% +6.3 6.27 ± 37% perf-profile.calltrace.cycles-pp.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter
> 0.00 +8.1 8.13 ± 28% +7.4 7.37 ± 52% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
> 0.00 +8.1 8.13 ± 28% +7.4 7.37 ± 52% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
> 0.00 +8.1 8.13 ± 28% +7.4 7.37 ± 52% perf-profile.calltrace.cycles-pp.read
> 0.00 +9.1 9.05 ± 35% +13.9 13.88 ± 19% perf-profile.calltrace.cycles-pp.handle_internal_command.main
> 0.00 +9.1 9.05 ± 35% +13.9 13.88 ± 19% perf-profile.calltrace.cycles-pp.main
> 0.00 +9.1 9.05 ± 35% +13.9 13.88 ± 19% perf-profile.calltrace.cycles-pp.run_builtin.handle_internal_command.main
> 0.00 +9.3 9.26 ± 30% +9.0 8.96 ± 31% perf-profile.calltrace.cycles-pp.perf_event_release_kernel.perf_release.__fput.task_work_run.do_exit
> 0.00 +9.3 9.26 ± 30% +9.0 8.96 ± 31% perf-profile.calltrace.cycles-pp.perf_release.__fput.task_work_run.do_exit.do_group_exit
> 0.00 +10.1 10.14 ± 28% +10.0 10.04 ± 34% perf-profile.calltrace.cycles-pp.__fput.task_work_run.do_exit.do_group_exit.get_signal
> 0.00 +10.2 10.23 ± 27% +10.7 10.65 ± 35% perf-profile.calltrace.cycles-pp.task_work_run.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
> 0.00 +11.0 10.98 ± 55% +13.0 13.00 ± 27% perf-profile.calltrace.cycles-pp.asm_sysvec_reschedule_ipi.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
> 0.00 +20.6 20.64 ± 30% +19.5 19.49 ± 43% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +20.6 20.64 ± 30% +19.5 19.49 ± 43% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
> 1.21 ± 3% +36.6 37.80 ± 12% +34.1 35.32 ± 11% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
> 1.21 ± 3% +36.6 37.80 ± 12% +34.4 35.62 ± 11% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.common_startup_64
> 1.22 ± 3% +36.8 38.00 ± 13% +34.8 36.05 ± 11% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.common_startup_64
> 1.22 ± 3% +36.9 38.10 ± 13% +34.8 36.05 ± 11% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.common_startup_64
> 1.22 ± 3% +36.9 38.10 ± 13% +34.8 36.05 ± 11% perf-profile.calltrace.cycles-pp.start_secondary.common_startup_64
> 1.21 ± 3% +37.2 38.43 ± 11% +34.2 35.40 ± 8% perf-profile.calltrace.cycles-pp.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
> 1.21 ± 3% +37.2 38.43 ± 11% +34.2 35.40 ± 8% perf-profile.calltrace.cycles-pp.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
> 1.21 ± 3% +37.3 38.54 ± 12% +34.7 35.87 ± 10% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
> 1.22 ± 3% +37.6 38.84 ± 12% +35.4 36.60 ± 11% perf-profile.calltrace.cycles-pp.common_startup_64
> 2.19 ± 3% +53.9 56.10 ± 19% +48.4 50.63 ± 13% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
> 95.60 -95.2 0.41 ±138% -94.9 0.72 ± 95% perf-profile.children.cycles-pp.__mmap
> 94.14 -93.7 0.49 ±130% -92.9 1.21 ± 33% perf-profile.children.cycles-pp.__mmap_new_vma
> 93.79 -93.5 0.31 ±134% -93.1 0.71 ± 78% perf-profile.children.cycles-pp.vma_link_file
> 93.40 -93.4 0.00 -93.4 0.00 perf-profile.children.cycles-pp.rwsem_down_write_slowpath
> 93.33 -93.3 0.00 -93.3 0.00 perf-profile.children.cycles-pp.rwsem_optimistic_spin
> 94.55 -93.1 1.42 ± 60% -93.0 1.55 ± 50% perf-profile.children.cycles-pp.ksys_mmap_pgoff
> 92.91 -92.9 0.00 -92.9 0.00 perf-profile.children.cycles-pp.osq_lock
> 93.44 -92.7 0.75 ±109% -93.4 0.06 ±264% perf-profile.children.cycles-pp.down_write
> 94.46 -92.6 1.84 ± 34% -92.0 2.48 ± 28% perf-profile.children.cycles-pp.vm_mmap_pgoff
> 94.45 -92.6 1.84 ± 34% -92.0 2.48 ± 28% perf-profile.children.cycles-pp.do_mmap
> 94.25 -92.6 1.66 ± 37% -91.9 2.40 ± 30% perf-profile.children.cycles-pp.__mmap_region
> 95.58 -44.8 50.78 ± 11% -42.8 52.76 ± 11% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
> 95.58 -44.8 50.78 ± 11% -42.8 52.76 ± 11% perf-profile.children.cycles-pp.do_syscall_64
> 0.00 +0.1 0.09 ±264% +1.0 0.96 ± 46% perf-profile.children.cycles-pp.kcpustat_cpu_fetch
> 0.25 ± 3% +0.2 0.45 ±133% +0.7 0.92 ± 41% perf-profile.children.cycles-pp.vma_interval_tree_insert
> 0.00 +0.3 0.29 ±129% +1.2 1.16 ± 26% perf-profile.children.cycles-pp.do_open
> 0.00 +0.3 0.32 ±129% +1.8 1.79 ± 43% perf-profile.children.cycles-pp.shmem_alloc_and_add_folio
> 0.00 +0.3 0.32 ±129% +1.8 1.83 ± 44% perf-profile.children.cycles-pp.shmem_get_folio_gfp
> 0.00 +0.5 0.49 ± 78% +1.8 1.83 ± 44% perf-profile.children.cycles-pp.shmem_write_begin
> 0.00 +1.1 1.09 ± 33% +0.5 0.48 ±160% perf-profile.children.cycles-pp.dup_mmap
> 0.00 +1.1 1.11 ±106% +1.6 1.60 ± 54% perf-profile.children.cycles-pp.__open64_nocancel
> 0.00 +1.1 1.15 ±102% +1.2 1.16 ± 86% perf-profile.children.cycles-pp.evlist_cpu_iterator__next
> 0.00 +1.3 1.32 ± 54% +1.4 1.36 ± 33% perf-profile.children.cycles-pp.filp_close
> 0.00 +1.3 1.32 ± 54% +1.5 1.47 ± 29% perf-profile.children.cycles-pp.put_files_struct
> 0.00 +1.4 1.37 ± 49% +1.8 1.77 ± 50% perf-profile.children.cycles-pp.setlocale
> 0.00 +1.4 1.39 ± 70% +1.8 1.80 ± 48% perf-profile.children.cycles-pp.seq_read
> 0.00 +1.5 1.55 ± 63% +1.7 1.75 ± 30% perf-profile.children.cycles-pp.do_read_fault
> 0.00 +1.7 1.66 ± 76% +0.9 0.91 ± 44% perf-profile.children.cycles-pp.event_function
> 0.00 +1.7 1.66 ± 76% +0.9 0.91 ± 44% perf-profile.children.cycles-pp.remote_function
> 0.00 +1.7 1.70 ± 71% +1.5 1.53 ± 73% perf-profile.children.cycles-pp.lookup_fast
> 0.00 +1.7 1.73 ± 53% +1.4 1.40 ± 77% perf-profile.children.cycles-pp.swevent_hlist_put_cpu
> 0.04 ± 44% +1.8 1.83 ± 96% +2.4 2.47 ± 44% perf-profile.children.cycles-pp.__schedule
> 0.00 +1.9 1.93 ± 26% +1.1 1.15 ±120% perf-profile.children.cycles-pp.dup_mm
> 0.03 ± 70% +2.0 1.99 ± 36% +1.2 1.23 ± 81% perf-profile.children.cycles-pp.handle_softirqs
> 0.00 +2.0 1.99 ± 36% +1.1 1.13 ± 67% perf-profile.children.cycles-pp.__irq_exit_rcu
> 0.00 +2.0 2.02 ± 38% +1.3 1.33 ± 57% perf-profile.children.cycles-pp.folios_put_refs
> 0.00 +2.1 2.06 ± 52% +1.4 1.38 ± 77% perf-profile.children.cycles-pp._raw_spin_lock
> 0.00 +2.1 2.12 ± 58% +3.7 3.71 ± 40% perf-profile.children.cycles-pp.open64
> 0.00 +2.2 2.16 ± 44% +1.7 1.75 ± 30% perf-profile.children.cycles-pp.do_pte_missing
> 0.00 +2.2 2.21 ± 68% +2.2 2.18 ± 58% perf-profile.children.cycles-pp.link_path_walk
> 0.00 +2.2 2.23 ± 33% +1.4 1.40 ± 99% perf-profile.children.cycles-pp.copy_process
> 0.00 +2.3 2.30 ± 40% +1.8 1.78 ± 48% perf-profile.children.cycles-pp.__tlb_batch_free_encoded_pages
> 0.00 +2.3 2.30 ± 40% +1.8 1.78 ± 48% perf-profile.children.cycles-pp.free_pages_and_swap_cache
> 0.00 +2.3 2.34 ±103% +5.1 5.09 ± 64% perf-profile.children.cycles-pp.perf_c2c__record
> 0.00 +2.3 2.34 ± 46% +1.5 1.52 ± 99% perf-profile.children.cycles-pp.walk_component
> 0.00 +2.4 2.37 ± 36% +2.0 2.04 ± 32% perf-profile.children.cycles-pp.zap_present_ptes
> 0.00 +2.5 2.48 ± 32% +2.5 2.51 ± 55% perf-profile.children.cycles-pp.get_cpu_sleep_time_us
> 0.00 +2.5 2.50 ± 73% +1.6 1.56 ± 76% perf-profile.children.cycles-pp.__evlist__enable
> 0.00 +2.6 2.62 ± 35% +1.6 1.57 ± 91% perf-profile.children.cycles-pp.__do_sys_clone
> 0.00 +2.6 2.62 ± 35% +1.6 1.57 ± 91% perf-profile.children.cycles-pp.kernel_clone
> 0.00 +2.7 2.67 ± 54% +2.6 2.59 ± 40% perf-profile.children.cycles-pp.load_elf_binary
> 0.00 +2.7 2.68 ± 35% +3.0 3.02 ± 45% perf-profile.children.cycles-pp.get_idle_time
> 0.00 +2.8 2.77 ± 33% +4.2 4.17 ± 35% perf-profile.children.cycles-pp.uptime_proc_show
> 0.00 +2.8 2.83 ± 48% +2.6 2.59 ± 40% perf-profile.children.cycles-pp.search_binary_handler
> 0.00 +2.8 2.83 ± 48% +2.7 2.68 ± 42% perf-profile.children.cycles-pp.exec_binprm
> 0.00 +2.9 2.91 ± 32% +1.8 1.83 ± 85% perf-profile.children.cycles-pp._Fork
> 0.00 +3.1 3.10 ± 64% +0.9 0.95 ±252% perf-profile.children.cycles-pp.proc_reg_read_iter
> 0.00 +3.2 3.24 ± 39% +2.8 2.85 ± 49% perf-profile.children.cycles-pp.bprm_execve
> 0.00 +3.2 3.24 ± 36% +2.0 2.00 ± 56% perf-profile.children.cycles-pp.__x64_sys_exit_group
> 0.00 +3.2 3.24 ± 36% +2.1 2.09 ± 53% perf-profile.children.cycles-pp.x64_sys_call
> 0.00 +3.8 3.85 ± 39% +3.3 3.29 ± 47% perf-profile.children.cycles-pp.execve
> 0.00 +3.8 3.85 ± 39% +3.3 3.34 ± 49% perf-profile.children.cycles-pp.__x64_sys_execve
> 0.00 +3.8 3.85 ± 39% +3.3 3.34 ± 49% perf-profile.children.cycles-pp.do_execveat_common
> 0.00 +4.0 3.99 ± 38% +4.1 4.06 ± 54% perf-profile.children.cycles-pp.mutex_unlock
> 0.00 +4.2 4.19 ± 31% +3.0 3.02 ± 20% perf-profile.children.cycles-pp.zap_pte_range
> 0.00 +4.2 4.25 ± 65% +8.0 7.98 ± 43% perf-profile.children.cycles-pp.generic_perform_write
> 0.00 +4.3 4.29 ± 29% +3.0 3.02 ± 20% perf-profile.children.cycles-pp.unmap_page_range
> 0.00 +4.3 4.29 ± 29% +3.0 3.02 ± 20% perf-profile.children.cycles-pp.zap_pmd_range
> 0.00 +4.3 4.31 ± 51% +5.3 5.31 ± 46% perf-profile.children.cycles-pp.do_filp_open
> 0.00 +4.3 4.31 ± 51% +5.3 5.31 ± 46% perf-profile.children.cycles-pp.path_openat
> 0.19 ± 23% +4.4 4.60 ± 26% +3.4 3.54 ± 27% perf-profile.children.cycles-pp.__handle_mm_fault
> 0.00 +4.5 4.46 ± 59% +8.1 8.07 ± 42% perf-profile.children.cycles-pp.shmem_file_write_iter
> 0.00 +4.5 4.55 ± 24% +4.0 3.97 ± 39% perf-profile.children.cycles-pp.smp_call_function_single
> 0.00 +4.5 4.55 ± 24% +4.1 4.06 ± 38% perf-profile.children.cycles-pp.event_function_call
> 0.00 +4.6 4.58 ± 30% +3.2 3.19 ± 24% perf-profile.children.cycles-pp.unmap_vmas
> 0.51 ± 6% +4.6 5.14 ± 24% +3.6 4.06 ± 30% perf-profile.children.cycles-pp.handle_mm_fault
> 0.00 +4.7 4.68 ± 55% +8.4 8.41 ± 39% perf-profile.children.cycles-pp.writen
> 0.00 +4.7 4.68 ± 55% +8.5 8.49 ± 39% perf-profile.children.cycles-pp.record__pushfn
> 0.00 +4.8 4.80 ± 48% +6.1 6.15 ± 34% perf-profile.children.cycles-pp.do_sys_openat2
> 0.77 ± 3% +4.8 5.59 ± 21% +4.3 5.07 ± 29% perf-profile.children.cycles-pp.exc_page_fault
> 0.76 ± 3% +4.8 5.59 ± 21% +4.3 5.07 ± 29% perf-profile.children.cycles-pp.do_user_addr_fault
> 0.00 +4.9 4.90 ± 57% +10.3 10.28 ± 65% perf-profile.children.cycles-pp.vfs_write
> 0.00 +4.9 4.90 ± 57% +10.4 10.41 ± 63% perf-profile.children.cycles-pp.ksys_write
> 0.00 +4.9 4.90 ± 48% +6.1 6.15 ± 34% perf-profile.children.cycles-pp.__x64_sys_openat
> 0.00 +4.9 4.92 ± 26% +4.7 4.66 ± 47% perf-profile.children.cycles-pp.sw_perf_event_destroy
> 0.00 +5.0 4.99 ±100% +2.6 2.64 ±101% perf-profile.children.cycles-pp.perf_rotate_context
> 0.00 +5.0 5.01 ± 54% +10.9 10.87 ± 59% perf-profile.children.cycles-pp.write
> 0.00 +5.1 5.09 ±102% +2.7 2.74 ± 94% perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
> 0.00 +5.4 5.43 ± 25% +5.0 4.97 ± 45% perf-profile.children.cycles-pp._free_event
> 1.18 +5.6 6.78 ± 20% +5.5 6.71 ± 24% perf-profile.children.cycles-pp.asm_exc_page_fault
> 0.46 +5.6 6.07 ± 90% +4.1 4.54 ± 53% perf-profile.children.cycles-pp.__hrtimer_run_queues
> 0.00 +5.7 5.75 ± 39% +10.2 10.22 ± 24% perf-profile.children.cycles-pp.perf_mmap__push
> 0.00 +5.7 5.75 ± 39% +10.4 10.38 ± 23% perf-profile.children.cycles-pp.record__mmap_read_evlist
> 0.53 +5.8 6.28 ± 89% +4.4 4.91 ± 50% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
> 0.53 +5.8 6.28 ± 89% +4.4 4.91 ± 50% perf-profile.children.cycles-pp.hrtimer_interrupt
> 0.00 +6.6 6.65 ± 77% +3.3 3.32 ± 91% perf-profile.children.cycles-pp.__intel_pmu_enable_all
> 0.00 +6.8 6.85 ± 20% +5.2 5.23 ± 25% perf-profile.children.cycles-pp.exit_mm
> 0.58 ± 2% +7.6 8.14 ± 75% +6.0 6.55 ± 38% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
> 0.00 +7.7 7.67 ± 23% +6.1 6.14 ± 15% perf-profile.children.cycles-pp.exit_mmap
> 0.00 +7.7 7.67 ± 30% +7.0 7.05 ± 50% perf-profile.children.cycles-pp.seq_read_iter
> 0.00 +7.7 7.72 ± 80% +8.2 8.15 ± 51% perf-profile.children.cycles-pp.asm_sysvec_reschedule_ipi
> 0.00 +7.8 7.75 ± 23% +6.1 6.14 ± 15% perf-profile.children.cycles-pp.__mmput
> 0.00 +8.0 8.03 ± 27% +7.4 7.37 ± 52% perf-profile.children.cycles-pp.ksys_read
> 0.00 +8.0 8.03 ± 27% +7.4 7.37 ± 52% perf-profile.children.cycles-pp.vfs_read
> 0.00 +8.1 8.13 ± 28% +7.4 7.37 ± 52% perf-profile.children.cycles-pp.read
> 0.02 ±141% +9.0 9.05 ± 35% +13.9 13.88 ± 19% perf-profile.children.cycles-pp.__cmd_record
> 0.02 ±141% +9.0 9.05 ± 35% +13.9 13.88 ± 19% perf-profile.children.cycles-pp.cmd_record
> 0.02 ±141% +9.0 9.05 ± 35% +13.9 13.88 ± 19% perf-profile.children.cycles-pp.handle_internal_command
> 0.02 ±141% +9.0 9.05 ± 35% +13.9 13.88 ± 19% perf-profile.children.cycles-pp.main
> 0.02 ±141% +9.0 9.05 ± 35% +13.9 13.88 ± 19% perf-profile.children.cycles-pp.run_builtin
> 0.00 +9.3 9.26 ± 30% +9.0 8.96 ± 31% perf-profile.children.cycles-pp.perf_event_release_kernel
> 0.00 +9.3 9.26 ± 30% +9.0 8.96 ± 31% perf-profile.children.cycles-pp.perf_release
> 1.02 ± 4% +9.3 10.33 ± 27% +9.8 10.80 ± 35% perf-profile.children.cycles-pp.task_work_run
> 0.00 +11.0 11.05 ± 28% +10.4 10.37 ± 32% perf-profile.children.cycles-pp.__fput
> 0.00 +15.8 15.85 ± 25% +16.1 16.11 ± 29% perf-profile.children.cycles-pp.get_signal
> 0.00 +15.8 15.85 ± 25% +16.2 16.17 ± 29% perf-profile.children.cycles-pp.arch_do_signal_or_restart
> 0.00 +19.1 19.09 ± 19% +18.1 18.06 ± 29% perf-profile.children.cycles-pp.do_exit
> 0.00 +19.1 19.09 ± 19% +18.1 18.06 ± 29% perf-profile.children.cycles-pp.do_group_exit
> 1.70 ± 2% +30.7 32.41 ± 21% +27.2 28.87 ± 12% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
> 1.22 ± 3% +36.9 38.10 ± 13% +34.8 36.05 ± 11% perf-profile.children.cycles-pp.start_secondary
> 1.21 ± 3% +37.2 38.43 ± 11% +34.2 35.40 ± 8% perf-profile.children.cycles-pp.acpi_idle_do_entry
> 1.21 ± 3% +37.2 38.43 ± 11% +34.2 35.40 ± 8% perf-profile.children.cycles-pp.acpi_idle_enter
> 1.21 ± 3% +37.2 38.43 ± 11% +34.2 35.40 ± 8% perf-profile.children.cycles-pp.acpi_safe_halt
> 1.22 ± 3% +37.3 38.54 ± 12% +35.0 36.18 ± 10% perf-profile.children.cycles-pp.cpuidle_idle_call
> 1.21 ± 3% +37.3 38.54 ± 12% +34.7 35.87 ± 10% perf-profile.children.cycles-pp.cpuidle_enter
> 1.21 ± 3% +37.3 38.54 ± 12% +34.7 35.87 ± 10% perf-profile.children.cycles-pp.cpuidle_enter_state
> 1.22 ± 3% +37.6 38.84 ± 12% +35.4 36.60 ± 11% perf-profile.children.cycles-pp.common_startup_64
> 1.22 ± 3% +37.6 38.84 ± 12% +35.4 36.60 ± 11% perf-profile.children.cycles-pp.cpu_startup_entry
> 1.22 ± 3% +37.6 38.84 ± 12% +35.4 36.60 ± 11% perf-profile.children.cycles-pp.do_idle
> 92.37 -92.4 0.00 -92.4 0.00 perf-profile.self.cycles-pp.osq_lock
> 0.00 +0.1 0.09 ±264% +0.8 0.84 ± 51% perf-profile.self.cycles-pp.kcpustat_cpu_fetch
> 0.00 +2.1 2.06 ± 52% +1.4 1.38 ± 77% perf-profile.self.cycles-pp._raw_spin_lock
> 0.00 +2.6 2.61 ± 36% +2.8 2.75 ± 48% perf-profile.self.cycles-pp.smp_call_function_single
> 0.00 +3.7 3.68 ± 37% +3.7 3.70 ± 64% perf-profile.self.cycles-pp.mutex_unlock
> 0.00 +6.6 6.65 ± 77% +3.3 3.32 ± 91% perf-profile.self.cycles-pp.__intel_pmu_enable_all
> 1.19 ± 3% +29.2 30.38 ± 15% +27.9 29.13 ± 13% perf-profile.self.cycles-pp.acpi_safe_halt
>
>
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-02-14 22:53 ` Yang Shi
@ 2025-02-18 6:30 ` Oliver Sang
2025-02-19 1:12 ` Yang Shi
0 siblings, 1 reply; 35+ messages in thread
From: Oliver Sang @ 2025-02-18 6:30 UTC (permalink / raw)
To: Yang Shi
Cc: oe-lkp, lkp, linux-kernel, arnd, gregkh, Liam.Howlett,
lorenzo.stoakes, vbabka, jannh, willy, liushixin2, akpm,
linux-mm, oliver.sang
hi, Yang Shi,
On Fri, Feb 14, 2025 at 02:53:37PM -0800, Yang Shi wrote:
>
> On 2/12/25 6:04 PM, Oliver Sang wrote:
> > hi, Yang Shi,
> >
> > On Fri, Feb 07, 2025 at 10:10:37AM -0800, Yang Shi wrote:
> > > On 2/6/25 12:02 AM, Oliver Sang wrote:
> > [...]
> >
> > > > since we applied your "/dev/zero: make private mapping full anonymous mapping"
> > > > patch upon a68d3cbfad like below:
> > > >
> > > > * 7143ee2391f1e /dev/zero: make private mapping full anonymous mapping
> > > > * a68d3cbfade64 memstick: core: fix kernel-doc notation
> > > >
> > > > so I applied below patch also upon a68d3cbfad.
> > > >
> > > > we saw big improvement but not that big.
> > > >
> > > > =========================================================================================
> > > > compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
> > > > gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
> > > >
> > > > commit:
> > > > a68d3cbfad ("memstick: core: fix kernel-doc notation")
> > > > 52ec85cb99 <--- your patch
> > > >
> > > >
> > > > a68d3cbfade64392 52ec85cb99e9b31dc304eae965a
> > > > ---------------- ---------------------------
> > > > %stddev %change %stddev
> > > > \ | \
> > > > 14364828 ± 4% +410.6% 73349239 ± 3% vm-scalability.throughput
> > > >
> > > > full comparison as below [1] just FYI.
> > > Thanks for the update. I stared at the profiling report for a whole day, but
> > > I didn't figure out where that 400% lost. I just saw the number of page
> > > faults was fewer. And it seems like the reduction of page faults match the
> > > 400% loss. So I did more trace and profiling.
> > >
> > > The test case did the below stuff in a tight loop:
> > > mmap 40K memory from /dev/zero (read only)
> > > read the area
> > >
> > > So two major factors to the performance: mmap and page fault. The
> > > alternative patch did reduce the overhead of mmap to the same level as the
> > > original patch.
> > >
> > > The further perf profiling showed the cost of page fault is higher than the
> > > original patch. But the profiling of page fault was interesting:
> > >
> > > - 44.87% 0.01% usemem [kernel.kallsyms] [k]
> > > do_translation_fault
> > > - 44.86% do_translation_fault
> > > - 44.83% do_page_fault
> > > - 44.53% handle_mm_fault
> > > 9.04% __handle_mm_fault
> > >
> > > Page fault consumed 40% of cpu time in handle_mm_fault, but
> > > __handle_mm_fault just consumed 9%, I expected it should be the major
> > > consumer.
> > >
> > > So I annotated handle_mm_fault, then found the most time was consumed by
> > > lru_gen_enter_fault() -> vma_has_recency() (my kernel has multi-gen LRU
> > > enabled):
> > >
> > > │ if (vma->vm_file && (vma->vm_file->f_mode & FMODE_NOREUSE))
> > > │ ↓ cbz x1, b4
> > > 0.00 │ ldr w0, [x1, #12]
> > > 99.59 │ eor x0, x0, #0x800000
> > > 0.00 │ ubfx w0, w0, #23, #1
> > > │ current->in_lru_fault = vma_has_recency(vma);
> > > 0.00 │ b4: ldrh w1, [x2, #1992]
> > > 0.01 │ bfi w1, w0, #5, #1
> > > 0.00 │ strh w1, [x2, #1992]
> > >
> > >
> > > vma_has_recency() read vma->vm_file->f_mode if vma->vm_file is not NULL. But
> > > that load took a long time. So I inspected struct file and saw:
> > >
> > > struct file {
> > > file_ref_t f_ref;
> > > spinlock_t f_lock;
> > > fmode_t f_mode;
> > > const struct file_operations *f_op;
> > > ...
> > > }
> > >
> > > The f_mode is in the same cache line with f_ref (my kernel does NOT have
> > > spin lock debug enabled). The test case mmap /dev/zero in a tight loop, so
> > > the refcount is modified (fget/fput) very frequently, this resulted in
> > > somehow false sharing.
> > >
> > > So I tried the below patch on top of the alternative patch:
> > >
> > > diff --git a/include/linux/mm_inline.h b/include/linux/mm_inline.h
> > > index f9157a0c42a5..ba11dc0b1c7c 100644
> > > --- a/include/linux/mm_inline.h
> > > +++ b/include/linux/mm_inline.h
> > > @@ -608,6 +608,9 @@ static inline bool vma_has_recency(struct vm_area_struct
> > > *vma)
> > > if (vma->vm_flags & (VM_SEQ_READ | VM_RAND_READ))
> > > return false;
> > >
> > > + if (vma_is_anonymous(vma))
> > > + return true;
> > > +
> > > if (vma->vm_file && (vma->vm_file->f_mode & FMODE_NOREUSE))
> > > return false;
> > >
> > > This made the profiling of page fault look normal:
> > >
> > > - 1.90% do_translation_fault
> > > - 1.87% do_page_fault
> > > - 1.49% handle_mm_fault
> > > - 1.36% __handle_mm_fault
> > >
> > > Please try this in your test.
> > >
> > > But AFAICT I have never seen performance issue reported due to the false
> > > sharing of refcount and other fields in struct file. This benchmark stressed
> > > this quite badly.
> > I applied your above patch upon alternative patch last time, then saw more
> > improvement (+445.2% vs a68d3cbfad), but still not that big as in our original
> > report.
>
> Thanks for the update. It looks like the problem is still in page faults. I
> did my test on arm64 machine. I also noticed struct file has
> "__randomize_layout", so it may have different layout on x86 than arm64?
>
> The page fault handler may also access other fields of struct file that may
> cause false sharing, for example, accessing f_mapping to read gfp flags.
> This may not be a problem on my machine, but may be more costly on yours
> depending on the real layout of struct file on the machines,
>
> Can you please try the below patch on top of the current patches? Thank you
> so much for your patience.
you are welcome!
now has more improvements. I just list "a68d3cbfad + 3 patches so far" vs
a68d3cbfad below, if you want more data, please let me know.
=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
commit:
a68d3cbfad ("memstick: core: fix kernel-doc notation")
edc84ea79f <--- a68d3cbfad + 3 patches so far
a68d3cbfade64392 edc84ea79f8dc11853076b96ad5
---------------- ---------------------------
%stddev %change %stddev
\ | \
14364828 ± 4% +685.6% 1.129e+08 ± 5% vm-scalability.throughput
full data is as below [1] FYI.
>
> diff --git a/mm/memory.c b/mm/memory.c
> index 539c0f7c6d54..1fa9dbce0f66 100644
> --- a/mm/memory.c
> +++ b/mm/memory.c
> @@ -3214,6 +3214,9 @@ static gfp_t __get_fault_gfp_mask(struct
> vm_area_struct *vma)
> {
> struct file *vm_file = vma->vm_file;
>
> + if (vma_is_anonymous(vma))
> + return GFP_KERNEL;
> +
> if (vm_file)
> return mapping_gfp_mask(vm_file->f_mapping) | __GFP_FS |
> __GFP_IO;
>
[1]
=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
commit:
a68d3cbfad ("memstick: core: fix kernel-doc notation")
edc84ea79f <--- a68d3cbfad + 3 patches so far
a68d3cbfade64392 edc84ea79f8dc11853076b96ad5
---------------- ---------------------------
%stddev %change %stddev
\ | \
5.262e+09 ± 3% -59.8% 2.114e+09 ± 2% cpuidle..time
7924008 ± 3% -83.9% 1275131 ± 5% cpuidle..usage
1871164 ± 4% -16.8% 1557233 ± 8% numa-numastat.node3.local_node
1952164 ± 3% -14.8% 1663189 ± 7% numa-numastat.node3.numa_hit
399.52 -75.0% 99.77 ± 2% uptime.boot
14507 -22.1% 11296 uptime.idle
3408 ± 5% -99.8% 7.25 ± 46% perf-c2c.DRAM.local
18076 ± 3% -99.8% 43.00 ±100% perf-c2c.DRAM.remote
8082 ± 5% -99.8% 12.50 ± 63% perf-c2c.HITM.local
6544 ± 6% -99.7% 22.88 ±151% perf-c2c.HITM.remote
14627 ± 4% -99.8% 35.38 ±114% perf-c2c.HITM.total
6.99 ± 3% +177.6% 19.41 ± 3% vmstat.cpu.id
91.35 -28.5% 65.31 vmstat.cpu.sy
1.71 +793.1% 15.25 ± 4% vmstat.cpu.us
34204 ± 5% -64.1% 12271 ± 9% vmstat.system.cs
266575 -21.2% 210049 vmstat.system.in
6.49 ± 3% +10.0 16.46 ± 3% mpstat.cpu.all.idle%
0.63 -0.3 0.34 ± 3% mpstat.cpu.all.irq%
0.03 ± 2% +0.3 0.31 ± 4% mpstat.cpu.all.soft%
91.17 -24.1 67.09 mpstat.cpu.all.sys%
1.68 ± 2% +14.1 15.80 ± 4% mpstat.cpu.all.usr%
337.33 -98.7% 4.25 ± 10% mpstat.max_utilization.seconds
352.76 -84.7% 53.95 ± 4% time.elapsed_time
352.76 -84.7% 53.95 ± 4% time.elapsed_time.max
225965 ± 7% -17.1% 187329 ± 12% time.involuntary_context_switches
9.592e+08 ± 4% +11.9% 1.074e+09 time.minor_page_faults
20852 -10.0% 18761 time.percent_of_cpu_this_job_got
72302 -88.6% 8227 ± 6% time.system_time
1260 ± 3% +50.7% 1899 time.user_time
5393707 ± 5% -98.8% 66895 ± 21% time.voluntary_context_switches
1609925 -50.7% 793216 meminfo.Active
1609925 -50.7% 793216 meminfo.Active(anon)
160837 ± 33% -72.5% 44155 ± 9% meminfo.AnonHugePages
4435665 -18.7% 3608195 meminfo.Cached
1775547 -44.2% 990889 meminfo.Committed_AS
148539 -47.4% 78096 meminfo.Mapped
4245538 ± 4% -24.6% 3202495 meminfo.PageTables
929777 -88.9% 102759 meminfo.Shmem
25676018 ± 3% +14.3% 29335678 meminfo.max_used_kB
64129 ± 4% +706.8% 517389 ± 7% vm-scalability.median
45.40 ± 5% +2248.9 2294 ± 2% vm-scalability.stddev%
14364828 ± 4% +685.6% 1.129e+08 ± 5% vm-scalability.throughput
352.76 -84.7% 53.95 ± 4% vm-scalability.time.elapsed_time
352.76 -84.7% 53.95 ± 4% vm-scalability.time.elapsed_time.max
225965 ± 7% -17.1% 187329 ± 12% vm-scalability.time.involuntary_context_switches
9.592e+08 ± 4% +11.9% 1.074e+09 vm-scalability.time.minor_page_faults
20852 -10.0% 18761 vm-scalability.time.percent_of_cpu_this_job_got
72302 -88.6% 8227 ± 6% vm-scalability.time.system_time
1260 ± 3% +50.7% 1899 vm-scalability.time.user_time
5393707 ± 5% -98.8% 66895 ± 21% vm-scalability.time.voluntary_context_switches
4.316e+09 ± 4% +11.9% 4.832e+09 vm-scalability.workload
1063552 ± 4% -24.9% 799008 ± 3% numa-meminfo.node0.PageTables
125455 ±106% -85.5% 18164 ±165% numa-meminfo.node0.Shmem
1062709 ± 4% -25.7% 789746 ± 4% numa-meminfo.node1.PageTables
176171 ± 71% -92.4% 13303 ±230% numa-meminfo.node1.Shmem
35515 ± 91% -97.3% 976.55 ± 59% numa-meminfo.node2.Mapped
1058901 ± 4% -25.3% 791392 ± 4% numa-meminfo.node2.PageTables
770405 ± 30% -79.2% 160245 ±101% numa-meminfo.node3.Active
770405 ± 30% -79.2% 160245 ±101% numa-meminfo.node3.Active(anon)
380096 ± 50% -62.5% 142513 ± 98% numa-meminfo.node3.AnonPages.max
1146977 ±108% -92.8% 82894 ± 60% numa-meminfo.node3.FilePages
52663 ± 47% -97.2% 1488 ± 39% numa-meminfo.node3.Mapped
1058539 ± 4% -22.3% 821992 ± 3% numa-meminfo.node3.PageTables
558943 ± 14% -93.7% 35227 ±124% numa-meminfo.node3.Shmem
265763 ± 4% -24.9% 199601 ± 3% numa-vmstat.node0.nr_page_table_pages
31364 ±106% -85.5% 4539 ±165% numa-vmstat.node0.nr_shmem
265546 ± 4% -25.5% 197854 ± 5% numa-vmstat.node1.nr_page_table_pages
44052 ± 71% -92.5% 3323 ±230% numa-vmstat.node1.nr_shmem
8961 ± 91% -97.3% 244.02 ± 59% numa-vmstat.node2.nr_mapped
264589 ± 4% -25.2% 197920 ± 3% numa-vmstat.node2.nr_page_table_pages
192683 ± 30% -79.2% 40126 ±101% numa-vmstat.node3.nr_active_anon
286819 ±108% -92.8% 20761 ± 60% numa-vmstat.node3.nr_file_pages
13124 ± 49% -97.2% 372.02 ± 39% numa-vmstat.node3.nr_mapped
264499 ± 4% -22.4% 205376 ± 3% numa-vmstat.node3.nr_page_table_pages
139810 ± 14% -93.7% 8844 ±124% numa-vmstat.node3.nr_shmem
192683 ± 30% -79.2% 40126 ±101% numa-vmstat.node3.nr_zone_active_anon
1951359 ± 3% -14.9% 1661427 ± 7% numa-vmstat.node3.numa_hit
1870359 ± 4% -16.8% 1555470 ± 8% numa-vmstat.node3.numa_local
402515 -50.7% 198246 proc-vmstat.nr_active_anon
170568 +1.8% 173591 proc-vmstat.nr_anon_pages
1109246 -18.7% 902238 proc-vmstat.nr_file_pages
37525 -47.3% 19768 proc-vmstat.nr_mapped
1059932 ± 4% -24.2% 803105 ± 2% proc-vmstat.nr_page_table_pages
232507 -89.0% 25623 proc-vmstat.nr_shmem
37297 -5.4% 35299 proc-vmstat.nr_slab_reclaimable
402515 -50.7% 198246 proc-vmstat.nr_zone_active_anon
61931 ± 8% -83.9% 9948 ± 59% proc-vmstat.numa_hint_faults
15755 ± 21% -96.6% 541.38 ± 36% proc-vmstat.numa_hint_faults_local
6916516 ± 3% -8.0% 6360040 proc-vmstat.numa_hit
6568542 ± 3% -8.5% 6012265 proc-vmstat.numa_local
293942 ± 3% -68.8% 91724 ± 48% proc-vmstat.numa_pte_updates
9.608e+08 ± 4% +11.8% 1.074e+09 proc-vmstat.pgfault
55981 ± 2% -68.7% 17541 ± 2% proc-vmstat.pgreuse
0.82 ± 4% -51.0% 0.40 ± 8% perf-stat.i.MPKI
2.714e+10 ± 2% +378.3% 1.298e+11 ± 9% perf-stat.i.branch-instructions
0.11 ± 3% +0.1 0.24 ± 8% perf-stat.i.branch-miss-rate%
24932893 +306.8% 1.014e+08 ± 9% perf-stat.i.branch-misses
64.93 -7.5 57.48 perf-stat.i.cache-miss-rate%
88563288 ± 3% +35.0% 1.196e+08 ± 7% perf-stat.i.cache-misses
1.369e+08 ± 3% +43.7% 1.968e+08 ± 7% perf-stat.i.cache-references
34508 ± 4% -47.3% 18199 ± 9% perf-stat.i.context-switches
7.67 -75.7% 1.87 ± 3% perf-stat.i.cpi
224605 +22.5% 275084 ± 6% perf-stat.i.cpu-clock
696.35 ± 2% -53.5% 323.77 ± 2% perf-stat.i.cpu-migrations
10834 ± 4% -24.1% 8224 ± 11% perf-stat.i.cycles-between-cache-misses
1.102e+11 +282.2% 4.212e+11 ± 9% perf-stat.i.instructions
0.14 +334.6% 0.62 ± 5% perf-stat.i.ipc
24.25 ± 3% +626.9% 176.25 ± 4% perf-stat.i.metric.K/sec
2722043 ± 3% +803.8% 24600740 ± 9% perf-stat.i.minor-faults
2722043 ± 3% +803.8% 24600739 ± 9% perf-stat.i.page-faults
224605 +22.5% 275084 ± 6% perf-stat.i.task-clock
0.81 ± 3% -62.2% 0.31 ± 11% perf-stat.overall.MPKI
0.09 -0.0 0.08 ± 2% perf-stat.overall.branch-miss-rate%
64.81 -2.4 62.37 perf-stat.overall.cache-miss-rate%
7.24 -70.7% 2.12 ± 5% perf-stat.overall.cpi
8933 ± 4% -21.9% 6978 ± 7% perf-stat.overall.cycles-between-cache-misses
0.14 +242.2% 0.47 ± 5% perf-stat.overall.ipc
9012 ± 2% -57.8% 3806 perf-stat.overall.path-length
2.701e+10 ± 2% +285.4% 1.041e+11 ± 5% perf-stat.ps.branch-instructions
24708939 +215.8% 78042343 ± 4% perf-stat.ps.branch-misses
89032538 ± 3% +15.9% 1.032e+08 ± 8% perf-stat.ps.cache-misses
1.374e+08 ± 3% +20.6% 1.656e+08 ± 9% perf-stat.ps.cache-references
34266 ± 5% -66.2% 11570 ± 10% perf-stat.ps.context-switches
223334 -1.6% 219861 perf-stat.ps.cpu-clock
7.941e+11 -9.9% 7.157e+11 perf-stat.ps.cpu-cycles
693.54 ± 2% -67.2% 227.38 ± 4% perf-stat.ps.cpu-migrations
1.097e+11 +208.3% 3.381e+11 ± 5% perf-stat.ps.instructions
2710577 ± 3% +626.7% 19698901 ± 5% perf-stat.ps.minor-faults
2710577 ± 3% +626.7% 19698902 ± 5% perf-stat.ps.page-faults
223334 -1.6% 219861 perf-stat.ps.task-clock
3.886e+13 ± 2% -52.7% 1.839e+13 perf-stat.total.instructions
64052898 ± 5% -99.9% 81213 ± 23% sched_debug.cfs_rq:/.avg_vruntime.avg
95701822 ± 7% -96.4% 3425672 ± 7% sched_debug.cfs_rq:/.avg_vruntime.max
43098762 ± 6% -100.0% 153.42 ± 36% sched_debug.cfs_rq:/.avg_vruntime.min
9223270 ± 9% -95.9% 380347 ± 16% sched_debug.cfs_rq:/.avg_vruntime.stddev
0.00 ± 22% -100.0% 0.00 sched_debug.cfs_rq:/.h_nr_delayed.avg
0.69 ± 8% -100.0% 0.00 sched_debug.cfs_rq:/.h_nr_delayed.max
0.05 ± 12% -100.0% 0.00 sched_debug.cfs_rq:/.h_nr_delayed.stddev
0.78 ± 2% -94.5% 0.04 ± 21% sched_debug.cfs_rq:/.h_nr_running.avg
1.97 ± 5% -49.3% 1.00 sched_debug.cfs_rq:/.h_nr_running.max
0.28 ± 7% -29.1% 0.20 ± 10% sched_debug.cfs_rq:/.h_nr_running.stddev
411536 ± 58% -100.0% 1.15 ±182% sched_debug.cfs_rq:/.left_deadline.avg
43049468 ± 22% -100.0% 258.27 ±182% sched_debug.cfs_rq:/.left_deadline.max
3836405 ± 37% -100.0% 17.22 ±182% sched_debug.cfs_rq:/.left_deadline.stddev
411536 ± 58% -100.0% 1.06 ±191% sched_debug.cfs_rq:/.left_vruntime.avg
43049467 ± 22% -100.0% 236.56 ±191% sched_debug.cfs_rq:/.left_vruntime.max
3836405 ± 37% -100.0% 15.77 ±191% sched_debug.cfs_rq:/.left_vruntime.stddev
64052901 ± 5% -99.9% 81213 ± 23% sched_debug.cfs_rq:/.min_vruntime.avg
95701822 ± 7% -96.4% 3425672 ± 7% sched_debug.cfs_rq:/.min_vruntime.max
43098762 ± 6% -100.0% 153.42 ± 36% sched_debug.cfs_rq:/.min_vruntime.min
9223270 ± 9% -95.9% 380347 ± 16% sched_debug.cfs_rq:/.min_vruntime.stddev
0.77 ± 2% -94.4% 0.04 ± 21% sched_debug.cfs_rq:/.nr_running.avg
1.50 ± 9% -33.3% 1.00 sched_debug.cfs_rq:/.nr_running.max
0.26 ± 10% -22.7% 0.20 ± 10% sched_debug.cfs_rq:/.nr_running.stddev
1.61 ± 24% +413.4% 8.24 ± 60% sched_debug.cfs_rq:/.removed.runnable_avg.avg
86.69 +508.6% 527.62 ± 4% sched_debug.cfs_rq:/.removed.runnable_avg.max
11.14 ± 13% +428.4% 58.87 ± 32% sched_debug.cfs_rq:/.removed.runnable_avg.stddev
1.61 ± 24% +413.3% 8.24 ± 60% sched_debug.cfs_rq:/.removed.util_avg.avg
86.69 +508.6% 527.62 ± 4% sched_debug.cfs_rq:/.removed.util_avg.max
11.14 ± 13% +428.4% 58.87 ± 32% sched_debug.cfs_rq:/.removed.util_avg.stddev
411536 ± 58% -100.0% 1.06 ±191% sched_debug.cfs_rq:/.right_vruntime.avg
43049467 ± 22% -100.0% 236.56 ±191% sched_debug.cfs_rq:/.right_vruntime.max
3836405 ± 37% -100.0% 15.77 ±191% sched_debug.cfs_rq:/.right_vruntime.stddev
769.03 -84.7% 117.79 ± 3% sched_debug.cfs_rq:/.util_avg.avg
1621 ± 5% -32.7% 1092 ± 16% sched_debug.cfs_rq:/.util_avg.max
159.12 ± 8% +33.2% 211.88 ± 7% sched_debug.cfs_rq:/.util_avg.stddev
724.17 ± 2% -98.6% 10.41 ± 32% sched_debug.cfs_rq:/.util_est.avg
1360 ± 15% -51.5% 659.38 ± 10% sched_debug.cfs_rq:/.util_est.max
234.34 ± 9% -68.2% 74.43 ± 18% sched_debug.cfs_rq:/.util_est.stddev
766944 ± 3% +18.9% 912012 sched_debug.cpu.avg_idle.avg
1067639 ± 5% +25.5% 1339736 ± 9% sched_debug.cpu.avg_idle.max
3799 ± 7% -38.3% 2346 ± 23% sched_debug.cpu.avg_idle.min
321459 ± 2% -36.6% 203909 ± 7% sched_debug.cpu.avg_idle.stddev
195573 -76.9% 45144 sched_debug.cpu.clock.avg
195596 -76.9% 45160 sched_debug.cpu.clock.max
195548 -76.9% 45123 sched_debug.cpu.clock.min
13.79 ± 3% -36.0% 8.83 ± 2% sched_debug.cpu.clock.stddev
194424 -76.8% 45019 sched_debug.cpu.clock_task.avg
194608 -76.8% 45145 sched_debug.cpu.clock_task.max
181834 -82.1% 32559 sched_debug.cpu.clock_task.min
4241 ± 2% -96.8% 136.38 ± 21% sched_debug.cpu.curr->pid.avg
9799 ± 2% -59.8% 3934 sched_debug.cpu.curr->pid.max
1365 ± 10% -49.1% 695.11 ± 10% sched_debug.cpu.curr->pid.stddev
537665 ± 4% +28.3% 690006 ± 6% sched_debug.cpu.max_idle_balance_cost.max
3119 ± 56% +479.5% 18078 ± 29% sched_debug.cpu.max_idle_balance_cost.stddev
0.00 ± 12% -68.3% 0.00 ± 17% sched_debug.cpu.next_balance.stddev
0.78 ± 2% -95.3% 0.04 ± 20% sched_debug.cpu.nr_running.avg
2.17 ± 8% -53.8% 1.00 sched_debug.cpu.nr_running.max
0.29 ± 8% -35.4% 0.19 ± 9% sched_debug.cpu.nr_running.stddev
25773 ± 5% -97.0% 764.82 ± 3% sched_debug.cpu.nr_switches.avg
48669 ± 10% -77.2% 11080 ± 12% sched_debug.cpu.nr_switches.max
19006 ± 7% -99.2% 151.12 ± 15% sched_debug.cpu.nr_switches.min
4142 ± 8% -69.5% 1264 ± 6% sched_debug.cpu.nr_switches.stddev
0.07 ± 23% -93.3% 0.01 ± 53% sched_debug.cpu.nr_uninterruptible.avg
240.19 ± 16% -80.2% 47.50 ± 44% sched_debug.cpu.nr_uninterruptible.max
-77.92 -88.1% -9.25 sched_debug.cpu.nr_uninterruptible.min
37.87 ± 5% -84.7% 5.78 ± 13% sched_debug.cpu.nr_uninterruptible.stddev
195549 -76.9% 45130 sched_debug.cpu_clk
194699 -77.3% 44280 sched_debug.ktime
0.00 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.avg
0.17 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.max
0.01 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.stddev
196368 -76.6% 45975 sched_debug.sched_clk
95.59 -95.6 0.00 perf-profile.calltrace.cycles-pp.__mmap
95.54 -95.5 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
95.54 -95.5 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap
94.54 -94.5 0.00 perf-profile.calltrace.cycles-pp.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
94.46 -94.4 0.07 ±264% perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
94.45 -94.0 0.41 ±158% perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
94.14 -93.9 0.29 ±134% perf-profile.calltrace.cycles-pp.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff
94.25 -93.8 0.41 ±158% perf-profile.calltrace.cycles-pp.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
93.79 -93.7 0.07 ±264% perf-profile.calltrace.cycles-pp.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff
93.44 -93.4 0.00 perf-profile.calltrace.cycles-pp.down_write.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap
93.40 -93.4 0.00 perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma.__mmap_region
93.33 -93.3 0.00 perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma
92.89 -92.9 0.00 perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file
0.00 +1.7 1.69 ± 65% perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64
0.00 +1.9 1.90 ± 55% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group
0.00 +1.9 1.90 ± 55% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
0.00 +1.9 1.93 ± 53% perf-profile.calltrace.cycles-pp.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +1.9 1.93 ± 53% perf-profile.calltrace.cycles-pp.seq_read_iter.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64
0.00 +2.0 1.99 ± 53% perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +2.0 2.02 ± 64% perf-profile.calltrace.cycles-pp.do_pte_missing.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
0.00 +2.3 2.27 ± 56% perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
0.00 +2.3 2.27 ± 56% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
0.00 +2.3 2.27 ± 56% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe._Fork
0.00 +2.3 2.27 ± 56% perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
0.00 +2.4 2.45 ± 53% perf-profile.calltrace.cycles-pp._Fork
0.00 +2.5 2.51 ± 52% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +2.5 2.51 ± 52% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64
0.00 +2.5 2.51 ± 52% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +2.5 2.51 ± 52% perf-profile.calltrace.cycles-pp.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +3.2 3.17 ± 42% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
0.00 +3.3 3.28 ± 52% perf-profile.calltrace.cycles-pp.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin
0.00 +3.3 3.28 ± 52% perf-profile.calltrace.cycles-pp.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin.handle_internal_command
0.00 +4.1 4.10 ± 45% perf-profile.calltrace.cycles-pp.__cmd_record.cmd_record.run_builtin.handle_internal_command.main
0.00 +4.1 4.10 ± 45% perf-profile.calltrace.cycles-pp.cmd_record.run_builtin.handle_internal_command.main
0.00 +4.8 4.80 ± 61% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter
0.00 +5.0 4.98 ± 69% perf-profile.calltrace.cycles-pp.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64
0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn
0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn.perf_mmap__push
0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen
0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.write.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist
0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record
0.00 +5.1 5.11 ± 47% perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit
0.00 +5.1 5.12 ± 70% perf-profile.calltrace.cycles-pp.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record
0.00 +6.1 6.08 ± 50% perf-profile.calltrace.cycles-pp.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter
0.00 +7.8 7.84 ± 21% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
0.00 +7.9 7.88 ± 20% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
0.00 +7.9 7.88 ± 20% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
0.00 +7.9 7.88 ± 20% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
0.00 +7.9 7.88 ± 20% perf-profile.calltrace.cycles-pp.read
0.00 +11.1 11.10 ± 41% perf-profile.calltrace.cycles-pp.handle_internal_command.main
0.00 +11.1 11.10 ± 41% perf-profile.calltrace.cycles-pp.main
0.00 +11.1 11.10 ± 41% perf-profile.calltrace.cycles-pp.run_builtin.handle_internal_command.main
0.00 +11.2 11.18 ± 73% perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
0.00 +15.9 15.94 ± 41% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +15.9 15.94 ± 41% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
0.00 +19.5 19.54 ± 41% perf-profile.calltrace.cycles-pp.asm_sysvec_reschedule_ipi.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
1.21 ± 3% +36.7 37.86 ± 7% perf-profile.calltrace.cycles-pp.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
1.21 ± 3% +36.7 37.86 ± 7% perf-profile.calltrace.cycles-pp.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
1.21 ± 3% +37.0 38.24 ± 7% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
1.21 ± 3% +37.2 38.41 ± 7% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.common_startup_64
1.21 ± 3% +37.4 38.57 ± 6% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
1.22 ± 3% +38.5 39.67 ± 7% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.common_startup_64
1.22 ± 3% +38.5 39.67 ± 7% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.common_startup_64
1.22 ± 3% +38.5 39.67 ± 7% perf-profile.calltrace.cycles-pp.start_secondary.common_startup_64
1.22 ± 3% +38.9 40.09 ± 6% perf-profile.calltrace.cycles-pp.common_startup_64
2.19 ± 3% +45.2 47.41 ± 14% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
95.60 -95.4 0.22 ±135% perf-profile.children.cycles-pp.__mmap
94.55 -93.9 0.60 ±103% perf-profile.children.cycles-pp.ksys_mmap_pgoff
94.14 -93.7 0.44 ±112% perf-profile.children.cycles-pp.__mmap_new_vma
93.79 -93.7 0.10 ±264% perf-profile.children.cycles-pp.vma_link_file
94.46 -93.5 0.96 ± 76% perf-profile.children.cycles-pp.vm_mmap_pgoff
94.45 -93.5 0.96 ± 76% perf-profile.children.cycles-pp.do_mmap
94.25 -93.4 0.86 ± 87% perf-profile.children.cycles-pp.__mmap_region
93.40 -93.4 0.00 perf-profile.children.cycles-pp.rwsem_down_write_slowpath
93.33 -93.3 0.00 perf-profile.children.cycles-pp.rwsem_optimistic_spin
93.44 -93.2 0.22 ±149% perf-profile.children.cycles-pp.down_write
92.91 -92.9 0.00 perf-profile.children.cycles-pp.osq_lock
95.58 -45.4 50.16 ± 8% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
95.58 -45.4 50.16 ± 8% perf-profile.children.cycles-pp.do_syscall_64
0.00 +1.1 1.12 ± 74% perf-profile.children.cycles-pp.filemap_map_pages
0.00 +1.1 1.12 ± 76% perf-profile.children.cycles-pp.vfs_fstatat
0.00 +1.2 1.19 ± 35% perf-profile.children.cycles-pp.vsnprintf
0.00 +1.2 1.20 ± 46% perf-profile.children.cycles-pp.seq_printf
0.00 +1.3 1.28 ± 78% perf-profile.children.cycles-pp.__do_sys_newfstatat
0.00 +1.5 1.54 ± 75% perf-profile.children.cycles-pp.folios_put_refs
0.00 +1.6 1.56 ± 52% perf-profile.children.cycles-pp.__cond_resched
0.00 +1.6 1.60 ± 32% perf-profile.children.cycles-pp.sched_balance_newidle
0.00 +1.7 1.69 ± 65% perf-profile.children.cycles-pp.dup_mm
0.00 +1.9 1.93 ± 53% perf-profile.children.cycles-pp.proc_reg_read_iter
0.00 +2.0 1.99 ± 53% perf-profile.children.cycles-pp.copy_process
0.00 +2.1 2.06 ± 51% perf-profile.children.cycles-pp.__x64_sys_ioctl
0.00 +2.1 2.08 ± 45% perf-profile.children.cycles-pp.proc_single_show
0.00 +2.1 2.14 ± 45% perf-profile.children.cycles-pp.seq_read
0.00 +2.2 2.16 ± 47% perf-profile.children.cycles-pp.ioctl
0.00 +2.2 2.17 ± 33% perf-profile.children.cycles-pp.schedule
0.00 +2.2 2.20 ± 28% perf-profile.children.cycles-pp.__pick_next_task
0.00 +2.2 2.21 ± 47% perf-profile.children.cycles-pp.perf_evsel__run_ioctl
0.00 +2.3 2.26 ± 58% perf-profile.children.cycles-pp.do_read_fault
0.00 +2.3 2.27 ± 56% perf-profile.children.cycles-pp.__do_sys_clone
0.00 +2.3 2.27 ± 56% perf-profile.children.cycles-pp.kernel_clone
0.00 +2.4 2.37 ± 58% perf-profile.children.cycles-pp.zap_present_ptes
0.00 +2.4 2.45 ± 53% perf-profile.children.cycles-pp._Fork
0.00 +2.6 2.59 ± 53% perf-profile.children.cycles-pp.__x64_sys_exit_group
0.00 +2.6 2.59 ± 53% perf-profile.children.cycles-pp.x64_sys_call
0.00 +2.6 2.64 ± 44% perf-profile.children.cycles-pp.do_pte_missing
0.00 +3.1 3.13 ± 59% perf-profile.children.cycles-pp.zap_pte_range
0.00 +3.2 3.21 ± 58% perf-profile.children.cycles-pp.zap_pmd_range
0.00 +3.4 3.40 ± 56% perf-profile.children.cycles-pp.unmap_page_range
0.00 +3.4 3.43 ± 55% perf-profile.children.cycles-pp.unmap_vmas
0.19 ± 23% +3.9 4.06 ± 45% perf-profile.children.cycles-pp.__handle_mm_fault
0.51 ± 6% +4.0 4.49 ± 38% perf-profile.children.cycles-pp.handle_mm_fault
0.04 ± 44% +4.0 4.04 ± 28% perf-profile.children.cycles-pp.__schedule
0.77 ± 3% +4.4 5.18 ± 39% perf-profile.children.cycles-pp.exc_page_fault
0.76 ± 3% +4.4 5.18 ± 39% perf-profile.children.cycles-pp.do_user_addr_fault
0.58 ± 2% +4.7 5.26 ± 53% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
0.00 +5.1 5.07 ± 71% perf-profile.children.cycles-pp.writen
0.00 +5.1 5.07 ± 69% perf-profile.children.cycles-pp.generic_perform_write
0.00 +5.1 5.12 ± 47% perf-profile.children.cycles-pp.exit_mm
0.00 +5.1 5.12 ± 70% perf-profile.children.cycles-pp.record__pushfn
0.00 +5.1 5.12 ± 70% perf-profile.children.cycles-pp.shmem_file_write_iter
1.18 +5.5 6.69 ± 33% perf-profile.children.cycles-pp.asm_exc_page_fault
0.00 +6.2 6.24 ± 43% perf-profile.children.cycles-pp.__mmput
0.00 +6.2 6.24 ± 43% perf-profile.children.cycles-pp.exit_mmap
0.00 +7.0 7.00 ± 51% perf-profile.children.cycles-pp.perf_mmap__push
0.00 +7.0 7.00 ± 51% perf-profile.children.cycles-pp.record__mmap_read_evlist
0.00 +7.2 7.25 ± 52% perf-profile.children.cycles-pp.__fput
0.00 +7.3 7.35 ± 20% perf-profile.children.cycles-pp.seq_read_iter
0.00 +7.8 7.84 ± 21% perf-profile.children.cycles-pp.vfs_read
0.00 +7.9 7.88 ± 20% perf-profile.children.cycles-pp.ksys_read
0.00 +7.9 7.88 ± 20% perf-profile.children.cycles-pp.read
0.00 +9.9 9.93 ± 41% perf-profile.children.cycles-pp.asm_sysvec_reschedule_ipi
0.02 ±141% +11.1 11.10 ± 41% perf-profile.children.cycles-pp.__cmd_record
0.02 ±141% +11.1 11.10 ± 41% perf-profile.children.cycles-pp.cmd_record
0.02 ±141% +11.1 11.10 ± 41% perf-profile.children.cycles-pp.handle_internal_command
0.02 ±141% +11.1 11.10 ± 41% perf-profile.children.cycles-pp.main
0.02 ±141% +11.1 11.10 ± 41% perf-profile.children.cycles-pp.run_builtin
0.00 +11.2 11.18 ± 73% perf-profile.children.cycles-pp.vfs_write
0.00 +11.2 11.23 ± 73% perf-profile.children.cycles-pp.ksys_write
0.00 +11.2 11.23 ± 73% perf-profile.children.cycles-pp.write
0.00 +13.6 13.61 ± 44% perf-profile.children.cycles-pp.do_exit
0.00 +13.6 13.61 ± 44% perf-profile.children.cycles-pp.do_group_exit
1.70 ± 2% +25.0 26.72 ± 15% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
1.21 ± 3% +36.6 37.81 ± 7% perf-profile.children.cycles-pp.acpi_safe_halt
1.21 ± 3% +36.6 37.86 ± 7% perf-profile.children.cycles-pp.acpi_idle_do_entry
1.21 ± 3% +36.6 37.86 ± 7% perf-profile.children.cycles-pp.acpi_idle_enter
1.21 ± 3% +37.4 38.57 ± 6% perf-profile.children.cycles-pp.cpuidle_enter_state
1.21 ± 3% +37.4 38.66 ± 6% perf-profile.children.cycles-pp.cpuidle_enter
1.22 ± 3% +37.6 38.82 ± 6% perf-profile.children.cycles-pp.cpuidle_idle_call
1.22 ± 3% +38.5 39.67 ± 7% perf-profile.children.cycles-pp.start_secondary
1.22 ± 3% +38.9 40.09 ± 6% perf-profile.children.cycles-pp.common_startup_64
1.22 ± 3% +38.9 40.09 ± 6% perf-profile.children.cycles-pp.cpu_startup_entry
1.22 ± 3% +38.9 40.09 ± 6% perf-profile.children.cycles-pp.do_idle
92.37 -92.4 0.00 perf-profile.self.cycles-pp.osq_lock
1.19 ± 3% +30.7 31.90 ± 7% perf-profile.self.cycles-pp.acpi_safe_halt
0.17 ±142% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
0.19 ± 34% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
0.14 ± 55% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
0.14 ± 73% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
0.10 ± 66% -99.9% 0.00 ±264% perf-sched.sch_delay.avg.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
0.11 ± 59% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
0.04 ±132% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
0.07 ±101% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
0.02 ± 31% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
0.02 ±143% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
0.10 ± 44% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
0.12 ±145% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
0.04 ± 55% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.25 ± 41% -98.5% 0.00 ±105% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
0.11 ± 59% -97.1% 0.00 ± 61% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
0.40 ± 50% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.32 ±104% -100.0% 0.00 perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
0.01 ± 12% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
0.08 ± 28% -99.5% 0.00 ±264% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
0.18 ± 57% -96.8% 0.01 ±193% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.03 ± 83% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
0.01 ± 20% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
0.02 ± 65% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
0.32 ± 47% -98.2% 0.01 ± 42% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
0.19 ±185% -96.5% 0.01 ± 33% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
0.07 ± 20% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
0.26 ± 17% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.02 ± 60% -94.2% 0.00 ±264% perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
0.01 ±128% -100.0% 0.00 perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
1.00 ±151% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
25.45 ± 94% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
4.56 ± 67% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
3.55 ± 97% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
2.13 ± 67% -100.0% 0.00 ±264% perf-sched.sch_delay.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
3.16 ± 78% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
0.30 ±159% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
1.61 ±100% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
0.03 ± 86% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
0.20 ±182% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
3.51 ± 21% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
0.83 ±160% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
0.09 ± 31% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
3.59 ± 11% -99.9% 0.00 ±105% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
1.60 ± 69% -99.6% 0.01 ±129% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
0.81 ± 43% -100.0% 0.00 perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
1.02 ± 88% -100.0% 0.00 perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
0.02 ± 7% -100.0% 0.00 perf-sched.sch_delay.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
9.68 ± 32% -100.0% 0.00 ±264% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
12.26 ±109% -100.0% 0.01 ±193% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
5.60 ±139% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
0.03 ±106% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
2.11 ± 61% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
3.67 ± 25% -99.8% 0.01 ± 16% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
1.65 ±187% -99.3% 0.01 ± 23% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
37.84 ± 47% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
4.68 ± 36% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.21 ±169% -99.6% 0.00 ±264% perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
7.92 ±131% -99.2% 0.06 ± 92% perf-sched.sch_delay.max.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.36 ±186% -100.0% 0.00 perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
33.45 ± 3% -91.6% 2.81 ± 90% perf-sched.total_wait_and_delay.average.ms
97903 ± 4% -98.2% 1776 ± 28% perf-sched.total_wait_and_delay.count.ms
2942 ± 23% -95.2% 141.09 ± 36% perf-sched.total_wait_and_delay.max.ms
33.37 ± 3% -91.9% 2.69 ± 95% perf-sched.total_wait_time.average.ms
2942 ± 23% -96.7% 97.14 ± 19% perf-sched.total_wait_time.max.ms
3.97 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
3.08 ± 4% -94.3% 0.18 ± 92% perf-sched.wait_and_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
119.91 ± 38% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
433.73 ± 41% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
302.41 ± 5% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
1.48 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
23.24 ± 25% -96.7% 0.76 ± 27% perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
327.16 ± 9% -99.8% 0.76 ±188% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
369.37 ± 2% -98.9% 4.03 ±204% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
0.96 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
453.60 -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
187.66 -96.7% 6.11 ±109% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
2.37 ± 29% -99.6% 0.01 ±264% perf-sched.wait_and_delay.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
750.07 -99.3% 5.10 ± 84% perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
1831 ± 9% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
1269 ± 8% -45.8% 688.12 ± 21% perf-sched.wait_and_delay.count.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
6.17 ± 45% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
5.00 -100.0% 0.00 perf-sched.wait_and_delay.count.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
14.33 ± 5% -100.0% 0.00 perf-sched.wait_and_delay.count.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
810.00 ± 10% -100.0% 0.00 perf-sched.wait_and_delay.count.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
3112 ± 24% -97.9% 65.75 ±106% perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64
40.50 ± 8% -98.8% 0.50 ±173% perf-sched.wait_and_delay.count.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
73021 ± 3% -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
40.00 -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_timeout.kcompactd.kthread.ret_from_fork
1122 -99.0% 10.88 ± 98% perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
11323 ± 3% -93.6% 722.25 ± 20% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1887 ± 45% -100.0% 0.88 ±264% perf-sched.wait_and_delay.count.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
1238 -93.9% 75.62 ± 79% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
35.19 ± 57% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
1002 -91.0% 89.82 ± 93% perf-sched.wait_and_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
318.48 ± 65% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1000 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
966.90 ± 7% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
20.79 ± 19% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
1043 -98.4% 16.64 ±214% perf-sched.wait_and_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
1240 ± 20% -99.9% 1.52 ±188% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
500.34 -96.9% 15.38 ±232% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
58.83 ± 39% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
505.17 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
19.77 ± 55% -62.8% 7.36 ± 85% perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
1237 ± 34% -91.7% 102.88 ± 33% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1001 -100.0% 0.05 ±264% perf-sched.wait_and_delay.max.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
2794 ± 24% -97.9% 59.20 ± 61% perf-sched.wait_and_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
49.27 ±119% -100.0% 0.01 ±264% perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
58.17 ±187% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
3.78 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
2.99 ± 4% -97.0% 0.09 ± 91% perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
3.92 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
4.71 ± 8% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
1.67 ± 20% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
2.10 ± 27% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
0.01 ± 44% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
1.67 ± 21% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
0.04 ±133% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
67.14 ± 73% -99.5% 0.32 ±177% perf-sched.wait_time.avg.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
1.65 ± 67% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
2.30 ± 14% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
42.44 ±200% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
152.73 ±152% -100.0% 0.06 ±249% perf-sched.wait_time.avg.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin
119.87 ± 38% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
3.80 ± 18% -99.9% 0.00 ±105% perf-sched.wait_time.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
433.32 ± 41% -100.0% 0.00 perf-sched.wait_time.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
250.23 ±107% -100.0% 0.00 perf-sched.wait_time.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
29.19 ± 5% -99.2% 0.25 ± 24% perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
302.40 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
1.40 ± 6% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
4.03 ± 8% -99.9% 0.01 ±193% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
35.38 ±192% -100.0% 0.00 ±264% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
0.05 ± 40% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
0.72 ±220% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
1.00 ±120% -99.9% 0.00 ±264% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
23.07 ± 24% -97.1% 0.67 ± 10% perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
326.84 ± 9% -99.6% 1.19 ±108% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
369.18 ± 2% -98.7% 4.72 ±167% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
0.89 ± 6% -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
1.17 ± 16% -99.7% 0.00 ±264% perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
453.58 -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
4.42 -25.4% 3.30 ± 17% perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
187.58 -96.8% 6.05 ±110% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
2.36 ± 29% -99.1% 0.02 ± 84% perf-sched.wait_time.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.01 ±156% -100.0% 0.00 perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
750.01 -99.5% 3.45 ±141% perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
340.69 ±135% -100.0% 0.01 ±264% perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
535.09 ±128% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
22.04 ± 32% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
1001 -95.5% 44.91 ± 93% perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
13.57 ± 17% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
13.54 ± 10% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
10.17 ± 19% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
11.35 ± 25% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
0.01 ± 32% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
10.62 ± 9% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
0.20 ±199% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
1559 ± 64% -100.0% 0.44 ±167% perf-sched.wait_time.max.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
6.93 ± 53% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
14.42 ± 22% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
159.10 ±148% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
391.02 ±171% -100.0% 0.12 ±256% perf-sched.wait_time.max.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin
318.43 ± 65% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
13.14 ± 21% -100.0% 0.00 ±105% perf-sched.wait_time.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
1000 -100.0% 0.00 perf-sched.wait_time.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
500.84 ± 99% -100.0% 0.00 perf-sched.wait_time.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
641.50 ± 23% -99.2% 5.27 ± 76% perf-sched.wait_time.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
10.75 ± 98% -89.8% 1.10 ± 78% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
966.89 ± 7% -100.0% 0.00 perf-sched.wait_time.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
15.80 ± 8% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
16.69 ± 10% -100.0% 0.01 ±193% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
41.71 ±158% -100.0% 0.00 ±264% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
11.64 ± 61% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
2.94 ±213% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
175.70 ±210% -100.0% 0.00 ±264% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
1043 -99.6% 4.46 ±105% perf-sched.wait_time.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
1240 ± 20% -99.8% 2.37 ±108% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
500.11 -96.5% 17.32 ±201% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
32.65 ± 33% -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
22.94 ± 56% -100.0% 0.00 ±264% perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
505.00 -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
12.20 ± 43% -59.2% 4.98 perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
1237 ± 34% -92.5% 92.94 ± 20% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1000 -100.0% 0.09 ±111% perf-sched.wait_time.max.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.36 ±190% -100.0% 0.00 perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
2794 ± 24% -98.9% 30.12 ±114% perf-sched.wait_time.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
^ permalink raw reply [flat|nested] 35+ messages in thread
* Re: [PATCH] /dev/zero: make private mapping full anonymous mapping
2025-02-18 6:30 ` Oliver Sang
@ 2025-02-19 1:12 ` Yang Shi
0 siblings, 0 replies; 35+ messages in thread
From: Yang Shi @ 2025-02-19 1:12 UTC (permalink / raw)
To: Oliver Sang
Cc: oe-lkp, lkp, linux-kernel, arnd, gregkh, Liam.Howlett,
lorenzo.stoakes, vbabka, jannh, willy, liushixin2, akpm,
linux-mm
On 2/17/25 10:30 PM, Oliver Sang wrote:
> hi, Yang Shi,
>
> On Fri, Feb 14, 2025 at 02:53:37PM -0800, Yang Shi wrote:
>> On 2/12/25 6:04 PM, Oliver Sang wrote:
>>> hi, Yang Shi,
>>>
>>> On Fri, Feb 07, 2025 at 10:10:37AM -0800, Yang Shi wrote:
>>>> On 2/6/25 12:02 AM, Oliver Sang wrote:
>>> [...]
>>>
>>>>> since we applied your "/dev/zero: make private mapping full anonymous mapping"
>>>>> patch upon a68d3cbfad like below:
>>>>>
>>>>> * 7143ee2391f1e /dev/zero: make private mapping full anonymous mapping
>>>>> * a68d3cbfade64 memstick: core: fix kernel-doc notation
>>>>>
>>>>> so I applied below patch also upon a68d3cbfad.
>>>>>
>>>>> we saw big improvement but not that big.
>>>>>
>>>>> =========================================================================================
>>>>> compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
>>>>> gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
>>>>>
>>>>> commit:
>>>>> a68d3cbfad ("memstick: core: fix kernel-doc notation")
>>>>> 52ec85cb99 <--- your patch
>>>>>
>>>>>
>>>>> a68d3cbfade64392 52ec85cb99e9b31dc304eae965a
>>>>> ---------------- ---------------------------
>>>>> %stddev %change %stddev
>>>>> \ | \
>>>>> 14364828 ± 4% +410.6% 73349239 ± 3% vm-scalability.throughput
>>>>>
>>>>> full comparison as below [1] just FYI.
>>>> Thanks for the update. I stared at the profiling report for a whole day, but
>>>> I didn't figure out where that 400% lost. I just saw the number of page
>>>> faults was fewer. And it seems like the reduction of page faults match the
>>>> 400% loss. So I did more trace and profiling.
>>>>
>>>> The test case did the below stuff in a tight loop:
>>>> mmap 40K memory from /dev/zero (read only)
>>>> read the area
>>>>
>>>> So two major factors to the performance: mmap and page fault. The
>>>> alternative patch did reduce the overhead of mmap to the same level as the
>>>> original patch.
>>>>
>>>> The further perf profiling showed the cost of page fault is higher than the
>>>> original patch. But the profiling of page fault was interesting:
>>>>
>>>> - 44.87% 0.01% usemem [kernel.kallsyms] [k]
>>>> do_translation_fault
>>>> - 44.86% do_translation_fault
>>>> - 44.83% do_page_fault
>>>> - 44.53% handle_mm_fault
>>>> 9.04% __handle_mm_fault
>>>>
>>>> Page fault consumed 40% of cpu time in handle_mm_fault, but
>>>> __handle_mm_fault just consumed 9%, I expected it should be the major
>>>> consumer.
>>>>
>>>> So I annotated handle_mm_fault, then found the most time was consumed by
>>>> lru_gen_enter_fault() -> vma_has_recency() (my kernel has multi-gen LRU
>>>> enabled):
>>>>
>>>> │ if (vma->vm_file && (vma->vm_file->f_mode & FMODE_NOREUSE))
>>>> │ ↓ cbz x1, b4
>>>> 0.00 │ ldr w0, [x1, #12]
>>>> 99.59 │ eor x0, x0, #0x800000
>>>> 0.00 │ ubfx w0, w0, #23, #1
>>>> │ current->in_lru_fault = vma_has_recency(vma);
>>>> 0.00 │ b4: ldrh w1, [x2, #1992]
>>>> 0.01 │ bfi w1, w0, #5, #1
>>>> 0.00 │ strh w1, [x2, #1992]
>>>>
>>>>
>>>> vma_has_recency() read vma->vm_file->f_mode if vma->vm_file is not NULL. But
>>>> that load took a long time. So I inspected struct file and saw:
>>>>
>>>> struct file {
>>>> file_ref_t f_ref;
>>>> spinlock_t f_lock;
>>>> fmode_t f_mode;
>>>> const struct file_operations *f_op;
>>>> ...
>>>> }
>>>>
>>>> The f_mode is in the same cache line with f_ref (my kernel does NOT have
>>>> spin lock debug enabled). The test case mmap /dev/zero in a tight loop, so
>>>> the refcount is modified (fget/fput) very frequently, this resulted in
>>>> somehow false sharing.
>>>>
>>>> So I tried the below patch on top of the alternative patch:
>>>>
>>>> diff --git a/include/linux/mm_inline.h b/include/linux/mm_inline.h
>>>> index f9157a0c42a5..ba11dc0b1c7c 100644
>>>> --- a/include/linux/mm_inline.h
>>>> +++ b/include/linux/mm_inline.h
>>>> @@ -608,6 +608,9 @@ static inline bool vma_has_recency(struct vm_area_struct
>>>> *vma)
>>>> if (vma->vm_flags & (VM_SEQ_READ | VM_RAND_READ))
>>>> return false;
>>>>
>>>> + if (vma_is_anonymous(vma))
>>>> + return true;
>>>> +
>>>> if (vma->vm_file && (vma->vm_file->f_mode & FMODE_NOREUSE))
>>>> return false;
>>>>
>>>> This made the profiling of page fault look normal:
>>>>
>>>> - 1.90% do_translation_fault
>>>> - 1.87% do_page_fault
>>>> - 1.49% handle_mm_fault
>>>> - 1.36% __handle_mm_fault
>>>>
>>>> Please try this in your test.
>>>>
>>>> But AFAICT I have never seen performance issue reported due to the false
>>>> sharing of refcount and other fields in struct file. This benchmark stressed
>>>> this quite badly.
>>> I applied your above patch upon alternative patch last time, then saw more
>>> improvement (+445.2% vs a68d3cbfad), but still not that big as in our original
>>> report.
>> Thanks for the update. It looks like the problem is still in page faults. I
>> did my test on arm64 machine. I also noticed struct file has
>> "__randomize_layout", so it may have different layout on x86 than arm64?
>>
>> The page fault handler may also access other fields of struct file that may
>> cause false sharing, for example, accessing f_mapping to read gfp flags.
>> This may not be a problem on my machine, but may be more costly on yours
>> depending on the real layout of struct file on the machines,
>>
>> Can you please try the below patch on top of the current patches? Thank you
>> so much for your patience.
> you are welcome!
>
> now has more improvements. I just list "a68d3cbfad + 3 patches so far" vs
> a68d3cbfad below, if you want more data, please let me know.
>
> =========================================================================================
> compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
> gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
>
> commit:
> a68d3cbfad ("memstick: core: fix kernel-doc notation")
> edc84ea79f <--- a68d3cbfad + 3 patches so far
>
> a68d3cbfade64392 edc84ea79f8dc11853076b96ad5
> ---------------- ---------------------------
> %stddev %change %stddev
> \ | \
> 14364828 ± 4% +685.6% 1.129e+08 ± 5% vm-scalability.throughput
>
> full data is as below [1] FYI.
Thank you for the update. It is close to the 800% target, and it looks
like there may be still some overhead in page fault handler due to the
false sharing. For example, the vma_is_dax() call in
__thp_vma_allowable_orders() which is called if pmd is null. I'm not
sure how much the impact could be. However, I'm not sure whether we
should continue chasing it or not. Because the false sharing in struct
file should be very rare for real life workload. The workload has to map
the same file then do page fault again and again in a tight loop, and
the struct file is shared by multiple processes. Such behavior should be
rare in real life.
And changing the layout of struct file to avoid the false sharing sounds
better than adding vma_is_anonymous() call in all the possible places.
But it may introduce new false sharing. Having refcount in a dedicated
cache line is doable too, however it will increase the size of struct
file (from 192 bytes to 256 bytes). So neither seems worth it.
We can split all the patches into two parts, the first part is to avoid
i_mmap_rwsem contention, the second part is the struct file false
sharing. IMHO the first part is more real. I can come up with a formal
patch then send to the mailing list
Thanks,
Yang
>> diff --git a/mm/memory.c b/mm/memory.c
>> index 539c0f7c6d54..1fa9dbce0f66 100644
>> --- a/mm/memory.c
>> +++ b/mm/memory.c
>> @@ -3214,6 +3214,9 @@ static gfp_t __get_fault_gfp_mask(struct
>> vm_area_struct *vma)
>> {
>> struct file *vm_file = vma->vm_file;
>>
>> + if (vma_is_anonymous(vma))
>> + return GFP_KERNEL;
>> +
>> if (vm_file)
>> return mapping_gfp_mask(vm_file->f_mapping) | __GFP_FS |
>> __GFP_IO;
>>
> [1]
> =========================================================================================
> compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
> gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability
>
> commit:
> a68d3cbfad ("memstick: core: fix kernel-doc notation")
> edc84ea79f <--- a68d3cbfad + 3 patches so far
>
> a68d3cbfade64392 edc84ea79f8dc11853076b96ad5
> ---------------- ---------------------------
> %stddev %change %stddev
> \ | \
> 5.262e+09 ± 3% -59.8% 2.114e+09 ± 2% cpuidle..time
> 7924008 ± 3% -83.9% 1275131 ± 5% cpuidle..usage
> 1871164 ± 4% -16.8% 1557233 ± 8% numa-numastat.node3.local_node
> 1952164 ± 3% -14.8% 1663189 ± 7% numa-numastat.node3.numa_hit
> 399.52 -75.0% 99.77 ± 2% uptime.boot
> 14507 -22.1% 11296 uptime.idle
> 3408 ± 5% -99.8% 7.25 ± 46% perf-c2c.DRAM.local
> 18076 ± 3% -99.8% 43.00 ±100% perf-c2c.DRAM.remote
> 8082 ± 5% -99.8% 12.50 ± 63% perf-c2c.HITM.local
> 6544 ± 6% -99.7% 22.88 ±151% perf-c2c.HITM.remote
> 14627 ± 4% -99.8% 35.38 ±114% perf-c2c.HITM.total
> 6.99 ± 3% +177.6% 19.41 ± 3% vmstat.cpu.id
> 91.35 -28.5% 65.31 vmstat.cpu.sy
> 1.71 +793.1% 15.25 ± 4% vmstat.cpu.us
> 34204 ± 5% -64.1% 12271 ± 9% vmstat.system.cs
> 266575 -21.2% 210049 vmstat.system.in
> 6.49 ± 3% +10.0 16.46 ± 3% mpstat.cpu.all.idle%
> 0.63 -0.3 0.34 ± 3% mpstat.cpu.all.irq%
> 0.03 ± 2% +0.3 0.31 ± 4% mpstat.cpu.all.soft%
> 91.17 -24.1 67.09 mpstat.cpu.all.sys%
> 1.68 ± 2% +14.1 15.80 ± 4% mpstat.cpu.all.usr%
> 337.33 -98.7% 4.25 ± 10% mpstat.max_utilization.seconds
> 352.76 -84.7% 53.95 ± 4% time.elapsed_time
> 352.76 -84.7% 53.95 ± 4% time.elapsed_time.max
> 225965 ± 7% -17.1% 187329 ± 12% time.involuntary_context_switches
> 9.592e+08 ± 4% +11.9% 1.074e+09 time.minor_page_faults
> 20852 -10.0% 18761 time.percent_of_cpu_this_job_got
> 72302 -88.6% 8227 ± 6% time.system_time
> 1260 ± 3% +50.7% 1899 time.user_time
> 5393707 ± 5% -98.8% 66895 ± 21% time.voluntary_context_switches
> 1609925 -50.7% 793216 meminfo.Active
> 1609925 -50.7% 793216 meminfo.Active(anon)
> 160837 ± 33% -72.5% 44155 ± 9% meminfo.AnonHugePages
> 4435665 -18.7% 3608195 meminfo.Cached
> 1775547 -44.2% 990889 meminfo.Committed_AS
> 148539 -47.4% 78096 meminfo.Mapped
> 4245538 ± 4% -24.6% 3202495 meminfo.PageTables
> 929777 -88.9% 102759 meminfo.Shmem
> 25676018 ± 3% +14.3% 29335678 meminfo.max_used_kB
> 64129 ± 4% +706.8% 517389 ± 7% vm-scalability.median
> 45.40 ± 5% +2248.9 2294 ± 2% vm-scalability.stddev%
> 14364828 ± 4% +685.6% 1.129e+08 ± 5% vm-scalability.throughput
> 352.76 -84.7% 53.95 ± 4% vm-scalability.time.elapsed_time
> 352.76 -84.7% 53.95 ± 4% vm-scalability.time.elapsed_time.max
> 225965 ± 7% -17.1% 187329 ± 12% vm-scalability.time.involuntary_context_switches
> 9.592e+08 ± 4% +11.9% 1.074e+09 vm-scalability.time.minor_page_faults
> 20852 -10.0% 18761 vm-scalability.time.percent_of_cpu_this_job_got
> 72302 -88.6% 8227 ± 6% vm-scalability.time.system_time
> 1260 ± 3% +50.7% 1899 vm-scalability.time.user_time
> 5393707 ± 5% -98.8% 66895 ± 21% vm-scalability.time.voluntary_context_switches
> 4.316e+09 ± 4% +11.9% 4.832e+09 vm-scalability.workload
> 1063552 ± 4% -24.9% 799008 ± 3% numa-meminfo.node0.PageTables
> 125455 ±106% -85.5% 18164 ±165% numa-meminfo.node0.Shmem
> 1062709 ± 4% -25.7% 789746 ± 4% numa-meminfo.node1.PageTables
> 176171 ± 71% -92.4% 13303 ±230% numa-meminfo.node1.Shmem
> 35515 ± 91% -97.3% 976.55 ± 59% numa-meminfo.node2.Mapped
> 1058901 ± 4% -25.3% 791392 ± 4% numa-meminfo.node2.PageTables
> 770405 ± 30% -79.2% 160245 ±101% numa-meminfo.node3.Active
> 770405 ± 30% -79.2% 160245 ±101% numa-meminfo.node3.Active(anon)
> 380096 ± 50% -62.5% 142513 ± 98% numa-meminfo.node3.AnonPages.max
> 1146977 ±108% -92.8% 82894 ± 60% numa-meminfo.node3.FilePages
> 52663 ± 47% -97.2% 1488 ± 39% numa-meminfo.node3.Mapped
> 1058539 ± 4% -22.3% 821992 ± 3% numa-meminfo.node3.PageTables
> 558943 ± 14% -93.7% 35227 ±124% numa-meminfo.node3.Shmem
> 265763 ± 4% -24.9% 199601 ± 3% numa-vmstat.node0.nr_page_table_pages
> 31364 ±106% -85.5% 4539 ±165% numa-vmstat.node0.nr_shmem
> 265546 ± 4% -25.5% 197854 ± 5% numa-vmstat.node1.nr_page_table_pages
> 44052 ± 71% -92.5% 3323 ±230% numa-vmstat.node1.nr_shmem
> 8961 ± 91% -97.3% 244.02 ± 59% numa-vmstat.node2.nr_mapped
> 264589 ± 4% -25.2% 197920 ± 3% numa-vmstat.node2.nr_page_table_pages
> 192683 ± 30% -79.2% 40126 ±101% numa-vmstat.node3.nr_active_anon
> 286819 ±108% -92.8% 20761 ± 60% numa-vmstat.node3.nr_file_pages
> 13124 ± 49% -97.2% 372.02 ± 39% numa-vmstat.node3.nr_mapped
> 264499 ± 4% -22.4% 205376 ± 3% numa-vmstat.node3.nr_page_table_pages
> 139810 ± 14% -93.7% 8844 ±124% numa-vmstat.node3.nr_shmem
> 192683 ± 30% -79.2% 40126 ±101% numa-vmstat.node3.nr_zone_active_anon
> 1951359 ± 3% -14.9% 1661427 ± 7% numa-vmstat.node3.numa_hit
> 1870359 ± 4% -16.8% 1555470 ± 8% numa-vmstat.node3.numa_local
> 402515 -50.7% 198246 proc-vmstat.nr_active_anon
> 170568 +1.8% 173591 proc-vmstat.nr_anon_pages
> 1109246 -18.7% 902238 proc-vmstat.nr_file_pages
> 37525 -47.3% 19768 proc-vmstat.nr_mapped
> 1059932 ± 4% -24.2% 803105 ± 2% proc-vmstat.nr_page_table_pages
> 232507 -89.0% 25623 proc-vmstat.nr_shmem
> 37297 -5.4% 35299 proc-vmstat.nr_slab_reclaimable
> 402515 -50.7% 198246 proc-vmstat.nr_zone_active_anon
> 61931 ± 8% -83.9% 9948 ± 59% proc-vmstat.numa_hint_faults
> 15755 ± 21% -96.6% 541.38 ± 36% proc-vmstat.numa_hint_faults_local
> 6916516 ± 3% -8.0% 6360040 proc-vmstat.numa_hit
> 6568542 ± 3% -8.5% 6012265 proc-vmstat.numa_local
> 293942 ± 3% -68.8% 91724 ± 48% proc-vmstat.numa_pte_updates
> 9.608e+08 ± 4% +11.8% 1.074e+09 proc-vmstat.pgfault
> 55981 ± 2% -68.7% 17541 ± 2% proc-vmstat.pgreuse
> 0.82 ± 4% -51.0% 0.40 ± 8% perf-stat.i.MPKI
> 2.714e+10 ± 2% +378.3% 1.298e+11 ± 9% perf-stat.i.branch-instructions
> 0.11 ± 3% +0.1 0.24 ± 8% perf-stat.i.branch-miss-rate%
> 24932893 +306.8% 1.014e+08 ± 9% perf-stat.i.branch-misses
> 64.93 -7.5 57.48 perf-stat.i.cache-miss-rate%
> 88563288 ± 3% +35.0% 1.196e+08 ± 7% perf-stat.i.cache-misses
> 1.369e+08 ± 3% +43.7% 1.968e+08 ± 7% perf-stat.i.cache-references
> 34508 ± 4% -47.3% 18199 ± 9% perf-stat.i.context-switches
> 7.67 -75.7% 1.87 ± 3% perf-stat.i.cpi
> 224605 +22.5% 275084 ± 6% perf-stat.i.cpu-clock
> 696.35 ± 2% -53.5% 323.77 ± 2% perf-stat.i.cpu-migrations
> 10834 ± 4% -24.1% 8224 ± 11% perf-stat.i.cycles-between-cache-misses
> 1.102e+11 +282.2% 4.212e+11 ± 9% perf-stat.i.instructions
> 0.14 +334.6% 0.62 ± 5% perf-stat.i.ipc
> 24.25 ± 3% +626.9% 176.25 ± 4% perf-stat.i.metric.K/sec
> 2722043 ± 3% +803.8% 24600740 ± 9% perf-stat.i.minor-faults
> 2722043 ± 3% +803.8% 24600739 ± 9% perf-stat.i.page-faults
> 224605 +22.5% 275084 ± 6% perf-stat.i.task-clock
> 0.81 ± 3% -62.2% 0.31 ± 11% perf-stat.overall.MPKI
> 0.09 -0.0 0.08 ± 2% perf-stat.overall.branch-miss-rate%
> 64.81 -2.4 62.37 perf-stat.overall.cache-miss-rate%
> 7.24 -70.7% 2.12 ± 5% perf-stat.overall.cpi
> 8933 ± 4% -21.9% 6978 ± 7% perf-stat.overall.cycles-between-cache-misses
> 0.14 +242.2% 0.47 ± 5% perf-stat.overall.ipc
> 9012 ± 2% -57.8% 3806 perf-stat.overall.path-length
> 2.701e+10 ± 2% +285.4% 1.041e+11 ± 5% perf-stat.ps.branch-instructions
> 24708939 +215.8% 78042343 ± 4% perf-stat.ps.branch-misses
> 89032538 ± 3% +15.9% 1.032e+08 ± 8% perf-stat.ps.cache-misses
> 1.374e+08 ± 3% +20.6% 1.656e+08 ± 9% perf-stat.ps.cache-references
> 34266 ± 5% -66.2% 11570 ± 10% perf-stat.ps.context-switches
> 223334 -1.6% 219861 perf-stat.ps.cpu-clock
> 7.941e+11 -9.9% 7.157e+11 perf-stat.ps.cpu-cycles
> 693.54 ± 2% -67.2% 227.38 ± 4% perf-stat.ps.cpu-migrations
> 1.097e+11 +208.3% 3.381e+11 ± 5% perf-stat.ps.instructions
> 2710577 ± 3% +626.7% 19698901 ± 5% perf-stat.ps.minor-faults
> 2710577 ± 3% +626.7% 19698902 ± 5% perf-stat.ps.page-faults
> 223334 -1.6% 219861 perf-stat.ps.task-clock
> 3.886e+13 ± 2% -52.7% 1.839e+13 perf-stat.total.instructions
> 64052898 ± 5% -99.9% 81213 ± 23% sched_debug.cfs_rq:/.avg_vruntime.avg
> 95701822 ± 7% -96.4% 3425672 ± 7% sched_debug.cfs_rq:/.avg_vruntime.max
> 43098762 ± 6% -100.0% 153.42 ± 36% sched_debug.cfs_rq:/.avg_vruntime.min
> 9223270 ± 9% -95.9% 380347 ± 16% sched_debug.cfs_rq:/.avg_vruntime.stddev
> 0.00 ± 22% -100.0% 0.00 sched_debug.cfs_rq:/.h_nr_delayed.avg
> 0.69 ± 8% -100.0% 0.00 sched_debug.cfs_rq:/.h_nr_delayed.max
> 0.05 ± 12% -100.0% 0.00 sched_debug.cfs_rq:/.h_nr_delayed.stddev
> 0.78 ± 2% -94.5% 0.04 ± 21% sched_debug.cfs_rq:/.h_nr_running.avg
> 1.97 ± 5% -49.3% 1.00 sched_debug.cfs_rq:/.h_nr_running.max
> 0.28 ± 7% -29.1% 0.20 ± 10% sched_debug.cfs_rq:/.h_nr_running.stddev
> 411536 ± 58% -100.0% 1.15 ±182% sched_debug.cfs_rq:/.left_deadline.avg
> 43049468 ± 22% -100.0% 258.27 ±182% sched_debug.cfs_rq:/.left_deadline.max
> 3836405 ± 37% -100.0% 17.22 ±182% sched_debug.cfs_rq:/.left_deadline.stddev
> 411536 ± 58% -100.0% 1.06 ±191% sched_debug.cfs_rq:/.left_vruntime.avg
> 43049467 ± 22% -100.0% 236.56 ±191% sched_debug.cfs_rq:/.left_vruntime.max
> 3836405 ± 37% -100.0% 15.77 ±191% sched_debug.cfs_rq:/.left_vruntime.stddev
> 64052901 ± 5% -99.9% 81213 ± 23% sched_debug.cfs_rq:/.min_vruntime.avg
> 95701822 ± 7% -96.4% 3425672 ± 7% sched_debug.cfs_rq:/.min_vruntime.max
> 43098762 ± 6% -100.0% 153.42 ± 36% sched_debug.cfs_rq:/.min_vruntime.min
> 9223270 ± 9% -95.9% 380347 ± 16% sched_debug.cfs_rq:/.min_vruntime.stddev
> 0.77 ± 2% -94.4% 0.04 ± 21% sched_debug.cfs_rq:/.nr_running.avg
> 1.50 ± 9% -33.3% 1.00 sched_debug.cfs_rq:/.nr_running.max
> 0.26 ± 10% -22.7% 0.20 ± 10% sched_debug.cfs_rq:/.nr_running.stddev
> 1.61 ± 24% +413.4% 8.24 ± 60% sched_debug.cfs_rq:/.removed.runnable_avg.avg
> 86.69 +508.6% 527.62 ± 4% sched_debug.cfs_rq:/.removed.runnable_avg.max
> 11.14 ± 13% +428.4% 58.87 ± 32% sched_debug.cfs_rq:/.removed.runnable_avg.stddev
> 1.61 ± 24% +413.3% 8.24 ± 60% sched_debug.cfs_rq:/.removed.util_avg.avg
> 86.69 +508.6% 527.62 ± 4% sched_debug.cfs_rq:/.removed.util_avg.max
> 11.14 ± 13% +428.4% 58.87 ± 32% sched_debug.cfs_rq:/.removed.util_avg.stddev
> 411536 ± 58% -100.0% 1.06 ±191% sched_debug.cfs_rq:/.right_vruntime.avg
> 43049467 ± 22% -100.0% 236.56 ±191% sched_debug.cfs_rq:/.right_vruntime.max
> 3836405 ± 37% -100.0% 15.77 ±191% sched_debug.cfs_rq:/.right_vruntime.stddev
> 769.03 -84.7% 117.79 ± 3% sched_debug.cfs_rq:/.util_avg.avg
> 1621 ± 5% -32.7% 1092 ± 16% sched_debug.cfs_rq:/.util_avg.max
> 159.12 ± 8% +33.2% 211.88 ± 7% sched_debug.cfs_rq:/.util_avg.stddev
> 724.17 ± 2% -98.6% 10.41 ± 32% sched_debug.cfs_rq:/.util_est.avg
> 1360 ± 15% -51.5% 659.38 ± 10% sched_debug.cfs_rq:/.util_est.max
> 234.34 ± 9% -68.2% 74.43 ± 18% sched_debug.cfs_rq:/.util_est.stddev
> 766944 ± 3% +18.9% 912012 sched_debug.cpu.avg_idle.avg
> 1067639 ± 5% +25.5% 1339736 ± 9% sched_debug.cpu.avg_idle.max
> 3799 ± 7% -38.3% 2346 ± 23% sched_debug.cpu.avg_idle.min
> 321459 ± 2% -36.6% 203909 ± 7% sched_debug.cpu.avg_idle.stddev
> 195573 -76.9% 45144 sched_debug.cpu.clock.avg
> 195596 -76.9% 45160 sched_debug.cpu.clock.max
> 195548 -76.9% 45123 sched_debug.cpu.clock.min
> 13.79 ± 3% -36.0% 8.83 ± 2% sched_debug.cpu.clock.stddev
> 194424 -76.8% 45019 sched_debug.cpu.clock_task.avg
> 194608 -76.8% 45145 sched_debug.cpu.clock_task.max
> 181834 -82.1% 32559 sched_debug.cpu.clock_task.min
> 4241 ± 2% -96.8% 136.38 ± 21% sched_debug.cpu.curr->pid.avg
> 9799 ± 2% -59.8% 3934 sched_debug.cpu.curr->pid.max
> 1365 ± 10% -49.1% 695.11 ± 10% sched_debug.cpu.curr->pid.stddev
> 537665 ± 4% +28.3% 690006 ± 6% sched_debug.cpu.max_idle_balance_cost.max
> 3119 ± 56% +479.5% 18078 ± 29% sched_debug.cpu.max_idle_balance_cost.stddev
> 0.00 ± 12% -68.3% 0.00 ± 17% sched_debug.cpu.next_balance.stddev
> 0.78 ± 2% -95.3% 0.04 ± 20% sched_debug.cpu.nr_running.avg
> 2.17 ± 8% -53.8% 1.00 sched_debug.cpu.nr_running.max
> 0.29 ± 8% -35.4% 0.19 ± 9% sched_debug.cpu.nr_running.stddev
> 25773 ± 5% -97.0% 764.82 ± 3% sched_debug.cpu.nr_switches.avg
> 48669 ± 10% -77.2% 11080 ± 12% sched_debug.cpu.nr_switches.max
> 19006 ± 7% -99.2% 151.12 ± 15% sched_debug.cpu.nr_switches.min
> 4142 ± 8% -69.5% 1264 ± 6% sched_debug.cpu.nr_switches.stddev
> 0.07 ± 23% -93.3% 0.01 ± 53% sched_debug.cpu.nr_uninterruptible.avg
> 240.19 ± 16% -80.2% 47.50 ± 44% sched_debug.cpu.nr_uninterruptible.max
> -77.92 -88.1% -9.25 sched_debug.cpu.nr_uninterruptible.min
> 37.87 ± 5% -84.7% 5.78 ± 13% sched_debug.cpu.nr_uninterruptible.stddev
> 195549 -76.9% 45130 sched_debug.cpu_clk
> 194699 -77.3% 44280 sched_debug.ktime
> 0.00 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.avg
> 0.17 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.max
> 0.01 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.stddev
> 196368 -76.6% 45975 sched_debug.sched_clk
> 95.59 -95.6 0.00 perf-profile.calltrace.cycles-pp.__mmap
> 95.54 -95.5 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
> 95.54 -95.5 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap
> 94.54 -94.5 0.00 perf-profile.calltrace.cycles-pp.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
> 94.46 -94.4 0.07 ±264% perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
> 94.45 -94.0 0.41 ±158% perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 94.14 -93.9 0.29 ±134% perf-profile.calltrace.cycles-pp.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff
> 94.25 -93.8 0.41 ±158% perf-profile.calltrace.cycles-pp.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 93.79 -93.7 0.07 ±264% perf-profile.calltrace.cycles-pp.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff
> 93.44 -93.4 0.00 perf-profile.calltrace.cycles-pp.down_write.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap
> 93.40 -93.4 0.00 perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 93.33 -93.3 0.00 perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma
> 92.89 -92.9 0.00 perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file
> 0.00 +1.7 1.69 ± 65% perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64
> 0.00 +1.9 1.90 ± 55% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group
> 0.00 +1.9 1.90 ± 55% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
> 0.00 +1.9 1.93 ± 53% perf-profile.calltrace.cycles-pp.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +1.9 1.93 ± 53% perf-profile.calltrace.cycles-pp.seq_read_iter.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64
> 0.00 +2.0 1.99 ± 53% perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +2.0 2.02 ± 64% perf-profile.calltrace.cycles-pp.do_pte_missing.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
> 0.00 +2.3 2.27 ± 56% perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
> 0.00 +2.3 2.27 ± 56% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
> 0.00 +2.3 2.27 ± 56% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe._Fork
> 0.00 +2.3 2.27 ± 56% perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
> 0.00 +2.4 2.45 ± 53% perf-profile.calltrace.cycles-pp._Fork
> 0.00 +2.5 2.51 ± 52% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +2.5 2.51 ± 52% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64
> 0.00 +2.5 2.51 ± 52% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +2.5 2.51 ± 52% perf-profile.calltrace.cycles-pp.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +3.2 3.17 ± 42% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
> 0.00 +3.3 3.28 ± 52% perf-profile.calltrace.cycles-pp.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin
> 0.00 +3.3 3.28 ± 52% perf-profile.calltrace.cycles-pp.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin.handle_internal_command
> 0.00 +4.1 4.10 ± 45% perf-profile.calltrace.cycles-pp.__cmd_record.cmd_record.run_builtin.handle_internal_command.main
> 0.00 +4.1 4.10 ± 45% perf-profile.calltrace.cycles-pp.cmd_record.run_builtin.handle_internal_command.main
> 0.00 +4.8 4.80 ± 61% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter
> 0.00 +5.0 4.98 ± 69% perf-profile.calltrace.cycles-pp.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64
> 0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn
> 0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn.perf_mmap__push
> 0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen
> 0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.write.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist
> 0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record
> 0.00 +5.1 5.11 ± 47% perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit
> 0.00 +5.1 5.12 ± 70% perf-profile.calltrace.cycles-pp.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record
> 0.00 +6.1 6.08 ± 50% perf-profile.calltrace.cycles-pp.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter
> 0.00 +7.8 7.84 ± 21% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
> 0.00 +7.9 7.88 ± 20% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
> 0.00 +7.9 7.88 ± 20% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
> 0.00 +7.9 7.88 ± 20% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
> 0.00 +7.9 7.88 ± 20% perf-profile.calltrace.cycles-pp.read
> 0.00 +11.1 11.10 ± 41% perf-profile.calltrace.cycles-pp.handle_internal_command.main
> 0.00 +11.1 11.10 ± 41% perf-profile.calltrace.cycles-pp.main
> 0.00 +11.1 11.10 ± 41% perf-profile.calltrace.cycles-pp.run_builtin.handle_internal_command.main
> 0.00 +11.2 11.18 ± 73% perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
> 0.00 +15.9 15.94 ± 41% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +15.9 15.94 ± 41% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
> 0.00 +19.5 19.54 ± 41% perf-profile.calltrace.cycles-pp.asm_sysvec_reschedule_ipi.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
> 1.21 ± 3% +36.7 37.86 ± 7% perf-profile.calltrace.cycles-pp.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
> 1.21 ± 3% +36.7 37.86 ± 7% perf-profile.calltrace.cycles-pp.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
> 1.21 ± 3% +37.0 38.24 ± 7% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
> 1.21 ± 3% +37.2 38.41 ± 7% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.common_startup_64
> 1.21 ± 3% +37.4 38.57 ± 6% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
> 1.22 ± 3% +38.5 39.67 ± 7% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.common_startup_64
> 1.22 ± 3% +38.5 39.67 ± 7% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.common_startup_64
> 1.22 ± 3% +38.5 39.67 ± 7% perf-profile.calltrace.cycles-pp.start_secondary.common_startup_64
> 1.22 ± 3% +38.9 40.09 ± 6% perf-profile.calltrace.cycles-pp.common_startup_64
> 2.19 ± 3% +45.2 47.41 ± 14% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state
> 95.60 -95.4 0.22 ±135% perf-profile.children.cycles-pp.__mmap
> 94.55 -93.9 0.60 ±103% perf-profile.children.cycles-pp.ksys_mmap_pgoff
> 94.14 -93.7 0.44 ±112% perf-profile.children.cycles-pp.__mmap_new_vma
> 93.79 -93.7 0.10 ±264% perf-profile.children.cycles-pp.vma_link_file
> 94.46 -93.5 0.96 ± 76% perf-profile.children.cycles-pp.vm_mmap_pgoff
> 94.45 -93.5 0.96 ± 76% perf-profile.children.cycles-pp.do_mmap
> 94.25 -93.4 0.86 ± 87% perf-profile.children.cycles-pp.__mmap_region
> 93.40 -93.4 0.00 perf-profile.children.cycles-pp.rwsem_down_write_slowpath
> 93.33 -93.3 0.00 perf-profile.children.cycles-pp.rwsem_optimistic_spin
> 93.44 -93.2 0.22 ±149% perf-profile.children.cycles-pp.down_write
> 92.91 -92.9 0.00 perf-profile.children.cycles-pp.osq_lock
> 95.58 -45.4 50.16 ± 8% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
> 95.58 -45.4 50.16 ± 8% perf-profile.children.cycles-pp.do_syscall_64
> 0.00 +1.1 1.12 ± 74% perf-profile.children.cycles-pp.filemap_map_pages
> 0.00 +1.1 1.12 ± 76% perf-profile.children.cycles-pp.vfs_fstatat
> 0.00 +1.2 1.19 ± 35% perf-profile.children.cycles-pp.vsnprintf
> 0.00 +1.2 1.20 ± 46% perf-profile.children.cycles-pp.seq_printf
> 0.00 +1.3 1.28 ± 78% perf-profile.children.cycles-pp.__do_sys_newfstatat
> 0.00 +1.5 1.54 ± 75% perf-profile.children.cycles-pp.folios_put_refs
> 0.00 +1.6 1.56 ± 52% perf-profile.children.cycles-pp.__cond_resched
> 0.00 +1.6 1.60 ± 32% perf-profile.children.cycles-pp.sched_balance_newidle
> 0.00 +1.7 1.69 ± 65% perf-profile.children.cycles-pp.dup_mm
> 0.00 +1.9 1.93 ± 53% perf-profile.children.cycles-pp.proc_reg_read_iter
> 0.00 +2.0 1.99 ± 53% perf-profile.children.cycles-pp.copy_process
> 0.00 +2.1 2.06 ± 51% perf-profile.children.cycles-pp.__x64_sys_ioctl
> 0.00 +2.1 2.08 ± 45% perf-profile.children.cycles-pp.proc_single_show
> 0.00 +2.1 2.14 ± 45% perf-profile.children.cycles-pp.seq_read
> 0.00 +2.2 2.16 ± 47% perf-profile.children.cycles-pp.ioctl
> 0.00 +2.2 2.17 ± 33% perf-profile.children.cycles-pp.schedule
> 0.00 +2.2 2.20 ± 28% perf-profile.children.cycles-pp.__pick_next_task
> 0.00 +2.2 2.21 ± 47% perf-profile.children.cycles-pp.perf_evsel__run_ioctl
> 0.00 +2.3 2.26 ± 58% perf-profile.children.cycles-pp.do_read_fault
> 0.00 +2.3 2.27 ± 56% perf-profile.children.cycles-pp.__do_sys_clone
> 0.00 +2.3 2.27 ± 56% perf-profile.children.cycles-pp.kernel_clone
> 0.00 +2.4 2.37 ± 58% perf-profile.children.cycles-pp.zap_present_ptes
> 0.00 +2.4 2.45 ± 53% perf-profile.children.cycles-pp._Fork
> 0.00 +2.6 2.59 ± 53% perf-profile.children.cycles-pp.__x64_sys_exit_group
> 0.00 +2.6 2.59 ± 53% perf-profile.children.cycles-pp.x64_sys_call
> 0.00 +2.6 2.64 ± 44% perf-profile.children.cycles-pp.do_pte_missing
> 0.00 +3.1 3.13 ± 59% perf-profile.children.cycles-pp.zap_pte_range
> 0.00 +3.2 3.21 ± 58% perf-profile.children.cycles-pp.zap_pmd_range
> 0.00 +3.4 3.40 ± 56% perf-profile.children.cycles-pp.unmap_page_range
> 0.00 +3.4 3.43 ± 55% perf-profile.children.cycles-pp.unmap_vmas
> 0.19 ± 23% +3.9 4.06 ± 45% perf-profile.children.cycles-pp.__handle_mm_fault
> 0.51 ± 6% +4.0 4.49 ± 38% perf-profile.children.cycles-pp.handle_mm_fault
> 0.04 ± 44% +4.0 4.04 ± 28% perf-profile.children.cycles-pp.__schedule
> 0.77 ± 3% +4.4 5.18 ± 39% perf-profile.children.cycles-pp.exc_page_fault
> 0.76 ± 3% +4.4 5.18 ± 39% perf-profile.children.cycles-pp.do_user_addr_fault
> 0.58 ± 2% +4.7 5.26 ± 53% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
> 0.00 +5.1 5.07 ± 71% perf-profile.children.cycles-pp.writen
> 0.00 +5.1 5.07 ± 69% perf-profile.children.cycles-pp.generic_perform_write
> 0.00 +5.1 5.12 ± 47% perf-profile.children.cycles-pp.exit_mm
> 0.00 +5.1 5.12 ± 70% perf-profile.children.cycles-pp.record__pushfn
> 0.00 +5.1 5.12 ± 70% perf-profile.children.cycles-pp.shmem_file_write_iter
> 1.18 +5.5 6.69 ± 33% perf-profile.children.cycles-pp.asm_exc_page_fault
> 0.00 +6.2 6.24 ± 43% perf-profile.children.cycles-pp.__mmput
> 0.00 +6.2 6.24 ± 43% perf-profile.children.cycles-pp.exit_mmap
> 0.00 +7.0 7.00 ± 51% perf-profile.children.cycles-pp.perf_mmap__push
> 0.00 +7.0 7.00 ± 51% perf-profile.children.cycles-pp.record__mmap_read_evlist
> 0.00 +7.2 7.25 ± 52% perf-profile.children.cycles-pp.__fput
> 0.00 +7.3 7.35 ± 20% perf-profile.children.cycles-pp.seq_read_iter
> 0.00 +7.8 7.84 ± 21% perf-profile.children.cycles-pp.vfs_read
> 0.00 +7.9 7.88 ± 20% perf-profile.children.cycles-pp.ksys_read
> 0.00 +7.9 7.88 ± 20% perf-profile.children.cycles-pp.read
> 0.00 +9.9 9.93 ± 41% perf-profile.children.cycles-pp.asm_sysvec_reschedule_ipi
> 0.02 ±141% +11.1 11.10 ± 41% perf-profile.children.cycles-pp.__cmd_record
> 0.02 ±141% +11.1 11.10 ± 41% perf-profile.children.cycles-pp.cmd_record
> 0.02 ±141% +11.1 11.10 ± 41% perf-profile.children.cycles-pp.handle_internal_command
> 0.02 ±141% +11.1 11.10 ± 41% perf-profile.children.cycles-pp.main
> 0.02 ±141% +11.1 11.10 ± 41% perf-profile.children.cycles-pp.run_builtin
> 0.00 +11.2 11.18 ± 73% perf-profile.children.cycles-pp.vfs_write
> 0.00 +11.2 11.23 ± 73% perf-profile.children.cycles-pp.ksys_write
> 0.00 +11.2 11.23 ± 73% perf-profile.children.cycles-pp.write
> 0.00 +13.6 13.61 ± 44% perf-profile.children.cycles-pp.do_exit
> 0.00 +13.6 13.61 ± 44% perf-profile.children.cycles-pp.do_group_exit
> 1.70 ± 2% +25.0 26.72 ± 15% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
> 1.21 ± 3% +36.6 37.81 ± 7% perf-profile.children.cycles-pp.acpi_safe_halt
> 1.21 ± 3% +36.6 37.86 ± 7% perf-profile.children.cycles-pp.acpi_idle_do_entry
> 1.21 ± 3% +36.6 37.86 ± 7% perf-profile.children.cycles-pp.acpi_idle_enter
> 1.21 ± 3% +37.4 38.57 ± 6% perf-profile.children.cycles-pp.cpuidle_enter_state
> 1.21 ± 3% +37.4 38.66 ± 6% perf-profile.children.cycles-pp.cpuidle_enter
> 1.22 ± 3% +37.6 38.82 ± 6% perf-profile.children.cycles-pp.cpuidle_idle_call
> 1.22 ± 3% +38.5 39.67 ± 7% perf-profile.children.cycles-pp.start_secondary
> 1.22 ± 3% +38.9 40.09 ± 6% perf-profile.children.cycles-pp.common_startup_64
> 1.22 ± 3% +38.9 40.09 ± 6% perf-profile.children.cycles-pp.cpu_startup_entry
> 1.22 ± 3% +38.9 40.09 ± 6% perf-profile.children.cycles-pp.do_idle
> 92.37 -92.4 0.00 perf-profile.self.cycles-pp.osq_lock
> 1.19 ± 3% +30.7 31.90 ± 7% perf-profile.self.cycles-pp.acpi_safe_halt
> 0.17 ±142% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
> 0.19 ± 34% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 0.14 ± 55% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
> 0.14 ± 73% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
> 0.10 ± 66% -99.9% 0.00 ±264% perf-sched.sch_delay.avg.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
> 0.11 ± 59% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 0.04 ±132% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
> 0.07 ±101% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 0.02 ± 31% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
> 0.02 ±143% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
> 0.10 ± 44% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
> 0.12 ±145% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
> 0.04 ± 55% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 0.25 ± 41% -98.5% 0.00 ±105% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
> 0.11 ± 59% -97.1% 0.00 ± 61% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
> 0.40 ± 50% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 0.32 ±104% -100.0% 0.00 perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
> 0.01 ± 12% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 0.08 ± 28% -99.5% 0.00 ±264% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 0.18 ± 57% -96.8% 0.01 ±193% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
> 0.03 ± 83% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
> 0.01 ± 20% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
> 0.02 ± 65% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
> 0.32 ± 47% -98.2% 0.01 ± 42% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 0.19 ±185% -96.5% 0.01 ± 33% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 0.07 ± 20% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 0.26 ± 17% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
> 0.02 ± 60% -94.2% 0.00 ±264% perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
> 0.01 ±128% -100.0% 0.00 perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
> 1.00 ±151% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
> 25.45 ± 94% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 4.56 ± 67% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
> 3.55 ± 97% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
> 2.13 ± 67% -100.0% 0.00 ±264% perf-sched.sch_delay.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
> 3.16 ± 78% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 0.30 ±159% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
> 1.61 ±100% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 0.03 ± 86% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
> 0.20 ±182% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
> 3.51 ± 21% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
> 0.83 ±160% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
> 0.09 ± 31% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 3.59 ± 11% -99.9% 0.00 ±105% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
> 1.60 ± 69% -99.6% 0.01 ±129% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
> 0.81 ± 43% -100.0% 0.00 perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 1.02 ± 88% -100.0% 0.00 perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
> 0.02 ± 7% -100.0% 0.00 perf-sched.sch_delay.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 9.68 ± 32% -100.0% 0.00 ±264% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 12.26 ±109% -100.0% 0.01 ±193% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
> 5.60 ±139% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
> 0.03 ±106% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
> 2.11 ± 61% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
> 3.67 ± 25% -99.8% 0.01 ± 16% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 1.65 ±187% -99.3% 0.01 ± 23% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 37.84 ± 47% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 4.68 ± 36% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
> 0.21 ±169% -99.6% 0.00 ±264% perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
> 7.92 ±131% -99.2% 0.06 ± 92% perf-sched.sch_delay.max.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 0.36 ±186% -100.0% 0.00 perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
> 33.45 ± 3% -91.6% 2.81 ± 90% perf-sched.total_wait_and_delay.average.ms
> 97903 ± 4% -98.2% 1776 ± 28% perf-sched.total_wait_and_delay.count.ms
> 2942 ± 23% -95.2% 141.09 ± 36% perf-sched.total_wait_and_delay.max.ms
> 33.37 ± 3% -91.9% 2.69 ± 95% perf-sched.total_wait_time.average.ms
> 2942 ± 23% -96.7% 97.14 ± 19% perf-sched.total_wait_time.max.ms
> 3.97 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 3.08 ± 4% -94.3% 0.18 ± 92% perf-sched.wait_and_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
> 119.91 ± 38% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 433.73 ± 41% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 302.41 ± 5% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 1.48 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 23.24 ± 25% -96.7% 0.76 ± 27% perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
> 327.16 ± 9% -99.8% 0.76 ±188% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 369.37 ± 2% -98.9% 4.03 ±204% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 0.96 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 453.60 -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
> 187.66 -96.7% 6.11 ±109% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 2.37 ± 29% -99.6% 0.01 ±264% perf-sched.wait_and_delay.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 750.07 -99.3% 5.10 ± 84% perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 1831 ± 9% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 1269 ± 8% -45.8% 688.12 ± 21% perf-sched.wait_and_delay.count.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
> 6.17 ± 45% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 5.00 -100.0% 0.00 perf-sched.wait_and_delay.count.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 14.33 ± 5% -100.0% 0.00 perf-sched.wait_and_delay.count.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 810.00 ± 10% -100.0% 0.00 perf-sched.wait_and_delay.count.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 3112 ± 24% -97.9% 65.75 ±106% perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64
> 40.50 ± 8% -98.8% 0.50 ±173% perf-sched.wait_and_delay.count.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 73021 ± 3% -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 40.00 -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_timeout.kcompactd.kthread.ret_from_fork
> 1122 -99.0% 10.88 ± 98% perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
> 11323 ± 3% -93.6% 722.25 ± 20% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 1887 ± 45% -100.0% 0.88 ±264% perf-sched.wait_and_delay.count.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 1238 -93.9% 75.62 ± 79% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 35.19 ± 57% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 1002 -91.0% 89.82 ± 93% perf-sched.wait_and_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
> 318.48 ± 65% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 1000 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 966.90 ± 7% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 20.79 ± 19% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 1043 -98.4% 16.64 ±214% perf-sched.wait_and_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
> 1240 ± 20% -99.9% 1.52 ±188% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 500.34 -96.9% 15.38 ±232% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 58.83 ± 39% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 505.17 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
> 19.77 ± 55% -62.8% 7.36 ± 85% perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
> 1237 ± 34% -91.7% 102.88 ± 33% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 1001 -100.0% 0.05 ±264% perf-sched.wait_and_delay.max.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 2794 ± 24% -97.9% 59.20 ± 61% perf-sched.wait_and_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 49.27 ±119% -100.0% 0.01 ±264% perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
> 58.17 ±187% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
> 3.78 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 2.99 ± 4% -97.0% 0.09 ± 91% perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
> 3.92 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
> 4.71 ± 8% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
> 1.67 ± 20% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
> 2.10 ± 27% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 0.01 ± 44% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
> 1.67 ± 21% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 0.04 ±133% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
> 67.14 ± 73% -99.5% 0.32 ±177% perf-sched.wait_time.avg.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
> 1.65 ± 67% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
> 2.30 ± 14% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
> 42.44 ±200% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
> 152.73 ±152% -100.0% 0.06 ±249% perf-sched.wait_time.avg.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin
> 119.87 ± 38% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 3.80 ± 18% -99.9% 0.00 ±105% perf-sched.wait_time.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
> 433.32 ± 41% -100.0% 0.00 perf-sched.wait_time.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 250.23 ±107% -100.0% 0.00 perf-sched.wait_time.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
> 29.19 ± 5% -99.2% 0.25 ± 24% perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
> 302.40 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 1.40 ± 6% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 4.03 ± 8% -99.9% 0.01 ±193% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
> 35.38 ±192% -100.0% 0.00 ±264% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
> 0.05 ± 40% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
> 0.72 ±220% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
> 1.00 ±120% -99.9% 0.00 ±264% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
> 23.07 ± 24% -97.1% 0.67 ± 10% perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
> 326.84 ± 9% -99.6% 1.19 ±108% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 369.18 ± 2% -98.7% 4.72 ±167% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 0.89 ± 6% -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 1.17 ± 16% -99.7% 0.00 ±264% perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
> 453.58 -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
> 4.42 -25.4% 3.30 ± 17% perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
> 187.58 -96.8% 6.05 ±110% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 2.36 ± 29% -99.1% 0.02 ± 84% perf-sched.wait_time.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 0.01 ±156% -100.0% 0.00 perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
> 750.01 -99.5% 3.45 ±141% perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 340.69 ±135% -100.0% 0.01 ±264% perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio
> 535.09 ±128% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault
> 22.04 ± 32% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region
> 1001 -95.5% 44.91 ± 93% perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
> 13.57 ± 17% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate
> 13.54 ± 10% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
> 10.17 ± 19% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap
> 11.35 ± 25% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region
> 0.01 ± 32% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
> 10.62 ± 9% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 0.20 ±199% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
> 1559 ± 64% -100.0% 0.44 ±167% perf-sched.wait_time.max.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
> 6.93 ± 53% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma
> 14.42 ± 22% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region
> 159.10 ±148% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
> 391.02 ±171% -100.0% 0.12 ±256% perf-sched.wait_time.max.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin
> 318.43 ± 65% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 13.14 ± 21% -100.0% 0.00 ±105% perf-sched.wait_time.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
> 1000 -100.0% 0.00 perf-sched.wait_time.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 500.84 ± 99% -100.0% 0.00 perf-sched.wait_time.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
> 641.50 ± 23% -99.2% 5.27 ± 76% perf-sched.wait_time.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
> 10.75 ± 98% -89.8% 1.10 ± 78% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 966.89 ± 7% -100.0% 0.00 perf-sched.wait_time.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
> 15.80 ± 8% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
> 16.69 ± 10% -100.0% 0.01 ±193% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
> 41.71 ±158% -100.0% 0.00 ±264% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
> 11.64 ± 61% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
> 2.94 ±213% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
> 175.70 ±210% -100.0% 0.00 ±264% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
> 1043 -99.6% 4.46 ±105% perf-sched.wait_time.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
> 1240 ± 20% -99.8% 2.37 ±108% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
> 500.11 -96.5% 17.32 ±201% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
> 32.65 ± 33% -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file
> 22.94 ± 56% -100.0% 0.00 ±264% perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
> 505.00 -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
> 12.20 ± 43% -59.2% 4.98 perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
> 1237 ± 34% -92.5% 92.94 ± 20% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
> 1000 -100.0% 0.09 ±111% perf-sched.wait_time.max.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
> 0.36 ±190% -100.0% 0.00 perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
> 2794 ± 24% -98.9% 30.12 ±114% perf-sched.wait_time.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
^ permalink raw reply [flat|nested] 35+ messages in thread
end of thread, other threads:[~2025-02-19 1:12 UTC | newest]
Thread overview: 35+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2025-01-13 22:30 [PATCH] /dev/zero: make private mapping full anonymous mapping Yang Shi
2025-01-14 12:05 ` Lorenzo Stoakes
2025-01-14 16:53 ` Yang Shi
2025-01-14 18:14 ` Lorenzo Stoakes
2025-01-14 18:19 ` Lorenzo Stoakes
2025-01-14 18:21 ` Lorenzo Stoakes
2025-01-14 18:22 ` Matthew Wilcox
2025-01-14 18:26 ` Lorenzo Stoakes
2025-01-14 18:32 ` Jann Horn
2025-01-14 18:38 ` Lorenzo Stoakes
2025-01-14 19:03 ` Yang Shi
2025-01-14 19:13 ` Lorenzo Stoakes
2025-01-14 21:24 ` Yang Shi
2025-01-15 12:10 ` Lorenzo Stoakes
2025-01-15 21:29 ` Yang Shi
2025-01-15 22:05 ` Christoph Lameter (Ampere)
2025-01-14 13:01 ` David Hildenbrand
2025-01-14 14:52 ` Lorenzo Stoakes
2025-01-14 15:06 ` David Hildenbrand
2025-01-14 17:01 ` Yang Shi
2025-01-14 17:23 ` David Hildenbrand
2025-01-14 17:38 ` Yang Shi
2025-01-14 17:46 ` David Hildenbrand
2025-01-14 18:05 ` Yang Shi
2025-01-14 17:02 ` David Hildenbrand
2025-01-14 17:20 ` Yang Shi
2025-01-14 17:24 ` David Hildenbrand
2025-01-28 3:14 ` kernel test robot
2025-01-31 18:38 ` Yang Shi
2025-02-06 8:02 ` Oliver Sang
2025-02-07 18:10 ` Yang Shi
2025-02-13 2:04 ` Oliver Sang
2025-02-14 22:53 ` Yang Shi
2025-02-18 6:30 ` Oliver Sang
2025-02-19 1:12 ` Yang Shi
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox