From: "Yin, Fengwei" <fengwei.yin@intel.com>
To: Matthew Wilcox <willy@infradead.org>, Yang Shi <shy828301@gmail.com>
Cc: kernel test robot <oliver.sang@intel.com>,
Rik van Riel <riel@surriel.com>, <oe-lkp@lists.linux.dev>,
<lkp@intel.com>,
"Linux Memory Management List" <linux-mm@kvack.org>,
Andrew Morton <akpm@linux-foundation.org>,
Christopher Lameter <cl@linux.com>, <ying.huang@intel.com>,
<feng.tang@intel.com>
Subject: Re: [linux-next:master] [mm] 1111d46b5c: stress-ng.pthread.ops_per_sec -84.3% regression
Date: Fri, 22 Dec 2023 09:06:03 +0800 [thread overview]
Message-ID: <ef768b7e-0c91-4ccc-b083-42e05da752f6@intel.com> (raw)
In-Reply-To: <ZYSAcThKd/FbH/Bt@casper.infradead.org>
On 12/22/2023 2:14 AM, Matthew Wilcox wrote:
> On Thu, Dec 21, 2023 at 10:07:09AM -0800, Yang Shi wrote:
>> On Wed, Dec 20, 2023 at 8:49 PM Matthew Wilcox <willy@infradead.org> wrote:
>>>
>>> On Thu, Dec 21, 2023 at 08:58:42AM +0800, Yin Fengwei wrote:
>>>> Yes. MAP_STACK is also mentioned in manpage of mmap. I did test to
>>>> filter out of the MAP_STACK mapping based on this patch. The regression
>>>> in stress-ng.pthread was gone. I suppose this is kind of safe because
>>>> the madvise call is only applied to glibc allocated stack.
>>>>
>>>>
>>>> But what I am not sure was whether it's worthy to do such kind of change
>>>> as the regression only is seen obviously in micro-benchmark. No evidence
>>>> showed the other regressionsin this report is related with madvise. At
>>>> least from the perf statstics. Need to check more on stream/ramspeed.
>>>
>>> FWIW, we had a customer report a significant performance problem when
>>> inadvertently using 2MB pages for stacks. They were able to avoid it by
>>> using 2044KiB sized stacks ...
>>
>> Thanks for the report. This provided more justification regarding
>> honoring MAP_STACK on Linux. Some applications, for example, pthread,
>> just allocate a fixed size area for stack. This confuses kernel
>> because kernel tell stack by VM_GROWSDOWN | VM_GROWSUP.
>>
>> But I'm still a little confused by why THP for stack could result in
>> significant performance problems. Unless the applications resize the
>> stack quite often.
>
> We didn't delve into what was causing the problem, only that it was
> happening. The application had many threads, so it could have been as
> simple as consuming all the available THP and leaving fewer available
> for other uses. Or it could have been a memory consumption problem;
> maybe the app would only have been using 16-32kB per thread but was
> now using 2MB per thread and if there were, say, 100 threads, that's an
> extra 199MB of memory in use.
One thing I know is related with the memory zeroing. This is from
the perf data in this report:
0.00 +16.7 16.69 ± 7%
perf-profile.calltrace.cycles-pp.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
Zeroing 2M memory costs much more CPU than zeroing 16-32KB memory if
there are many threads.
Regards
Yin, Fengwei
next prev parent reply other threads:[~2023-12-22 1:06 UTC|newest]
Thread overview: 24+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-12-19 15:41 kernel test robot
2023-12-20 5:27 ` Yang Shi
2023-12-20 8:29 ` Yin Fengwei
2023-12-20 15:42 ` Christoph Lameter (Ampere)
2023-12-20 20:14 ` Yang Shi
2023-12-20 20:09 ` Yang Shi
2023-12-21 0:26 ` Yang Shi
2023-12-21 0:58 ` Yin Fengwei
2023-12-21 1:02 ` Yin Fengwei
2023-12-21 4:49 ` Matthew Wilcox
2023-12-21 4:58 ` Yin Fengwei
2023-12-21 18:07 ` Yang Shi
2023-12-21 18:14 ` Matthew Wilcox
2023-12-22 1:06 ` Yin, Fengwei [this message]
2023-12-22 2:23 ` Huang, Ying
2023-12-21 13:39 ` Yin, Fengwei
2023-12-21 18:11 ` Yang Shi
2023-12-22 1:13 ` Yin, Fengwei
2024-01-04 1:32 ` Yang Shi
2024-01-04 8:18 ` Yin Fengwei
2024-01-04 8:39 ` Oliver Sang
2024-01-05 9:29 ` Oliver Sang
2024-01-05 14:52 ` Yin, Fengwei
2024-01-05 18:49 ` Yang Shi
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=ef768b7e-0c91-4ccc-b083-42e05da752f6@intel.com \
--to=fengwei.yin@intel.com \
--cc=akpm@linux-foundation.org \
--cc=cl@linux.com \
--cc=feng.tang@intel.com \
--cc=linux-mm@kvack.org \
--cc=lkp@intel.com \
--cc=oe-lkp@lists.linux.dev \
--cc=oliver.sang@intel.com \
--cc=riel@surriel.com \
--cc=shy828301@gmail.com \
--cc=willy@infradead.org \
--cc=ying.huang@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox