linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
* Expensive memory.stat + cpu.stat reads
@ 2023-06-30 23:22 Ivan Babrou
  2023-07-06  6:20 ` Shakeel Butt
  2023-07-10 14:44 ` Michal Koutný
  0 siblings, 2 replies; 21+ messages in thread
From: Ivan Babrou @ 2023-06-30 23:22 UTC (permalink / raw)
  To: cgroups
  Cc: Linux MM, kernel-team, Johannes Weiner, Michal Hocko,
	Roman Gushchin, Shakeel Butt, Muchun Song, Andrew Morton,
	linux-kernel

Hello,

We're seeing CPU load issues with cgroup stats retrieval. I made a
public gist with all the details, including the repro code (which
unfortunately requires heavily loaded hardware) and some flamegraphs:

* https://gist.github.com/bobrik/5ba58fb75a48620a1965026ad30a0a13

I'll repeat the gist of that gist here. Our repro has the following
output after a warm-up run:

completed:  5.17s [manual / mem-stat + cpu-stat]
completed:  5.59s [manual / cpu-stat + mem-stat]
completed:  0.52s [manual / mem-stat]
completed:  0.04s [manual / cpu-stat]

The first two lines do effectively the following:

for _ in $(seq 1 1000); do cat /sys/fs/cgroup/system.slice/memory.stat
/sys/fs/cgroup/system.slice/cpu.stat > /dev/null

The latter two are the same thing, but via two loops:

for _ in $(seq 1 1000); do cat /sys/fs/cgroup/system.slice/cpu.stat >
/dev/null; done
for _ in $(seq 1 1000); do cat /sys/fs/cgroup/system.slice/memory.stat
> /dev/null; done

As you might've noticed from the output, splitting the loop into two
makes the code run 10x faster. This isn't great, because most
monitoring software likes to get all stats for one service before
reading the stats for the next one, which maps to the slow and
expensive way of doing this.

We're running Linux v6.1 (the output is from v6.1.25) with no patches
that touch the cgroup or mm subsystems, so you can assume vanilla
kernel.

From the flamegraph it just looks like rstat flushing takes longer. I
used the following flags on an AMD EPYC 7642 system (our usual pick
cpu-clock was blaming spinlock irqrestore, which was questionable):

perf -e cycles -g --call-graph fp -F 999 -- /tmp/repro

Naturally, there are two questions that arise:

* Is this expected (I guess not, but good to be sure)?
* What can we do to make this better?

I am happy to try out patches or to do some tracing to help understand
this better.


^ permalink raw reply	[flat|nested] 21+ messages in thread

end of thread, other threads:[~2023-08-15  0:31 UTC | newest]

Thread overview: 21+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2023-06-30 23:22 Expensive memory.stat + cpu.stat reads Ivan Babrou
2023-07-06  6:20 ` Shakeel Butt
2023-07-10 23:21   ` Ivan Babrou
2023-07-11  0:44     ` Waiman Long
2023-07-13 23:25       ` Ivan Babrou
2023-07-14 17:23         ` Waiman Long
2023-07-15  0:00           ` Ivan Babrou
2023-07-15  0:30         ` Ivan Babrou
2023-08-11 22:03           ` Ivan Babrou
2023-08-11 22:27             ` Waiman Long
2023-08-11 22:35               ` Ivan Babrou
2023-08-12  2:33                 ` Shakeel Butt
2023-08-14 17:56                   ` Ivan Babrou
2023-08-11 23:43             ` Yosry Ahmed
2023-08-12  0:01               ` Yosry Ahmed
2023-08-15  0:18                 ` Tejun Heo
2023-08-15  0:30                   ` Ivan Babrou
2023-08-15  0:31                     ` Yosry Ahmed
2023-07-15  0:14     ` Ivan Babrou
2023-07-10 14:44 ` Michal Koutný
2023-07-10 23:23   ` Ivan Babrou

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox