From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 16D12C433F5 for ; Wed, 18 May 2022 16:27:30 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id A93828D0001; Wed, 18 May 2022 12:27:29 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id A1D016B0074; Wed, 18 May 2022 12:27:29 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 896118D0001; Wed, 18 May 2022 12:27:29 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0013.hostedemail.com [216.40.44.13]) by kanga.kvack.org (Postfix) with ESMTP id 77B9A6B0073 for ; Wed, 18 May 2022 12:27:29 -0400 (EDT) Received: from smtpin29.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay10.hostedemail.com (Postfix) with ESMTP id 46C806F4 for ; Wed, 18 May 2022 16:27:29 +0000 (UTC) X-FDA: 79479394218.29.D3B72EA Received: from alexa-out-sd-01.qualcomm.com (alexa-out-sd-01.qualcomm.com [199.106.114.38]) by imf13.hostedemail.com (Postfix) with ESMTP id D8ECC200A3 for ; Wed, 18 May 2022 16:27:05 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=quicinc.com; i=@quicinc.com; q=dns/txt; s=qcdkim; t=1652891248; x=1684427248; h=date:from:to:cc:subject:message-id:references: mime-version:in-reply-to; bh=9ct5rUyFTf0yi2HTXDPN5qQLoFF6S7TELmBqsKJ2GXU=; b=D7geF970THsFLBVUdX82kncA9JCypRP7km5TVeEawUB2kVT+GsfEmzGy 7dxZJi0G0ZFdBud8z4m8Pn5jZWJeH9sGzrdU21JI13h0OtjePll+5FzmG qHSWH0gz7XxNU5/3QlaymjkprTOarDJxljOwPubw6B8G3NTFu/Q2MYMCr Q=; Received: from unknown (HELO ironmsg02-sd.qualcomm.com) ([10.53.140.142]) by alexa-out-sd-01.qualcomm.com with ESMTP; 18 May 2022 09:27:27 -0700 X-QCInternal: smtphost Received: from nasanex01c.na.qualcomm.com ([10.47.97.222]) by ironmsg02-sd.qualcomm.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 18 May 2022 09:27:25 -0700 Received: from nalasex01a.na.qualcomm.com (10.47.209.196) by nasanex01c.na.qualcomm.com (10.47.97.222) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.986.22; Wed, 18 May 2022 09:27:26 -0700 Received: from qian (10.80.80.8) by nalasex01a.na.qualcomm.com (10.47.209.196) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.986.22; Wed, 18 May 2022 09:27:25 -0700 Date: Wed, 18 May 2022 12:27:22 -0400 From: Qian Cai To: Mel Gorman CC: Andrew Morton , Nicolas Saenz Julienne , Marcelo Tosatti , Vlastimil Babka , Michal Hocko , LKML , Linux-MM , "Paul E. McKenney" Subject: Re: [PATCH 0/6] Drain remote per-cpu directly v3 Message-ID: References: <20220512085043.5234-1-mgorman@techsingularity.net> <20220517233507.GA423@qian> <20220518125152.GQ3441@techsingularity.net> MIME-Version: 1.0 Content-Type: text/plain; charset="us-ascii" Content-Disposition: inline In-Reply-To: <20220518125152.GQ3441@techsingularity.net> X-Originating-IP: [10.80.80.8] X-ClientProxiedBy: nasanex01b.na.qualcomm.com (10.46.141.250) To nalasex01a.na.qualcomm.com (10.47.209.196) X-Stat-Signature: fidozqp3kicum85nkad6orozydectcax Authentication-Results: imf13.hostedemail.com; dkim=pass header.d=quicinc.com header.s=qcdkim header.b=D7geF970; spf=pass (imf13.hostedemail.com: domain of quic_qiancai@quicinc.com designates 199.106.114.38 as permitted sender) smtp.mailfrom=quic_qiancai@quicinc.com; dmarc=pass (policy=none) header.from=quicinc.com X-Rspam-User: X-Rspamd-Server: rspam01 X-Rspamd-Queue-Id: D8ECC200A3 X-HE-Tag: 1652891225-267472 X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: On Wed, May 18, 2022 at 01:51:52PM +0100, Mel Gorman wrote: > On Tue, May 17, 2022 at 07:35:07PM -0400, Qian Cai wrote: > > On Thu, May 12, 2022 at 09:50:37AM +0100, Mel Gorman wrote: > > > Changelog since v2 > > > o More conversions from page->lru to page->[pcp_list|buddy_list] > > > o Additional test results in changelogs > > > > > > Changelog since v1 > > > o Fix unsafe RT locking scheme > > > o Use spin_trylock on UP PREEMPT_RT > > > > > > This series has the same intent as Nicolas' series "mm/page_alloc: Remote > > > per-cpu lists drain support" -- avoid interference of a high priority > > > task due to a workqueue item draining per-cpu page lists. While many > > > workloads can tolerate a brief interruption, it may be cause a real-time > > > task runnning on a NOHZ_FULL CPU to miss a deadline and at minimum, > > > the draining in non-deterministic. > > > > > > Currently an IRQ-safe local_lock protects the page allocator per-cpu lists. > > > The local_lock on its own prevents migration and the IRQ disabling protects > > > from corruption due to an interrupt arriving while a page allocation is > > > in progress. The locking is inherently unsafe for remote access unless > > > the CPU is hot-removed. > > > > > > This series adjusts the locking. A spinlock is added to struct > > > per_cpu_pages to protect the list contents while local_lock_irq continues > > > to prevent migration and IRQ reentry. This allows a remote CPU to safely > > > drain a remote per-cpu list. > > > > > > This series is a partial series. Follow-on work should allow the > > > local_irq_save to be converted to a local_irq to avoid IRQs being > > > disabled/enabled in most cases. Consequently, there are some TODO comments > > > highlighting the places that would change if local_irq was used. However, > > > there are enough corner cases that it deserves a series on its own > > > separated by one kernel release and the priority right now is to avoid > > > interference of high priority tasks. > > > > Reverting the whole series fixed an issue that offlining a memory > > section blocking for hours on today's linux-next tree. > > > > __wait_rcu_gp > > synchronize_rcu at kernel/rcu/tree.c:3915 > > lru_cache_disable at mm/swap.c:886 > > __alloc_contig_migrate_range at mm/page_alloc.c:9078 > > isolate_single_pageblock at mm/page_isolation.c:405 > > start_isolate_page_range > > offline_pages > > memory_subsys_offline > > device_offline > > online_store > > dev_attr_store > > sysfs_kf_write > > kernfs_fop_write_iter > > new_sync_write > > vfs_write > > ksys_write > > __arm64_sys_write > > invoke_syscall > > el0_svc_common.constprop.0 > > do_el0_svc > > el0_svc > > el0t_64_sync_handler > > el0t_64_sync > > > > For full disclosure, I have also reverted the commit 0d523026abd4 > > ("mm/page_alloc: fix tracepoint mm_page_alloc_zone_locked()"), so the > > series can be reverted cleanly. But, I can't see how the commit > > 0d523026abd4 could cause this issue at all. > > This is halting in __lru_add_drain_all where it calls synchronize_rcu > before the drain even happens. It's also an LRU drain and not PCP which > is what the series affects and the allocator doesn't use rcu. In a KVM > machine, I can do > > $ for BANK in `(for i in {1..20}; do echo $((RANDOM%416)); done) | sort -n | uniq`; do BEFORE=`cat /sys/devices/system/memory/memory$BANK/online`; echo 0 > /sys/devices/system/memory/memory$BANK/online; AFTER=`cat /sys/devices/system/memory/memory$BANK/online`; printf "%4d %d -> %d\n" $BANK $BEFORE $AFTER; done > 3 1 -> 0 > 57 1 -> 0 > 74 1 -> 0 > 93 1 -> 0 > 101 1 -> 0 > 128 1 -> 0 > 133 1 -> 0 > 199 1 -> 0 > 223 1 -> 0 > 225 1 -> 0 > 229 1 -> 0 > 243 1 -> 0 > 263 1 -> 0 > 300 1 -> 0 > 309 1 -> 0 > 329 1 -> 0 > 355 1 -> 0 > 365 1 -> 0 > 372 1 -> 0 > 383 1 -> 0 > > It offlines 20 sections although after several attempts free -m starts > reporting negative used memory so there is a bug of some description. > How are you testing this exactly? Is it every time or intermittent? Are > you confident that reverting the series makes the problem go away? Cc'ing Paul. Either reverting this series or Paul's 3 patches below from today's linux-next tree fixed the issue. ca52639daa5b rcu-tasks: Drive synchronous grace periods from calling task 89ad98e93ce8 rcu-tasks: Move synchronize_rcu_tasks_generic() down 0d90e7225fb1 rcu-tasks: Split rcu_tasks_one_gp() from rcu_tasks_kthread() It was reproduced by running this script below on an arm64 server. I can reproduce it every time within 5 attempts. I noticed that when it happens, we have a few rcu kthreads all are stuck in this line, rcuwait_wait_event(&rtp->cbs_wait, (needgpcb = rcu_tasks_need_gpcb(rtp)), TASK_IDLE); rcu_tasks_kthread rcu_tasks_rude_kthread [rcu_tasks_trace_kthread #!/usr/bin/env python3 # SPDX-License-Identifier: GPL-2.0 import os import re import subprocess def mem_iter(): base_dir = '/sys/devices/system/memory/' for curr_dir in os.listdir(base_dir): if re.match(r'memory\d+', curr_dir): yield base_dir + curr_dir if __name__ == '__main__': print('- Try to remove each memory section and then add it back.') for mem_dir in mem_iter(): status = f'{mem_dir}/online' if open(status).read().rstrip() == '1': # This could expectedly fail due to many reasons. section = os.path.basename(mem_dir) print(f'- Try to remove {section}.') proc = subprocess.run([f'echo 0 | sudo tee {status}'], shell=True) if proc.returncode == 0: print(f'- Try to add {section}.') subprocess.check_call([f'echo 1 | sudo tee {status}'], shell=True)