* Re: [PATCH 2/3] rcu: Equip sleepable RCU with lockdep dependency graph checks [not found] ` <20230113065955.815667-3-boqun.feng@gmail.com> @ 2023-01-13 13:03 ` Hillf Danton 2023-01-13 17:58 ` Boqun Feng 0 siblings, 1 reply; 8+ messages in thread From: Hillf Danton @ 2023-01-13 13:03 UTC (permalink / raw) To: Boqun Feng Cc: linux-kernel, linux-mm, Peter Zijlstra, Paul E. McKenney, Paolo Bonzini, Joel Fernandes On 12 Jan 2023 22:59:54 -0800 Boqun Feng <boqun.feng@gmail.com> > --- a/kernel/rcu/srcutree.c > +++ b/kernel/rcu/srcutree.c > @@ -1267,6 +1267,8 @@ static void __synchronize_srcu(struct srcu_struct *ssp, bool do_norm) > { > struct rcu_synchronize rcu; > > + srcu_lock_sync(&ssp->dep_map); > + > RCU_LOCKDEP_WARN(lockdep_is_held(ssp) || > lock_is_held(&rcu_bh_lock_map) || > lock_is_held(&rcu_lock_map) || > -- > 2.38.1 The following deadlock is able to escape srcu_lock_sync() because the __lock_release folded in sync leaves one lock on the sync side. cpu9 cpu0 --- --- lock A srcu_lock_acquire(&ssp->dep_map); srcu_lock_sync(&ssp->dep_map); lock A ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH 2/3] rcu: Equip sleepable RCU with lockdep dependency graph checks 2023-01-13 13:03 ` [PATCH 2/3] rcu: Equip sleepable RCU with lockdep dependency graph checks Hillf Danton @ 2023-01-13 17:58 ` Boqun Feng 2023-01-13 23:58 ` Hillf Danton 0 siblings, 1 reply; 8+ messages in thread From: Boqun Feng @ 2023-01-13 17:58 UTC (permalink / raw) To: Hillf Danton Cc: linux-kernel, linux-mm, Peter Zijlstra, Paul E. McKenney, Paolo Bonzini, Joel Fernandes On Fri, Jan 13, 2023 at 09:03:30PM +0800, Hillf Danton wrote: > On 12 Jan 2023 22:59:54 -0800 Boqun Feng <boqun.feng@gmail.com> > > --- a/kernel/rcu/srcutree.c > > +++ b/kernel/rcu/srcutree.c > > @@ -1267,6 +1267,8 @@ static void __synchronize_srcu(struct srcu_struct *ssp, bool do_norm) > > { > > struct rcu_synchronize rcu; > > > > + srcu_lock_sync(&ssp->dep_map); > > + > > RCU_LOCKDEP_WARN(lockdep_is_held(ssp) || > > lock_is_held(&rcu_bh_lock_map) || > > lock_is_held(&rcu_lock_map) || > > -- > > 2.38.1 > > The following deadlock is able to escape srcu_lock_sync() because the > __lock_release folded in sync leaves one lock on the sync side. > > cpu9 cpu0 > --- --- > lock A srcu_lock_acquire(&ssp->dep_map); > srcu_lock_sync(&ssp->dep_map); > lock A But isn't it just the srcu_mutex_ABBA test case in patch #3, and my run of lockdep selftest shows we can catch it. Anything subtle I'm missing? Regards, Boqun ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH 2/3] rcu: Equip sleepable RCU with lockdep dependency graph checks 2023-01-13 17:58 ` Boqun Feng @ 2023-01-13 23:58 ` Hillf Danton 2023-01-14 0:17 ` Boqun Feng 0 siblings, 1 reply; 8+ messages in thread From: Hillf Danton @ 2023-01-13 23:58 UTC (permalink / raw) To: Boqun Feng Cc: linux-kernel, linux-mm, Peter Zijlstra, Paul E. McKenney, Paolo Bonzini, Joel Fernandes On 13 Jan 2023 09:58:10 -0800 Boqun Feng <boqun.feng@gmail.com> > On Fri, Jan 13, 2023 at 09:03:30PM +0800, Hillf Danton wrote: > > On 12 Jan 2023 22:59:54 -0800 Boqun Feng <boqun.feng@gmail.com> > > > --- a/kernel/rcu/srcutree.c > > > +++ b/kernel/rcu/srcutree.c > > > @@ -1267,6 +1267,8 @@ static void __synchronize_srcu(struct srcu_struct *ssp, bool do_norm) > > > { > > > struct rcu_synchronize rcu; > > > > > > + srcu_lock_sync(&ssp->dep_map); > > > + > > > RCU_LOCKDEP_WARN(lockdep_is_held(ssp) || > > > lock_is_held(&rcu_bh_lock_map) || > > > lock_is_held(&rcu_lock_map) || > > > -- > > > 2.38.1 > > > > The following deadlock is able to escape srcu_lock_sync() because the > > __lock_release folded in sync leaves one lock on the sync side. > > > > cpu9 cpu0 > > --- --- > > lock A srcu_lock_acquire(&ssp->dep_map); > > srcu_lock_sync(&ssp->dep_map); > > lock A > > But isn't it just the srcu_mutex_ABBA test case in patch #3, and my run > of lockdep selftest shows we can catch it. Anything subtle I'm missing? I am leaning to not call it ABBA deadlock, because B is unlocked. task X task Y --- --- lock A lock B lock B unlock B wait_for_completion E lock A complete E And no deadlock should be detected/caught after B goes home. ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH 2/3] rcu: Equip sleepable RCU with lockdep dependency graph checks 2023-01-13 23:58 ` Hillf Danton @ 2023-01-14 0:17 ` Boqun Feng 2023-01-14 7:18 ` Hillf Danton 0 siblings, 1 reply; 8+ messages in thread From: Boqun Feng @ 2023-01-14 0:17 UTC (permalink / raw) To: Hillf Danton Cc: linux-kernel, linux-mm, Peter Zijlstra, Paul E. McKenney, Paolo Bonzini, Joel Fernandes On Sat, Jan 14, 2023 at 07:58:09AM +0800, Hillf Danton wrote: > On 13 Jan 2023 09:58:10 -0800 Boqun Feng <boqun.feng@gmail.com> > > On Fri, Jan 13, 2023 at 09:03:30PM +0800, Hillf Danton wrote: > > > On 12 Jan 2023 22:59:54 -0800 Boqun Feng <boqun.feng@gmail.com> > > > > --- a/kernel/rcu/srcutree.c > > > > +++ b/kernel/rcu/srcutree.c > > > > @@ -1267,6 +1267,8 @@ static void __synchronize_srcu(struct srcu_struct *ssp, bool do_norm) > > > > { > > > > struct rcu_synchronize rcu; > > > > > > > > + srcu_lock_sync(&ssp->dep_map); > > > > + > > > > RCU_LOCKDEP_WARN(lockdep_is_held(ssp) || > > > > lock_is_held(&rcu_bh_lock_map) || > > > > lock_is_held(&rcu_lock_map) || > > > > -- > > > > 2.38.1 > > > > > > The following deadlock is able to escape srcu_lock_sync() because the > > > __lock_release folded in sync leaves one lock on the sync side. > > > > > > cpu9 cpu0 > > > --- --- > > > lock A srcu_lock_acquire(&ssp->dep_map); > > > srcu_lock_sync(&ssp->dep_map); > > > lock A > > > > But isn't it just the srcu_mutex_ABBA test case in patch #3, and my run > > of lockdep selftest shows we can catch it. Anything subtle I'm missing? > > I am leaning to not call it ABBA deadlock, because B is unlocked. > > task X task Y > --- --- > lock A > lock B > lock B > unlock B > wait_for_completion E > > lock A > complete E > > And no deadlock should be detected/caught after B goes home. Your example makes me more confused.. given the case: task X task Y --- --- mutex_lock(A); srcu_read_lock(B); synchronze_srcu(B); mutex_lock(A); isn't it a deadlock? If your example, A, B or E which one is srcu? Regards, Boqun ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH 2/3] rcu: Equip sleepable RCU with lockdep dependency graph checks 2023-01-14 0:17 ` Boqun Feng @ 2023-01-14 7:18 ` Hillf Danton 2023-01-14 7:32 ` Boqun Feng 0 siblings, 1 reply; 8+ messages in thread From: Hillf Danton @ 2023-01-14 7:18 UTC (permalink / raw) To: Boqun Feng Cc: linux-kernel, linux-mm, Peter Zijlstra, Paul E. McKenney, Paolo Bonzini, Joel Fernandes On Fri, 13 Jan 2023 16:17:59 -0800 Boqun Feng <boqun.feng@gmail.com> > On Sat, Jan 14, 2023 at 07:58:09AM +0800, Hillf Danton wrote: > > On 13 Jan 2023 09:58:10 -0800 Boqun Feng <boqun.feng@gmail.com> > > > On Fri, Jan 13, 2023 at 09:03:30PM +0800, Hillf Danton wrote: > > > > On 12 Jan 2023 22:59:54 -0800 Boqun Feng <boqun.feng@gmail.com> > > > > > --- a/kernel/rcu/srcutree.c > > > > > +++ b/kernel/rcu/srcutree.c > > > > > @@ -1267,6 +1267,8 @@ static void __synchronize_srcu(struct srcu_struct *ssp, bool do_norm) > > > > > { > > > > > struct rcu_synchronize rcu; > > > > > > > > > > + srcu_lock_sync(&ssp->dep_map); > > > > > + > > > > > RCU_LOCKDEP_WARN(lockdep_is_held(ssp) || > > > > > lock_is_held(&rcu_bh_lock_map) || > > > > > lock_is_held(&rcu_lock_map) || > > > > > -- > > > > > 2.38.1 > > > > > > > > The following deadlock is able to escape srcu_lock_sync() because the > > > > __lock_release folded in sync leaves one lock on the sync side. > > > > > > > > cpu9 cpu0 > > > > --- --- > > > > lock A srcu_lock_acquire(&ssp->dep_map); > > > > srcu_lock_sync(&ssp->dep_map); > > > > lock A > > > > > > But isn't it just the srcu_mutex_ABBA test case in patch #3, and my run > > > of lockdep selftest shows we can catch it. Anything subtle I'm missing? > > > > I am leaning to not call it ABBA deadlock, because B is unlocked. > > > > task X task Y > > --- --- > > lock A > > lock B > > lock B > > unlock B > > wait_for_completion E > > > > lock A > > complete E > > > > And no deadlock should be detected/caught after B goes home. > > Your example makes me more confused.. given the case: > > task X task Y > --- --- > mutex_lock(A); > srcu_read_lock(B); > synchronze_srcu(B); > mutex_lock(A); > > isn't it a deadlock? Yes and nope, see below. > If your example, A, B or E which one is srcu? A and B are mutex, and E is completion in my example to show the failure of catching deadlock in case of non-fake lock. Now see srcu after your change. task X task Y --- --- mutex_lock(A); srcu_read_lock(B); srcu_lock_acquire(&B->dep_map); a) lock_map_acquire_read(&B->dep_map); synchronze_srcu(B); __synchronize_srcu(B); srcu_lock_sync(&B->dep_map); lock_map_sync(&B->dep_map); lock_sync(&B->dep_map); __lock_acquire(&B->dep_map); b) lock_map_acquire_read(&B->dep_map); __lock_release(&B->dep_map); c) lock_map_acquire_read(&B->dep_map); mutex_lock(A); No deadlock could be detected if taskY takes mutexA after taskX releases B, and how taskY acquires B does not matter as per the a), b) and c) modes in the above chart, again because releasing lock can break deadlock in general. ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH 2/3] rcu: Equip sleepable RCU with lockdep dependency graph checks 2023-01-14 7:18 ` Hillf Danton @ 2023-01-14 7:32 ` Boqun Feng 2023-01-14 10:26 ` Hillf Danton 0 siblings, 1 reply; 8+ messages in thread From: Boqun Feng @ 2023-01-14 7:32 UTC (permalink / raw) To: Hillf Danton Cc: linux-kernel, linux-mm, Peter Zijlstra, Paul E. McKenney, Paolo Bonzini, Joel Fernandes On Sat, Jan 14, 2023 at 03:18:32PM +0800, Hillf Danton wrote: > On Fri, 13 Jan 2023 16:17:59 -0800 Boqun Feng <boqun.feng@gmail.com> > > On Sat, Jan 14, 2023 at 07:58:09AM +0800, Hillf Danton wrote: > > > On 13 Jan 2023 09:58:10 -0800 Boqun Feng <boqun.feng@gmail.com> > > > > On Fri, Jan 13, 2023 at 09:03:30PM +0800, Hillf Danton wrote: > > > > > On 12 Jan 2023 22:59:54 -0800 Boqun Feng <boqun.feng@gmail.com> > > > > > > --- a/kernel/rcu/srcutree.c > > > > > > +++ b/kernel/rcu/srcutree.c > > > > > > @@ -1267,6 +1267,8 @@ static void __synchronize_srcu(struct srcu_struct *ssp, bool do_norm) > > > > > > { > > > > > > struct rcu_synchronize rcu; > > > > > > > > > > > > + srcu_lock_sync(&ssp->dep_map); > > > > > > + > > > > > > RCU_LOCKDEP_WARN(lockdep_is_held(ssp) || > > > > > > lock_is_held(&rcu_bh_lock_map) || > > > > > > lock_is_held(&rcu_lock_map) || > > > > > > -- > > > > > > 2.38.1 > > > > > > > > > > The following deadlock is able to escape srcu_lock_sync() because the > > > > > __lock_release folded in sync leaves one lock on the sync side. > > > > > > > > > > cpu9 cpu0 > > > > > --- --- > > > > > lock A srcu_lock_acquire(&ssp->dep_map); > > > > > srcu_lock_sync(&ssp->dep_map); > > > > > lock A > > > > > > > > But isn't it just the srcu_mutex_ABBA test case in patch #3, and my run > > > > of lockdep selftest shows we can catch it. Anything subtle I'm missing? > > > > > > I am leaning to not call it ABBA deadlock, because B is unlocked. > > > > > > task X task Y > > > --- --- > > > lock A > > > lock B > > > lock B > > > unlock B > > > wait_for_completion E > > > > > > lock A > > > complete E > > > > > > And no deadlock should be detected/caught after B goes home. > > > > Your example makes me more confused.. given the case: > > > > task X task Y > > --- --- > > mutex_lock(A); > > srcu_read_lock(B); > > synchronze_srcu(B); > > mutex_lock(A); > > > > isn't it a deadlock? > > Yes and nope, see below. > > > If your example, A, B or E which one is srcu? > > A and B are mutex, and E is completion in my example to show the failure > of catching deadlock in case of non-fake lock. Now see srcu after your change. > > task X task Y > --- --- > mutex_lock(A); > srcu_read_lock(B); > srcu_lock_acquire(&B->dep_map); > a) lock_map_acquire_read(&B->dep_map); > synchronze_srcu(B); > __synchronize_srcu(B); > srcu_lock_sync(&B->dep_map); > lock_map_sync(&B->dep_map); > lock_sync(&B->dep_map); > __lock_acquire(&B->dep_map); At this time, lockdep add dependency A -> B in the dependency graph. > b) lock_map_acquire_read(&B->dep_map); > __lock_release(&B->dep_map); > c) lock_map_acquire_read(&B->dep_map); > mutex_lock(A); and here, lockdep will try to add dependency B -> A into the dependency graph, and find that A -> B -> A will form a circle (with strong dependency), therefore lockdep knows it's a deadlock. > > No deadlock could be detected if taskY takes mutexA after taskX releases B, The timing that taskX releases B doesn't master, since lockdep uses graph to detect deadlocks rather than after-fact detection. > and how taskY acquires B does not matter as per the a), b) and c) modes in > the above chart, again because releasing lock can break deadlock in general. I have test cases showing the above deadlock can be detected, so if you think there is a deadlock that may dodge from my change, feel free to add a test case in lib/locking-selftest.c ;-) Regards, Boqun ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH 2/3] rcu: Equip sleepable RCU with lockdep dependency graph checks 2023-01-14 7:32 ` Boqun Feng @ 2023-01-14 10:26 ` Hillf Danton 2023-01-15 0:18 ` Boqun Feng 0 siblings, 1 reply; 8+ messages in thread From: Hillf Danton @ 2023-01-14 10:26 UTC (permalink / raw) To: Boqun Feng Cc: linux-kernel, linux-mm, Peter Zijlstra, Paul E. McKenney, Paolo Bonzini, Joel Fernandes On Fri, 13 Jan 2023 23:32:01 -0800 Boqun Feng <boqun.feng@gmail.com> > On Sat, Jan 14, 2023 at 03:18:32PM +0800, Hillf Danton wrote: > > > > task X task Y > > --- --- > > mutex_lock(A); > > srcu_read_lock(B); > > srcu_lock_acquire(&B->dep_map); > > a) lock_map_acquire_read(&B->dep_map); > > synchronze_srcu(B); > > __synchronize_srcu(B); > > srcu_lock_sync(&B->dep_map); > > lock_map_sync(&B->dep_map); > > lock_sync(&B->dep_map); > > __lock_acquire(&B->dep_map); > > At this time, lockdep add dependency A -> B in the dependency graph. > > > b) lock_map_acquire_read(&B->dep_map); > > __lock_release(&B->dep_map); > > c) lock_map_acquire_read(&B->dep_map); > > mutex_lock(A); > > and here, lockdep will try to add dependency B -> A into the dependency > graph, and find that A -> B -> A will form a circle (with strong > dependency), therefore lockdep knows it's a deadlock. Is the strong dependency applying to mode c)? If yes then deadlock should be also detected in the following locking pattern that has no deadlock. cpu0 cpu1 --- --- mutex_lock A mutex_lock B mutex_unlock B mutex_lock B mutex_lock A > > > > > No deadlock could be detected if taskY takes mutexA after taskX releases B, > > The timing that taskX releases B doesn't master, since lockdep uses > graph to detect deadlocks rather than after-fact detection. > > > and how taskY acquires B does not matter as per the a), b) and c) modes in > > the above chart, again because releasing lock can break deadlock in general. > > I have test cases showing the above deadlock can be detected, so if you > think there is a deadlock that may dodge from my change, feel free to > add a test case in lib/locking-selftest.c ;-) > > Regards, > Boqun ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH 2/3] rcu: Equip sleepable RCU with lockdep dependency graph checks 2023-01-14 10:26 ` Hillf Danton @ 2023-01-15 0:18 ` Boqun Feng 0 siblings, 0 replies; 8+ messages in thread From: Boqun Feng @ 2023-01-15 0:18 UTC (permalink / raw) To: Hillf Danton Cc: linux-kernel, linux-mm, Peter Zijlstra, Paul E. McKenney, Paolo Bonzini, Joel Fernandes On Sat, Jan 14, 2023 at 06:26:59PM +0800, Hillf Danton wrote: > On Fri, 13 Jan 2023 23:32:01 -0800 Boqun Feng <boqun.feng@gmail.com> > > On Sat, Jan 14, 2023 at 03:18:32PM +0800, Hillf Danton wrote: > > > > > > task X task Y > > > --- --- > > > mutex_lock(A); > > > srcu_read_lock(B); > > > srcu_lock_acquire(&B->dep_map); > > > a) lock_map_acquire_read(&B->dep_map); > > > synchronze_srcu(B); > > > __synchronize_srcu(B); > > > srcu_lock_sync(&B->dep_map); > > > lock_map_sync(&B->dep_map); > > > lock_sync(&B->dep_map); > > > __lock_acquire(&B->dep_map); > > > > At this time, lockdep add dependency A -> B in the dependency graph. > > > > > b) lock_map_acquire_read(&B->dep_map); > > > __lock_release(&B->dep_map); > > > c) lock_map_acquire_read(&B->dep_map); > > > mutex_lock(A); > > > > and here, lockdep will try to add dependency B -> A into the dependency > > graph, and find that A -> B -> A will form a circle (with strong > > dependency), therefore lockdep knows it's a deadlock. > > Is the strong dependency applying to mode c)? > If yes then deadlock should be also detected in the following locking > pattern that has no deadlock. > > cpu0 cpu1 > --- --- > mutex_lock A > mutex_lock B > mutex_unlock B > mutex_lock B > mutex_lock A Well, of course, this is how lockdep works. Lockdep detects the *potential* deadlocks rather than detects the deadlocks when they really happen. Otherwise lockdep is useless. The execution in your example shows the potential deadlocks, i.e. one task acquires A and then acquires B, the other task acquires B and then acquires A. Potential deadlocks mean given a correct timing, a deadlock may happen. Regards, Boqun > > > > > > > > No deadlock could be detected if taskY takes mutexA after taskX releases B, > > > > The timing that taskX releases B doesn't master, since lockdep uses > > graph to detect deadlocks rather than after-fact detection. > > > > > and how taskY acquires B does not matter as per the a), b) and c) modes in > > > the above chart, again because releasing lock can break deadlock in general. > > > > I have test cases showing the above deadlock can be detected, so if you > > think there is a deadlock that may dodge from my change, feel free to > > add a test case in lib/locking-selftest.c ;-) > > > > Regards, > > Boqun ^ permalink raw reply [flat|nested] 8+ messages in thread
end of thread, other threads:[~2023-01-15 0:19 UTC | newest]
Thread overview: 8+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
[not found] <20230113065955.815667-1-boqun.feng@gmail.com>
[not found] ` <20230113065955.815667-3-boqun.feng@gmail.com>
2023-01-13 13:03 ` [PATCH 2/3] rcu: Equip sleepable RCU with lockdep dependency graph checks Hillf Danton
2023-01-13 17:58 ` Boqun Feng
2023-01-13 23:58 ` Hillf Danton
2023-01-14 0:17 ` Boqun Feng
2023-01-14 7:18 ` Hillf Danton
2023-01-14 7:32 ` Boqun Feng
2023-01-14 10:26 ` Hillf Danton
2023-01-15 0:18 ` Boqun Feng
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox