linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
* Re: [PATCH 3/6] locking/rwsem: Rework writer wakeup
       [not found] ` <20230223123319.487908155@infradead.org>
@ 2023-02-24 10:11   ` Hillf Danton
  0 siblings, 0 replies; 2+ messages in thread
From: Hillf Danton @ 2023-02-24 10:11 UTC (permalink / raw)
  To: Peter Zijlstra; +Cc: longman, mingo, will, linux-kernel, linux-mm, boqun.feng

On Thu, 23 Feb 2023 13:26:45 +0100 Peter Zijlstra <peterz@infradead.org>
> +static void rwsem_writer_wake(struct rw_semaphore *sem,
> +			      struct rwsem_waiter *waiter,
> +			      struct wake_q_head *wake_q)
> +{
> +	struct rwsem_waiter *first = rwsem_first_waiter(sem);
> +	long count, new;
> +
> +	lockdep_assert_held(&sem->wait_lock);
> +
> +	count = atomic_long_read(&sem->count);
> +	do {
> +		bool has_handoff = !!(count & RWSEM_FLAG_HANDOFF);
> +
> +		if (has_handoff) {
> +			/*
> +			 * Honor handoff bit and yield only when the first
> +			 * waiter is the one that set it. Otherwisee, we
> +			 * still try to acquire the rwsem.
> +			 */
> +			if (first->handoff_set && (waiter != first))
> +				return;

Given that HANDOFF disables all spinning and stealing in 2/6, what sense
made by still trying to acquire the rwsem? The answer is perhaps it is called
at wake time.
> +		}
> +
> +		new = count;
> +
> +		if (count & RWSEM_LOCK_MASK) {

If it is only called at wake time, the chance for a transient RWSEM_READER_BIAS
to ruin the wakeup is not zero. 

> +			/*
> +			 * A waiter (first or not) can set the handoff bit
> +			 * if it is an RT task or wait in the wait queue
> +			 * for too long.
> +			 */
> +			if (has_handoff || (!rt_task(waiter->task) &&
> +					    !time_after(jiffies, waiter->timeout)))
> +				return;
> +
> +			new |= RWSEM_FLAG_HANDOFF;
> +		} else {
> +			new |= RWSEM_WRITER_LOCKED;
> +			new &= ~RWSEM_FLAG_HANDOFF;
> +
> +			if (list_is_singular(&sem->wait_list))
> +				new &= ~RWSEM_FLAG_WAITERS;
> +		}
> +	} while (!atomic_long_try_cmpxchg_acquire(&sem->count, &count, new));
> +
> +	/*
> +	 * We have either acquired the lock with handoff bit cleared or set
> +	 * the handoff bit. Only the first waiter can have its handoff_set
> +	 * set here to enable optimistic spinning in slowpath loop.
> +	 */
> +	if (new & RWSEM_FLAG_HANDOFF) {
> +		first->handoff_set = true;
> +		lockevent_inc(rwsem_wlock_handoff);
> +		return;
> +	}
> +
> +	/*
> +	 * Have rwsem_writer_wake() fully imply rwsem_del_waiter() on
> +	 * success.
> +	 */
> +	list_del(&waiter->list);
> +	rwsem_set_owner(sem);
> +	rwsem_waiter_wake(waiter, wake_q);
> +}
> +
>  /*
>   * handle the lock release when processes blocked on it that can now run
>   * - if we come here from up_xxxx(), then the RWSEM_FLAG_WAITERS bit must
> @@ -424,23 +518,12 @@ static void rwsem_mark_wake(struct rw_se
>  	 */
>  	waiter = rwsem_first_waiter(sem);
>  
> -	if (waiter->type != RWSEM_WAITING_FOR_WRITE)
> -		goto wake_readers;
> -
> -	if (wake_type == RWSEM_WAKE_ANY) {
> -		/*
> -		 * Mark writer at the front of the queue for wakeup.
> -		 * Until the task is actually later awoken later by
> -		 * the caller, other writers are able to steal it.
> -		 * Readers, on the other hand, will block as they
> -		 * will notice the queued writer.
> -		 */
> -		wake_q_add(wake_q, waiter->task);
> -		lockevent_inc(rwsem_wake_writer);
> +	if (waiter->type == RWSEM_WAITING_FOR_WRITE) {
> +		if (wake_type == RWSEM_WAKE_ANY)
> +			rwsem_writer_wake(sem, waiter, wake_q);
> +		return;


^ permalink raw reply	[flat|nested] 2+ messages in thread

* Re: [PATCH 4/6] locking/rwsem: Split out rwsem_reader_wake()
       [not found] ` <20230223123319.548254615@infradead.org>
@ 2023-02-25 10:15   ` Hillf Danton
  0 siblings, 0 replies; 2+ messages in thread
From: Hillf Danton @ 2023-02-25 10:15 UTC (permalink / raw)
  To: Peter Zijlstra; +Cc: longman, mingo, will, linux-kernel, linux-mm, boqun.feng

On Thu, 23 Feb 2023 13:26:46 +0100 Peter Zijlstra (Intel) <peterz@infradead.org>
> +static void rwsem_mark_wake(struct rw_semaphore *sem,
> +			    enum rwsem_wake_type wake_type,
> +			    struct wake_q_head *wake_q)
> +{
> +	struct rwsem_waiter *waiter;
> +
> +	lockdep_assert_held(&sem->wait_lock);
> +
> +	/*
> +	 * Take a peek at the queue head waiter such that we can determine
> +	 * the wakeup(s) to perform.
> +	 */
> +	waiter = rwsem_first_waiter(sem);
> +
> +	if (waiter->type == RWSEM_WAITING_FOR_WRITE) {
> +		if (wake_type == RWSEM_WAKE_ANY)
> +			rwsem_writer_wake(sem, waiter, wake_q);
> +	} else {
> +		rwsem_reader_wake(sem, wake_type, waiter, wake_q);
> +	}
> +}

Hm... missing wakeup, and feel free to pick up the diff below if it
makes sense to you.

+++ b/kernel/locking/rwsem.c
@@ -661,12 +661,13 @@ static void rwsem_mark_wake(struct rw_se
 	 */
 	waiter = rwsem_first_waiter(sem);
 
-	if (waiter->type == RWSEM_WAITING_FOR_WRITE) {
-		if (wake_type == RWSEM_WAKE_ANY)
+	if (waiter->type == RWSEM_WAITING_FOR_WRITE)
+		if (wake_type == RWSEM_WAKE_ANY) {
 			rwsem_writer_wake(sem, waiter, wake_q);
-	} else {
-		rwsem_reader_wake(sem, wake_type, waiter, wake_q);
-	}
+			return;
+		}
+
+	rwsem_reader_wake(sem, wake_type, waiter, wake_q);
 }
 
 /*


^ permalink raw reply	[flat|nested] 2+ messages in thread

end of thread, other threads:[~2023-02-25 10:15 UTC | newest]

Thread overview: 2+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
     [not found] <20230223122642.491637862@infradead.org>
     [not found] ` <20230223123319.487908155@infradead.org>
2023-02-24 10:11   ` [PATCH 3/6] locking/rwsem: Rework writer wakeup Hillf Danton
     [not found] ` <20230223123319.548254615@infradead.org>
2023-02-25 10:15   ` [PATCH 4/6] locking/rwsem: Split out rwsem_reader_wake() Hillf Danton

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox