Re: [PATCH] rwsem: fix missed wakeup due to reordering of load
On Wed, 20 Sep 2017, Andrea Parri wrote: Instead, how about just removing the release from atomic_long_sub_return_release() such that the osq load is not hoisted over the atomic compound (along with Peter's comment): This solution will actually enforce a stronger (full) ordering w.r.t. the solution described by Prateek and Peter. Also, it will "trade" two lwsync for two sync (powerpc), one dmb.ld for one dmb (arm64). What are the reasons you would prefer this? It was mainly to maintain consistency about dealing with sem->count, but sure I won't argue with the above.
Re: [PATCH] rwsem: fix missed wakeup due to reordering of load
On 09/07/2017 08:00 PM, Prateek Sood wrote: > If a spinner is present, there is a chance that the load of > rwsem_has_spinner() in rwsem_wake() can be reordered with > respect to decrement of rwsem count in __up_write() leading > to wakeup being missed. > > spinning writer up_write caller > --- --- > [S] osq_unlock() [L] osq > spin_lock(wait_lock) > sem->count=0x0001 > +0x > count=sem->count > MB >sem->count=0xFFFE0001 > -0x0001 >spin_trylock(wait_lock) >return > rwsem_try_write_lock(count) > spin_unlock(wait_lock) > schedule() > > Reordering of atomic_long_sub_return_release() in __up_write() > and rwsem_has_spinner() in rwsem_wake() can cause missing of > wakeup in up_write() context. In spinning writer, sem->count > and local variable count is 0XFFFE0001. It would result > in rwsem_try_write_lock() failing to acquire rwsem and spinning > writer going to sleep in rwsem_down_write_failed(). > > The smp_rmb() will make sure that the spinner state is > consulted after sem->count is updated in up_write context. > > Signed-off-by: Prateek Sood > --- > kernel/locking/rwsem-xadd.c | 27 +++ > 1 file changed, 27 insertions(+) > > diff --git a/kernel/locking/rwsem-xadd.c b/kernel/locking/rwsem-xadd.c > index 02f6606..1fefe6d 100644 > --- a/kernel/locking/rwsem-xadd.c > +++ b/kernel/locking/rwsem-xadd.c > @@ -613,6 +613,33 @@ struct rw_semaphore *rwsem_wake(struct rw_semaphore *sem) > DEFINE_WAKE_Q(wake_q); > > /* > + * __rwsem_down_write_failed_common(sem) > + * rwsem_optimistic_spin(sem) > + * osq_unlock(sem->osq) > + * ... > + * atomic_long_add_return(&sem->count) > + * > + * - VS - > + * > + * __up_write() > + *if (atomic_long_sub_return_release(&sem->count) < 0) > + * rwsem_wake(sem) > + *osq_is_locked(&sem->osq) > + * > + * And __up_write() must observe !osq_is_locked() when it observes the > + * atomic_long_add_return() in order to not miss a wakeup. > + * > + * This boils down to: > + * > + * [S.rel] X = 1[RmW] r0 = (Y += 0) > + * MB RMB > + * [RmW] Y += 1 [L] r1 = X > + * > + * exists (r0=1 /\ r1=0) > + */ > + smp_rmb(); > + > + /* >* If a spinner is present, it is not necessary to do the wakeup. >* Try to do wakeup only if the trylock succeeds to minimize >* spinlock contention which may introduce too much delay in the > Hi Folks, Do you have any more suggestion/feedback on this patch. Regards Prateek -- Qualcomm India Private Limited, on behalf of Qualcomm Innovation Center, Inc., is a member of Code Aurora Forum, a Linux Foundation Collaborative Project
Re: [PATCH] rwsem: fix missed wakeup due to reordering of load
On Wed, Sep 20, 2017 at 07:52:54AM -0700, Davidlohr Bueso wrote: > On Thu, 07 Sep 2017, Prateek Sood wrote: > > /* > >+* __rwsem_down_write_failed_common(sem) > >+* rwsem_optimistic_spin(sem) > >+* osq_unlock(sem->osq) > >+* ... > >+* atomic_long_add_return(&sem->count) > >+* > >+* - VS - > >+* > >+* __up_write() > >+*if (atomic_long_sub_return_release(&sem->count) < 0) > >+* rwsem_wake(sem) > >+*osq_is_locked(&sem->osq) > >+* > >+* And __up_write() must observe !osq_is_locked() when it observes the > >+* atomic_long_add_return() in order to not miss a wakeup. > >+* > >+* This boils down to: > >+* > >+* [S.rel] X = 1[RmW] r0 = (Y += 0) > >+* MB RMB > >+* [RmW] Y += 1 [L] r1 = X > >+* > >+* exists (r0=1 /\ r1=0) > >+*/ > >+smp_rmb(); > > Instead, how about just removing the release from > atomic_long_sub_return_release() > such that the osq load is not hoisted over the atomic compound (along with > Peter's > comment): This solution will actually enforce a stronger (full) ordering w.r.t. the solution described by Prateek and Peter. Also, it will "trade" two lwsync for two sync (powerpc), one dmb.ld for one dmb (arm64). What are the reasons you would prefer this? Andrea > > diff --git a/include/asm-generic/rwsem.h b/include/asm-generic/rwsem.h > index 6c6a2141f271..487ce31078ff 100644 > --- a/include/asm-generic/rwsem.h > +++ b/include/asm-generic/rwsem.h > @@ -101,7 +101,7 @@ static inline void __up_read(struct rw_semaphore *sem) > */ > static inline void __up_write(struct rw_semaphore *sem) > { > - if (unlikely(atomic_long_sub_return_release(RWSEM_ACTIVE_WRITE_BIAS, > + if (unlikely(atomic_long_sub_return(RWSEM_ACTIVE_WRITE_BIAS, > &sem->count) < 0)) > rwsem_wake(sem); > }
Re: [PATCH] rwsem: fix missed wakeup due to reordering of load
On Thu, 07 Sep 2017, Prateek Sood wrote: /* + * __rwsem_down_write_failed_common(sem) + * rwsem_optimistic_spin(sem) + * osq_unlock(sem->osq) + * ... + * atomic_long_add_return(&sem->count) + * + * - VS - + * + * __up_write() + *if (atomic_long_sub_return_release(&sem->count) < 0) + * rwsem_wake(sem) + *osq_is_locked(&sem->osq) + * + * And __up_write() must observe !osq_is_locked() when it observes the + * atomic_long_add_return() in order to not miss a wakeup. + * + * This boils down to: + * + * [S.rel] X = 1[RmW] r0 = (Y += 0) + * MB RMB + * [RmW] Y += 1 [L] r1 = X + * + * exists (r0=1 /\ r1=0) + */ + smp_rmb(); Instead, how about just removing the release from atomic_long_sub_return_release() such that the osq load is not hoisted over the atomic compound (along with Peter's comment): diff --git a/include/asm-generic/rwsem.h b/include/asm-generic/rwsem.h index 6c6a2141f271..487ce31078ff 100644 --- a/include/asm-generic/rwsem.h +++ b/include/asm-generic/rwsem.h @@ -101,7 +101,7 @@ static inline void __up_read(struct rw_semaphore *sem) */ static inline void __up_write(struct rw_semaphore *sem) { - if (unlikely(atomic_long_sub_return_release(RWSEM_ACTIVE_WRITE_BIAS, + if (unlikely(atomic_long_sub_return(RWSEM_ACTIVE_WRITE_BIAS, &sem->count) < 0)) rwsem_wake(sem); }
Re: [PATCH] rwsem: fix missed wakeup due to reordering of load
On Thu, Sep 07, 2017 at 08:00:58PM +0530, Prateek Sood wrote: > If a spinner is present, there is a chance that the load of > rwsem_has_spinner() in rwsem_wake() can be reordered with > respect to decrement of rwsem count in __up_write() leading > to wakeup being missed. > > spinning writer up_write caller > --- --- > [S] osq_unlock() [L] osq > spin_lock(wait_lock) > sem->count=0x0001 > +0x > count=sem->count > MB >sem->count=0xFFFE0001 > -0x0001 >spin_trylock(wait_lock) >return > rwsem_try_write_lock(count) > spin_unlock(wait_lock) > schedule() > > Reordering of atomic_long_sub_return_release() in __up_write() > and rwsem_has_spinner() in rwsem_wake() can cause missing of > wakeup in up_write() context. In spinning writer, sem->count > and local variable count is 0XFFFE0001. It would result > in rwsem_try_write_lock() failing to acquire rwsem and spinning > writer going to sleep in rwsem_down_write_failed(). > > The smp_rmb() will make sure that the spinner state is > consulted after sem->count is updated in up_write context. > > Signed-off-by: Prateek Sood Reviewed-by: Andrea Parri I understand that the merge window and LPC made this stalls for a while... what am I missing? are there other changes that need to be considered for this patch? Andrea > --- > kernel/locking/rwsem-xadd.c | 27 +++ > 1 file changed, 27 insertions(+) > > diff --git a/kernel/locking/rwsem-xadd.c b/kernel/locking/rwsem-xadd.c > index 02f6606..1fefe6d 100644 > --- a/kernel/locking/rwsem-xadd.c > +++ b/kernel/locking/rwsem-xadd.c > @@ -613,6 +613,33 @@ struct rw_semaphore *rwsem_wake(struct rw_semaphore *sem) > DEFINE_WAKE_Q(wake_q); > > /* > + * __rwsem_down_write_failed_common(sem) > + * rwsem_optimistic_spin(sem) > + * osq_unlock(sem->osq) > + * ... > + * atomic_long_add_return(&sem->count) > + * > + * - VS - > + * > + * __up_write() > + *if (atomic_long_sub_return_release(&sem->count) < 0) > + * rwsem_wake(sem) > + *osq_is_locked(&sem->osq) > + * > + * And __up_write() must observe !osq_is_locked() when it observes the > + * atomic_long_add_return() in order to not miss a wakeup. > + * > + * This boils down to: > + * > + * [S.rel] X = 1[RmW] r0 = (Y += 0) > + * MB RMB > + * [RmW] Y += 1 [L] r1 = X > + * > + * exists (r0=1 /\ r1=0) > + */ > + smp_rmb(); > + > + /* >* If a spinner is present, it is not necessary to do the wakeup. >* Try to do wakeup only if the trylock succeeds to minimize >* spinlock contention which may introduce too much delay in the > -- > Qualcomm India Private Limited, on behalf of Qualcomm Innovation Center, > Inc., > is a member of Code Aurora Forum, a Linux Foundation Collaborative Project. >
Re: [PATCH] rwsem: fix missed wakeup due to reordering of load
On 08/24/2017 06:22 PM, Peter Zijlstra wrote: > On Thu, Aug 24, 2017 at 02:33:04PM +0200, Peter Zijlstra wrote: >> On Thu, Aug 24, 2017 at 01:29:27PM +0200, Peter Zijlstra wrote: >>> >>> WTH did you not Cc the people that commented on your patch last time? >>> >>> On Wed, Aug 23, 2017 at 04:58:55PM +0530, Prateek Sood wrote: If a spinner is present, there is a chance that the load of rwsem_has_spinner() in rwsem_wake() can be reordered with respect to decrement of rwsem count in __up_write() leading to wakeup being missed. >>> spinning writer up_write caller --- --- [S] osq_unlock() [L] osq spin_lock(wait_lock) sem->count=0x0001 +0x count=sem->count MB sem->count=0xFFFE0001 -0x0001 RMB >>> >>> This doesn't make sense, it appears to order a STORE against something >>> else. >>> spin_trylock(wait_lock) return rwsem_try_write_lock(count) spin_unlock(wait_lock) schedule() >> >> Is this what you wanted to write? > > And ideally there should be a comment near the atomic_long_add_return() > in __rwsem_down_write_failed_common() to indicate we rely on the implied > smp_mb() before it -- just in case someone goes and makes it > atomic_long_add_return_relaxed(). > > And I suppose someone should look at the waiting branch of that thing > too.. because I'm not sure what happens if waiting is true but count > isn't big enough. > > I bloody hate the rwsem code, that BIAS stuff forever confuses me. I > have a start at rewriting the thing to put the owner in the lock word > just like we now do for mutex, but never seem to get around to finishing > it. > >> --- >> kernel/locking/rwsem-xadd.c | 27 +++ >> 1 file changed, 27 insertions(+) >> >> diff --git a/kernel/locking/rwsem-xadd.c b/kernel/locking/rwsem-xadd.c >> index 02f660666ab8..813b5d3654ce 100644 >> --- a/kernel/locking/rwsem-xadd.c >> +++ b/kernel/locking/rwsem-xadd.c >> @@ -613,6 +613,33 @@ struct rw_semaphore *rwsem_wake(struct rw_semaphore >> *sem) >> DEFINE_WAKE_Q(wake_q); >> >> /* >> + * __rwsem_down_write_failed_common(sem) >> + * rwsem_optimistic_spin(sem) >> + * osq_unlock(sem->osq) >> + * ... >> + * atomic_long_add_return(&sem->count) >> + * >> + * - VS - >> + * >> + * __up_write() >> + *if >> (atomic_long_sub_return_release(&sem->count) < 0) >> + * rwsem_wake(sem) >> + *osq_is_locked(&sem->osq) >> + * >> + * And __up_write() must observe !osq_is_locked() when it observes the >> + * atomic_long_add_return() in order to not miss a wakeup. >> + * >> + * This boils down to: >> + * >> + * [S.rel] X = 1[RmW] r0 = (Y += 0) >> + * MB RMB >> + * [RmW] Y += 1 [L] r1 = X >> + * >> + * exists (r0=1 /\ r1=0) >> + */ >> +smp_rmb(); >> + >> +/* >> * If a spinner is present, it is not necessary to do the wakeup. >> * Try to do wakeup only if the trylock succeeds to minimize >> * spinlock contention which may introduce too much delay in the Thanks Peter for your suggestion on comments. I will resend the patch with updated comments -- Qualcomm India Private Limited, on behalf of Qualcomm Innovation Center, Inc., is a member of Code Aurora Forum, a Linux Foundation Collaborative Project
Re: [PATCH] rwsem: fix missed wakeup due to reordering of load
On Thu, Aug 24, 2017 at 02:33:04PM +0200, Peter Zijlstra wrote: > On Thu, Aug 24, 2017 at 01:29:27PM +0200, Peter Zijlstra wrote: > > > > WTH did you not Cc the people that commented on your patch last time? > > > > On Wed, Aug 23, 2017 at 04:58:55PM +0530, Prateek Sood wrote: > > > If a spinner is present, there is a chance that the load of > > > rwsem_has_spinner() in rwsem_wake() can be reordered with > > > respect to decrement of rwsem count in __up_write() leading > > > to wakeup being missed. > > > > > spinning writer up_write caller > > > --- --- > > > [S] osq_unlock() [L] osq > > > spin_lock(wait_lock) > > > sem->count=0x0001 > > > +0x > > > count=sem->count > > > MB > > >sem->count=0xFFFE0001 > > > -0x0001 > > >RMB > > > > This doesn't make sense, it appears to order a STORE against something > > else. > > > > >spin_trylock(wait_lock) > > >return > > > rwsem_try_write_lock(count) > > > spin_unlock(wait_lock) > > > schedule() > > Is this what you wanted to write? And ideally there should be a comment near the atomic_long_add_return() in __rwsem_down_write_failed_common() to indicate we rely on the implied smp_mb() before it -- just in case someone goes and makes it atomic_long_add_return_relaxed(). And I suppose someone should look at the waiting branch of that thing too.. because I'm not sure what happens if waiting is true but count isn't big enough. I bloody hate the rwsem code, that BIAS stuff forever confuses me. I have a start at rewriting the thing to put the owner in the lock word just like we now do for mutex, but never seem to get around to finishing it. > --- > kernel/locking/rwsem-xadd.c | 27 +++ > 1 file changed, 27 insertions(+) > > diff --git a/kernel/locking/rwsem-xadd.c b/kernel/locking/rwsem-xadd.c > index 02f660666ab8..813b5d3654ce 100644 > --- a/kernel/locking/rwsem-xadd.c > +++ b/kernel/locking/rwsem-xadd.c > @@ -613,6 +613,33 @@ struct rw_semaphore *rwsem_wake(struct rw_semaphore *sem) > DEFINE_WAKE_Q(wake_q); > > /* > + * __rwsem_down_write_failed_common(sem) > + * rwsem_optimistic_spin(sem) > + * osq_unlock(sem->osq) > + * ... > + * atomic_long_add_return(&sem->count) > + * > + * - VS - > + * > + * __up_write() > + *if > (atomic_long_sub_return_release(&sem->count) < 0) > + * rwsem_wake(sem) > + *osq_is_locked(&sem->osq) > + * > + * And __up_write() must observe !osq_is_locked() when it observes the > + * atomic_long_add_return() in order to not miss a wakeup. > + * > + * This boils down to: > + * > + * [S.rel] X = 1[RmW] r0 = (Y += 0) > + * MB RMB > + * [RmW] Y += 1 [L] r1 = X > + * > + * exists (r0=1 /\ r1=0) > + */ > + smp_rmb(); > + > + /* >* If a spinner is present, it is not necessary to do the wakeup. >* Try to do wakeup only if the trylock succeeds to minimize >* spinlock contention which may introduce too much delay in the
Re: [PATCH] rwsem: fix missed wakeup due to reordering of load
On Thu, Aug 24, 2017 at 01:29:27PM +0200, Peter Zijlstra wrote: > > WTH did you not Cc the people that commented on your patch last time? > > On Wed, Aug 23, 2017 at 04:58:55PM +0530, Prateek Sood wrote: > > If a spinner is present, there is a chance that the load of > > rwsem_has_spinner() in rwsem_wake() can be reordered with > > respect to decrement of rwsem count in __up_write() leading > > to wakeup being missed. > > > spinning writer up_write caller > > --- --- > > [S] osq_unlock() [L] osq > > spin_lock(wait_lock) > > sem->count=0x0001 > > +0x > > count=sem->count > > MB > >sem->count=0xFFFE0001 > > -0x0001 > >RMB > > This doesn't make sense, it appears to order a STORE against something > else. > > >spin_trylock(wait_lock) > >return > > rwsem_try_write_lock(count) > > spin_unlock(wait_lock) > > schedule() Is this what you wanted to write? --- kernel/locking/rwsem-xadd.c | 27 +++ 1 file changed, 27 insertions(+) diff --git a/kernel/locking/rwsem-xadd.c b/kernel/locking/rwsem-xadd.c index 02f660666ab8..813b5d3654ce 100644 --- a/kernel/locking/rwsem-xadd.c +++ b/kernel/locking/rwsem-xadd.c @@ -613,6 +613,33 @@ struct rw_semaphore *rwsem_wake(struct rw_semaphore *sem) DEFINE_WAKE_Q(wake_q); /* +* __rwsem_down_write_failed_common(sem) +* rwsem_optimistic_spin(sem) +* osq_unlock(sem->osq) +* ... +* atomic_long_add_return(&sem->count) +* +* - VS - +* +* __up_write() +*if (atomic_long_sub_return_release(&sem->count) < 0) +* rwsem_wake(sem) +*osq_is_locked(&sem->osq) +* +* And __up_write() must observe !osq_is_locked() when it observes the +* atomic_long_add_return() in order to not miss a wakeup. +* +* This boils down to: +* +* [S.rel] X = 1[RmW] r0 = (Y += 0) +* MB RMB +* [RmW] Y += 1 [L] r1 = X +* +* exists (r0=1 /\ r1=0) +*/ + smp_rmb(); + + /* * If a spinner is present, it is not necessary to do the wakeup. * Try to do wakeup only if the trylock succeeds to minimize * spinlock contention which may introduce too much delay in the
Re: [PATCH] rwsem: fix missed wakeup due to reordering of load
WTH did you not Cc the people that commented on your patch last time? On Wed, Aug 23, 2017 at 04:58:55PM +0530, Prateek Sood wrote: > If a spinner is present, there is a chance that the load of > rwsem_has_spinner() in rwsem_wake() can be reordered with > respect to decrement of rwsem count in __up_write() leading > to wakeup being missed. > spinning writer up_write caller > --- --- > [S] osq_unlock() [L] osq > spin_lock(wait_lock) > sem->count=0x0001 > +0x > count=sem->count > MB >sem->count=0xFFFE0001 > -0x0001 >RMB This doesn't make sense, it appears to order a STORE against something else. >spin_trylock(wait_lock) >return > rwsem_try_write_lock(count) > spin_unlock(wait_lock) > schedule()
Re: [PATCH] rwsem: fix missed wakeup due to reordering of load
On Thu, Jul 27, 2017 at 01:47:52AM +0530, Prateek Sood wrote: > diff --git a/kernel/locking/rwsem-xadd.c b/kernel/locking/rwsem-xadd.c > index 34e727f..21c111a 100644 > --- a/kernel/locking/rwsem-xadd.c > +++ b/kernel/locking/rwsem-xadd.c > @@ -585,6 +585,40 @@ struct rw_semaphore *rwsem_wake(struct rw_semaphore *sem) > unsigned long flags; > DEFINE_WAKE_Q(wake_q); > > +/* > + * If a spinner is present, there is a chance that the load of > + * rwsem_has_spinner() in rwsem_wake() can be reordered with > + * respect to decrement of rwsem count in __up_write() leading > + * to wakeup being missed. > + * > + * spinning writer up_write caller > + * --- --- > + * [S] osq_unlock() [L] osq > + * spin_lock(wait_lock) > + * sem->count=0x0001 > + *+0x > + * count=sem->count > + * MB > + * sem->count=0xFFFE0001 > + * -0x0001 > + * spin_trylock(wait_lock) > + * return > + * rwsem_try_write_lock(count) > + * spin_unlock(wait_lock) > + * schedule() > + * > + * Reordering of atomic_long_sub_return_release() in __up_write() > + * and rwsem_has_spinner() in rwsem_wake() can cause missing of > + * wakeup in up_write() context. In spinning writer, sem->count > + * and local variable count is 0XFFFE0001. It would result > + * in rwsem_try_write_lock() failing to acquire rwsem and spinning > + * writer going to sleep in rwsem_down_write_failed(). > + * > + * The smp_rmb() here is to make sure that the spinner state is > + * consulted after sem->count is updated in up_write context. I feel that comment can use help.. for example the RMB you add below is not present at all. > + */ > +smp_rmb(); > + > /* >* If a spinner is present, it is not necessary to do the wakeup. >* Try to do wakeup only if the trylock succeeds to minimize Your patch is whitespace damaged, all the indentation on the + lines is with spaces. Please resend with \t.
Re: [PATCH] rwsem: fix missed wakeup due to reordering of load
On Thu, Aug 10, 2017 at 10:32:56AM +0200, Andrea Parri wrote: > On Thu, Jul 27, 2017 at 11:48:53AM -0400, Waiman Long wrote: > > On 07/26/2017 04:17 PM, Prateek Sood wrote: > > > If a spinner is present, there is a chance that the load of > > > rwsem_has_spinner() in rwsem_wake() can be reordered with > > > respect to decrement of rwsem count in __up_write() leading > > > to wakeup being missed. > > > > > > spinning writer up_write caller > > > --- --- > > > [S] osq_unlock() [L] osq > > > spin_lock(wait_lock) > > > sem->count=0x0001 > > > +0x > > > count=sem->count > > > MB > > >sem->count=0xFFFE0001 > > > -0x0001 > > >spin_trylock(wait_lock) > > >return > > > rwsem_try_write_lock(count) > > > spin_unlock(wait_lock) > > > schedule() > > > > > > Reordering of atomic_long_sub_return_release() in __up_write() > > > and rwsem_has_spinner() in rwsem_wake() can cause missing of > > > wakeup in up_write() context. In spinning writer, sem->count > > > and local variable count is 0XFFFE0001. It would result > > > in rwsem_try_write_lock() failing to acquire rwsem and spinning > > > writer going to sleep in rwsem_down_write_failed(). > > > > > > The smp_rmb() will make sure that the spinner state is > > > consulted after sem->count is updated in up_write context. > > > > > > Signed-off-by: Prateek Sood > > > > Did you actually observe that the reordering happens? > > > > I am not sure if some architectures can actually speculatively execute > > instruction ahead of a branch and then ahead into a function call. I > > know it can happen if the function call is inlined, but rwsem_wake() > > will not be inlined into __up_read() or __up_write(). > > Branches/control dependencies targeting a read do not necessarily preserve > program order; this is for example the case for PowerPC and ARM. > > I'd not expect more than a compiler barrier from a function call (in fact, > not even that if the function happens to be inlined). Indeed. Reads can be speculated by deep out-of-order CPUs no problem. That's what branch predictors are for. > > Even if that is the case, I am not sure if smp_rmb() alone is enough to > > guarantee the ordering as I think it will depend on how the > > atomic_long_sub_return_release() is implmented. > > AFAICT, the pattern under discussion is MP with: > > - a store-release to osq->tail(unlock) followed by a store to sem->count, > separated by a MB (from atomic_long_add_return()) on CPU0; > > - a load of sem->count (for atomic_long_sub_return_release()) followed by Which is a regular load, as 'release' only need apply to the store. > a load of osq->tail (rwsem_has_spinner()) on CPU1. > > Thus a RMW between the two loads suffices to forbid the weak behaviour. Agreed.
Re: [PATCH] rwsem: fix missed wakeup due to reordering of load
On Thu, Jul 27, 2017 at 11:48:53AM -0400, Waiman Long wrote: > On 07/26/2017 04:17 PM, Prateek Sood wrote: > > If a spinner is present, there is a chance that the load of > > rwsem_has_spinner() in rwsem_wake() can be reordered with > > respect to decrement of rwsem count in __up_write() leading > > to wakeup being missed. > > > > spinning writer up_write caller > > --- --- > > [S] osq_unlock() [L] osq > > spin_lock(wait_lock) > > sem->count=0x0001 > > +0x > > count=sem->count > > MB > >sem->count=0xFFFE0001 > > -0x0001 > >spin_trylock(wait_lock) > >return > > rwsem_try_write_lock(count) > > spin_unlock(wait_lock) > > schedule() > > > > Reordering of atomic_long_sub_return_release() in __up_write() > > and rwsem_has_spinner() in rwsem_wake() can cause missing of > > wakeup in up_write() context. In spinning writer, sem->count > > and local variable count is 0XFFFE0001. It would result > > in rwsem_try_write_lock() failing to acquire rwsem and spinning > > writer going to sleep in rwsem_down_write_failed(). > > > > The smp_rmb() will make sure that the spinner state is > > consulted after sem->count is updated in up_write context. > > > > Signed-off-by: Prateek Sood > > Did you actually observe that the reordering happens? > > I am not sure if some architectures can actually speculatively execute > instruction ahead of a branch and then ahead into a function call. I > know it can happen if the function call is inlined, but rwsem_wake() > will not be inlined into __up_read() or __up_write(). Branches/control dependencies targeting a read do not necessarily preserve program order; this is for example the case for PowerPC and ARM. I'd not expect more than a compiler barrier from a function call (in fact, not even that if the function happens to be inlined). > > Even if that is the case, I am not sure if smp_rmb() alone is enough to > guarantee the ordering as I think it will depend on how the > atomic_long_sub_return_release() is implmented. AFAICT, the pattern under discussion is MP with: - a store-release to osq->tail(unlock) followed by a store to sem->count, separated by a MB (from atomic_long_add_return()) on CPU0; - a load of sem->count (for atomic_long_sub_return_release()) followed by a load of osq->tail (rwsem_has_spinner()) on CPU1. Thus a RMW between the two loads suffices to forbid the weak behaviour. Andrea > > Cheers, > Longman >
Re: [PATCH] rwsem: fix missed wakeup due to reordering of load
On Thu, Jul 27, 2017 at 11:48:53AM -0400, Waiman Long wrote: > atomic_long_sub_return_release() is implmented. I've not had time to really thing about the problem at hand, but this I can answer: TSO (x86, s390, sparc): fully serialized PPC: lwsync; ll/sc (RCpc) ARM64: ll/sc-release(RCsc) other: smp_mb(); atomic_sub_return_relaxed();
Re: [PATCH] rwsem: fix missed wakeup due to reordering of load
On 07/26/2017 04:17 PM, Prateek Sood wrote: > If a spinner is present, there is a chance that the load of > rwsem_has_spinner() in rwsem_wake() can be reordered with > respect to decrement of rwsem count in __up_write() leading > to wakeup being missed. > > spinning writer up_write caller > --- --- > [S] osq_unlock() [L] osq > spin_lock(wait_lock) > sem->count=0x0001 > +0x > count=sem->count > MB >sem->count=0xFFFE0001 > -0x0001 >spin_trylock(wait_lock) >return > rwsem_try_write_lock(count) > spin_unlock(wait_lock) > schedule() > > Reordering of atomic_long_sub_return_release() in __up_write() > and rwsem_has_spinner() in rwsem_wake() can cause missing of > wakeup in up_write() context. In spinning writer, sem->count > and local variable count is 0XFFFE0001. It would result > in rwsem_try_write_lock() failing to acquire rwsem and spinning > writer going to sleep in rwsem_down_write_failed(). > > The smp_rmb() will make sure that the spinner state is > consulted after sem->count is updated in up_write context. > > Signed-off-by: Prateek Sood Did you actually observe that the reordering happens? I am not sure if some architectures can actually speculatively execute instruction ahead of a branch and then ahead into a function call. I know it can happen if the function call is inlined, but rwsem_wake() will not be inlined into __up_read() or __up_write(). Even if that is the case, I am not sure if smp_rmb() alone is enough to guarantee the ordering as I think it will depend on how the atomic_long_sub_return_release() is implmented. Cheers, Longman