Hi Wathsala,

> >
> > Thanks for the clarification, I now see your concern.
> >
> > So, in summary what you are saying is that tail update from
> > Thread #1 that happened at T0 is not observed by the thread #2
> > at T2 when it computed new_head and entries calculation.

Yes

> > That cannot happen in Arm v8/v9 because tail update is a
> > store-release and a load-acquire that program order follows it
> > can only be issued after all the cores have observed the
> > store-release (there is a synchronization with relationship to
> > store-release and load-acquire pairs).
> >
> > In the example you have provided Thread #1's
> > store(&prod.tail, 2, release) is observed by all the cores in the
> > system by the time same thread performs load(&prod.tail, acquire)
> > at T2. As per Arm v8/v9 memory model Thread #2 should observe
> > prod.tail=2 (not prod.tail=1).
> >
> > Arm Architecture Reference Manual section B2.10.11 states…
> >
> > "Where a Load-Acquire appears in program order after a Store-Release,
> > the memory access generated by the Store-Release instruction is
> > observed by each PE to the extent that PE is required to observe the
> > access coherently, before the memory access generated by the
> > Load-Acquire instruction is observed by that PE, to the extent that the
> > PE is required to observe the access coherently."

Interesting, thanks for pointing to that.
Indeed, according to the docs, acquire/release operations use a sequentially 
consistent model.
But we not always getting pure LDAR instruction here, it depends on compiler 
march/mcpu  flags.
For some HW arch gcc (and clang) generate LDAR instruction, for others - LDPAPR.
And according to the doc:
https://developer.arm.com/documentation/102336/0100/Load-Acquire-and-Store-Release-instructions
for Load-AcquirePC  the requirement that Load-Acquires are observed in order 
with preceding Store-Releases is dropped.  
To be more specific here, on my box:
1) meson with -Dmachine=default (gcc options: "-march=armv8-a+crc") generates 
code with LDAR instruction
https://godbolt.org/z/5cEscdxrb
2) meson with -Dmachine=native (gcc options: " -mcpu=neoverse-n1") generates 
code with LDAPR instruction 
https://godbolt.org/z/8KjcvYxch

soring_stress_test for 1) (LDAR) passes
soring_stress_test for 2) (LDAPR) fails with symptoms I described earlier 

> 
> In soring is this the pair that update the tail and move head?
> 
> __rte_soring_update_tail:
>       __rte_ring_update_tail(&rht->ht, head, next, st, enq);

That function is used to update prod.tail and cons.tail

> __rte_soring_move_cons_head:
>        __rte_ring_headtail_move_head(&r->cons.ht,  &r->stage[stage].ht, 0, 
> ...);
> 

__rte_ring_headtail_move_head is invoked to update prod.head (by 
__rte_soring_move_prod_head)
 and cons.head (by __rte_soring_move_cons_head). 

> If so, &rht->ht and &r->stage[stage].ht are the same address?

No.
In conventional ring we have just two stages: 'prod'  and 'cons'.
So moving prod.head depends on cons.tail and visa-versa: moving cons.head 
depends on prod.tail.
With soring we have multiple (N) extra stages in between.
So the dependency chain is a bit longer:
prod.head depends on cons.tail
stage[0].head depends on prod.tail,
stage[1].head depends on stage[0].tail
...
stage{N-1].head depends on stage[N-2].tail
cons.head depends on stage[N-1].tail

> If they are
> not, then you will run into the issue you have seen (a.k.a "Other-multi-copy
>  atomic" which is legit in Arm v8 and above).

Can you probably elaborate a bit more for me here?
Thanks
Konstantin
 

> Thanks.
> 
> --wathsala
> 
> 
> > > -----Original Message-----
> > > From: Konstantin Ananyev <konstantin.anan...@huawei.com>
> > > Sent: Monday, May 26, 2025 6:54 AM
> > > To: Wathsala Wathawana Vithanage <wathsala.vithan...@arm.com>;
> > > dev@dpdk.org
> > > Cc: Honnappa Nagarahalli <honnappa.nagaraha...@arm.com>;
> > > jer...@marvell.com; hemant.agra...@nxp.com; d...@linux.ibm.com; nd
> > > <n...@arm.com>
> > > Subject: RE: [PATCH v1 3/4] ring: fix potential sync issue between head 
> > > and
> > tail
> > > values
> > >
> > > Hi Wathsala,
> > >
> > > >
> > > > Hi Konstanin,
> > > >
> > > > In rte_ring the store-release on tail update guarantees that CAS
> > > > won't get reordered with the store-released of the tail update.
> > > >
> > > > So, the sequence of events would look like this (combined view
> > > > of head and tail update)
> > > >
> > > > Releaxed-load(new_head,  N)                              
> > > > ----------------> (A)
> > > > Relaxed-CAS(d->head, new_head, old_head)   ----------------> (B)
> > > > Store-release-store(d->tail, new_head)             ----------------> (C)
> > > >
> > > > If we look at address dependencies, then...
> > > >
> > > > (B) depends on (A) due to new_head address dependency.
> > > > (C) depends on (A) due to new_head address dependency.
> > > >
> > > > So, dependency graph looks like this
> > > >        (A)
> > > >     /       \
> > > >    v        v
> > > >  (B)     (C)
> > > >
> > > > There is no implicit dependence between (B) and (C), I think
> > > > this is the issue you are brining up.
> > > > Even though there is no dependence between the two,
> > > > the store-release of (C) ensures that (B) won't drop below it.
> > > > Therefore, the above graph can be turned into an ordered
> > > > sequence as shown below..
> > > >
> > > > (A) -> (B) -> (C)
> > >
> > > I do agree that with current implementation of
> > > __rte_ring_headtail_move_head()
> > > in lib/ring/rte_ring_c11_pvt.h the order of these 3 operations (A->B->C)
> > > should be sequential.
> > > The problem I am talking about is a different one:
> > > thread can see 'latest' 'cons.head' value, with 'previous' value for 
> > > 'prod.tail' or
> > > visa-versa.
> > > In other words: 'cons.head' value depends on 'prod.tail', so before making
> > > latest 'cons.head'
> > > value visible to other threads, we need to ensure that latest 'prod.tail' 
> > > is also
> > > visible.
> > > Let me try to explain it on the example:
> > >
> > > Suppose at some moment we have:
> > > prod={.head=2,.tail=1};
> > > cons={.head=0,.tail=0};
> > > I.e. thead #1 is in process to enqueue one more element into the ring.
> > >
> > >        Thread #1                                                          
> > >                    Thread #2
> > > T0:
> > >   store(&prod.tail, 2, release);
> > >   /*AFAIU: this is somewhat equivalent to: wmb(); prod.tail=2;
> > >   * I.E. - it guarantees that all stores initiated before that operation 
> > > will be
> > > visible
> > >   * by other threads at the same moment or before new value of prod.tail 
> > > wlll
> > > become
> > >   * visible, but it doesn't guarantee that new prod.tail  value will be 
> > > visible  to
> > > other
> > >   * threads immediately.
> > >   */
> > > ...
> > > move_cons_head(...,n=2)
> > move_cons_head(...,n=1)
> > > ...                                                                       
> > >                        ...
> > > T1:
> > >   *old_head = load(&cons.head, relaxed);
> > >   fence(acquire);
> > >   /*old_head == 0, no surprises */
> > >   stail = load(&prod.tail, acquire);
> > >   /*stail == 2, no surprises */
> > >  *entries = (capacity + stail - *old_head);
> > > *new_head = *old_head + n;
> > >  /* *entries == (2 - 0) == 2; *new_head==2; all good */
> > > ...
> > > T2:
> > >                                                                           
> > >                       *old_head =
> > > load(&cons.head, relaxed);
> > >                                                                           
> > >                       fence(acquire);
> > >                                                                           
> > >                       /*old_head == 0, no
> > surprises
> > > */
> > >                                                                           
> > >                       stail = load(&prod.tail,
> > > acquire);
> > > /* !!!!! stail == 1  !!!!! for Thread 2
> > >  * Even though we do use acquire here - there was no *release* after
> > > store(&prod.tail).
> > >  * So, still no guarantee that Thread 2 will see latest prod.tail value.
> > >  */
> > >                                                                           
> > >                      *entries = (capacity + stail -
> > > *old_head);
> > >                                                                           
> > >                      /* *entries == (1 - 0) == 1,
> > still
> > > ok */
> > >                                                                           
> > >                      *new_head = *old_head + n;
> > >                                                                           
> > >                      /* *new_head == 1 */
> > > T3:
> > >   success = CAS(&cons.head,
> > >     old_head /*==0/, *new_head /*==2*/,
> > >     relaxed, relaxed);
> > >   /*success==1, cons.head==2*/
> > >  ...
> > > T4:
> > >                                                                           
> > >                  success = CAS(&cons.head,
> > >                                                                           
> > >                  old_head /*==0/, *new_head
> > > /*==1*/,
> > >                                                                           
> > >                 relaxed, relaxed);
> > >                                                                           
> > >                 /*success==0, *old_head==2*/
> > > /* CAS() failed and provided Thread 2 with latest valued for 
> > > cons.head(==2)
> > >  *  Thread 2 repeats attempt, starts second iteration
> > >  */
> > >                                                                           
> > >                 fence(acquire);
> > >                                                                           
> > >                 stail = load(&prod.tail, acquire);
> > > /* !!!!! stail == 1  !!!!! for Thread 2
> > >  * Still no *release* had happened after store(&prod.tail) at T0.
> > >  * So, still no guarantee that Thread 2 will see latest prod.tail value.
> > >  */
> > >                                                                           
> > >                 *entries = (capacity + stail -
> > > *old_head);
> > >                                                                           
> > >                 *new_head = *old_head + n;
> > >
> > > /* !!!!! *entries == (1 - 2) == -1(UINT32_MAX); *new_head==(2+1)==3; !!!!!
> > >  *  we are about to corrupt our ring !!!!!
> > >  */
> > >
> > > >
> > > > I haven't looked at the so-ring yet. Could it be possible that the
> > > > issue is due to something else introduced in that code?
> > >
> > > Well, as I said, so far I wasn't able to re-produce this problem with
> > > conventional
> > > ring (ring_stress_autotest), only soring_stress_autotest is failing and
> > > for now - only on one specific ARM platform.
> > > Regarding soring specific fix (without touching common code) -
> > > sure it is also possible, pls see patch #2.
> > > There I just added 'fence(release);' straight after 'store(&tail);'
> > > That's seems enough to fix that problem within the soring only.
> > > Though, from my understanding rte_ring might also be affected,
> > > that's why I went ahead and prepared that patch.
> > > If you feel, that I a missing something here - pls shout.
> > > Konstantin
> > >
> > >
> > > > Thanks,
> > > >
> > > > --wathsala
> > > >
> > > > > This patch aims several purposes:
> > > > > - provide an alternative (and I think a better) way to fix the
> > > > >   issue discussed in previous patch:
> > > > >   "ring/soring: fix synchronization issue between head and tail 
> > > > > values"
> > > > > - make sure that such problem wouldn’t happen within other usages of
> > > > >   __rte_ring_headtail_move_head() – both current rte_ring
> > > > >   implementation and possible future use-cases.
> > > > > - step towards unification of move_head() implementations and
> > > > >   removing rte_ring_generic_pvt.h
> > > > > It uses Acquire-Release memory ordering for CAS operation in
> > > move_head().
> > > > > That guarantees that corresponding ‘tail’ updates will be visible 
> > > > > before
> > > current
> > > > > ‘head’ is updated.
> > > > > As I said before: I think that in theory the problem described in 
> > > > > previous
> > > patch
> > > > > might happen with our conventional rte_ring too (when
> > > > > RTE_USE_C11_MEM_MODEL enabled).
> > > > > But, so far I didn’t manage to reproduce it in reality.
> > > > > For that reason and also because it touches a critical rte_ring 
> > > > > code-path, I
> > > put
> > > > > these changes into a separate patch. Expect all interested 
> > > > > stakeholders to
> > > come-
> > > > > up with their comments and observations.
> > > > > Regarding performance impact – on my boxes both ring_perf_autotest
> > and
> > > > > ring_stress_autotest – show a mixed set of results: some of them
> > become
> > > few
> > > > > cycles faster, another few cycles slower.
> > > > > But so far, I didn’t notice any real degradations with that patch.
> > > > >
> > > > > Fixes: b5458e2cc483 ("ring: introduce staged ordered ring")
> > > > > Fixes: 1cc363b8ce06 ("ring: introduce HTS ring mode")
> > > > > Fixes: e6ba4731c0f3 ("ring: introduce RTS ring mode")
> > > > > Fixes: 49594a63147a ("ring/c11: relax ordering for load and store of 
> > > > > the
> > > head")
> > > > >
> > > > > Signed-off-by: Konstantin Ananyev <konstantin.anan...@huawei.com>
> > > > > ---
> > > > >  lib/ring/rte_ring_c11_pvt.h      | 27 +++++++++++++++++----------
> > > > >  lib/ring/rte_ring_hts_elem_pvt.h |  6 ++++--
> > > lib/ring/rte_ring_rts_elem_pvt.h
> > > > > |  6 ++++--
> > > > >  lib/ring/soring.c                |  5 -----
> > > > >  4 files changed, 25 insertions(+), 19 deletions(-)
> > > > >
> > > > > diff --git a/lib/ring/rte_ring_c11_pvt.h b/lib/ring/rte_ring_c11_pvt.h
> > index
> > > > > 0845cd6dcf..6d1c46df9a 100644
> > > > > --- a/lib/ring/rte_ring_c11_pvt.h
> > > > > +++ b/lib/ring/rte_ring_c11_pvt.h
> > > > > @@ -77,20 +77,19 @@ __rte_ring_headtail_move_head(struct
> > > > > rte_ring_headtail *d,
> > > > >       int success;
> > > > >       unsigned int max = n;
> > > > >
> > > > > +     /* Ensure the head is read before tail */
> > > > >       *old_head = rte_atomic_load_explicit(&d->head,
> > > > > -                     rte_memory_order_relaxed);
> > > > > +                     rte_memory_order_acquire);
> > > > >       do {
> > > > >               /* Reset n to the initial burst count */
> > > > >               n = max;
> > > > >
> > > > > -             /* Ensure the head is read before tail */
> > > > > -             rte_atomic_thread_fence(rte_memory_order_acquire);
> > > > > -
> > > > > -             /* load-acquire synchronize with store-release of 
> > > > > ht->tail
> > > > > -              * in update_tail.
> > > > > +             /*
> > > > > +              * Read s->tail value. Note that it will be loaded after
> > > > > +              * d->head load, but before CAS operation for the 
> > > > > d->head.
> > > > >                */
> > > > >               stail = rte_atomic_load_explicit(&s->tail,
> > > > > -                                     rte_memory_order_acquire);
> > > > > +                                     rte_memory_order_relaxed);
> > > > >
> > > > >               /* The subtraction is done between two unsigned 32bits 
> > > > > value
> > > > >                * (the result is always modulo 32 bits even if we have 
> > > > > @@ -
> > > > > 112,11 +111,19 @@ __rte_ring_headtail_move_head(struct
> > > rte_ring_headtail
> > > > > *d,
> > > > >                       d->head = *new_head;
> > > > >                       success = 1;
> > > > >               } else
> > > > > -                     /* on failure, *old_head is updated */
> > > > > +                     /*
> > > > > +                      * on failure, *old_head is updated.
> > > > > +                      * this CAS(ACQ_REL, ACQUIRE) serves as a hoist
> > > > > +                      * barrier to prevent:
> > > > > +                      *  - OOO reads of cons tail value
> > > > > +                      *  - OOO copy of elems from the ring
> > > > > +                      *  Also RELEASE guarantees that latest tail 
> > > > > value
> > > > > +                      *  will become visible before the new head 
> > > > > value.
> > > > > +                      */
> > > > >                       success =
> > > > > rte_atomic_compare_exchange_strong_explicit(
> > > > >                                       &d->head, old_head, *new_head,
> > > > > -                                     rte_memory_order_relaxed,
> > > > > -                                     rte_memory_order_relaxed);
> > > > > +                                     rte_memory_order_acq_rel,
> > > > > +                                     rte_memory_order_acquire);
> > > > >       } while (unlikely(success == 0));
> > > > >       return n;
> > > > >  }
> > > > > diff --git a/lib/ring/rte_ring_hts_elem_pvt.h
> > > b/lib/ring/rte_ring_hts_elem_pvt.h
> > > > > index c59e5f6420..cc593433b9 100644
> > > > > --- a/lib/ring/rte_ring_hts_elem_pvt.h
> > > > > +++ b/lib/ring/rte_ring_hts_elem_pvt.h
> > > > > @@ -116,13 +116,15 @@ __rte_ring_hts_move_head(struct
> > > > > rte_ring_hts_headtail *d,
> > > > >               np.pos.head = op.pos.head + n;
> > > > >
> > > > >       /*
> > > > > -      * this CAS(ACQUIRE, ACQUIRE) serves as a hoist barrier to 
> > > > > prevent:
> > > > > +      * this CAS(ACQ_REL, ACQUIRE) serves as a hoist barrier to 
> > > > > prevent:
> > > > >        *  - OOO reads of cons tail value
> > > > >        *  - OOO copy of elems from the ring
> > > > > +      *   Also RELEASE guarantees that latest tail value
> > > > > +      *   will become visible before the new head value.
> > > > >        */
> > > > >       } while (rte_atomic_compare_exchange_strong_explicit(&d->ht.raw,
> > > > >                       (uint64_t *)(uintptr_t)&op.raw, np.raw,
> > > > > -                     rte_memory_order_acquire,
> > > > > +                     rte_memory_order_acq_rel,
> > > > >                       rte_memory_order_acquire) == 0);
> > > > >
> > > > >       *old_head = op.pos.head;
> > > > > diff --git a/lib/ring/rte_ring_rts_elem_pvt.h
> > > b/lib/ring/rte_ring_rts_elem_pvt.h
> > > > > index 509fa674fb..860b13cc61 100644
> > > > > --- a/lib/ring/rte_ring_rts_elem_pvt.h
> > > > > +++ b/lib/ring/rte_ring_rts_elem_pvt.h
> > > > > @@ -131,13 +131,15 @@ __rte_ring_rts_move_head(struct
> > > > > rte_ring_rts_headtail *d,
> > > > >               nh.val.cnt = oh.val.cnt + 1;
> > > > >
> > > > >       /*
> > > > > -      * this CAS(ACQUIRE, ACQUIRE) serves as a hoist barrier to 
> > > > > prevent:
> > > > > +      * this CAS(ACQ_REL, ACQUIRE) serves as a hoist barrier to 
> > > > > prevent:
> > > > >        *  - OOO reads of cons tail value
> > > > >        *  - OOO copy of elems to the ring
> > > > > +      *  Also RELEASE guarantees that latest tail value
> > > > > +      *  will become visible before the new head value.
> > > > >        */
> > > > >       } while (rte_atomic_compare_exchange_strong_explicit(&d-
> > > >head.raw,
> > > > >                       (uint64_t *)(uintptr_t)&oh.raw, nh.raw,
> > > > > -                     rte_memory_order_acquire,
> > > > > +                     rte_memory_order_acq_rel,
> > > > >                       rte_memory_order_acquire) == 0);
> > > > >
> > > > >       *old_head = oh.val.pos;
> > > > > diff --git a/lib/ring/soring.c b/lib/ring/soring.c index
> > > 7bcbf35516..21a1a27e24
> > > > > 100644
> > > > > --- a/lib/ring/soring.c
> > > > > +++ b/lib/ring/soring.c
> > > > > @@ -123,8 +123,6 @@ __rte_soring_stage_finalize(struct
> > > > > soring_stage_headtail *sht, uint32_t stage,
> > > > >       rte_atomic_store_explicit(&sht->tail.raw, ot.raw,
> > > > >                       rte_memory_order_release);
> > > > >
> > > > > -     /* make sure that new tail value is visible */
> > > > > -     rte_atomic_thread_fence(rte_memory_order_release);
> > > > >       return i;
> > > > >  }
> > > > >
> > > > > @@ -219,9 +217,6 @@ __rte_soring_update_tail(struct
> > > __rte_ring_headtail
> > > > > *rht,
> > > > >               /* unsupported mode, shouldn't be here */
> > > > >               RTE_ASSERT(0);
> > > > >       }
> > > > > -
> > > > > -     /* make sure that new tail value is visible */
> > > > > -     rte_atomic_thread_fence(rte_memory_order_release);
> > > > >  }
> > > > >
> > > > >  static __rte_always_inline uint32_t
> > > > > --
> > > > > 2.43.0

Reply via email to