On 22.01.2019 20:01, Tiwei Bie wrote:
> This patch mainly fixed below issues in the packed ring based
> control vq support in virtio driver:
> 
> 1. When parsing the used descriptors, we have to track the
>    number of descs that we need to skip;
> 2. vq->vq_free_cnt was decreased twice for a same desc;
> 
> Meanwhile, make the function name consistent with other parts.
> 
> Fixes: ec194c2f1895 ("net/virtio: support packed queue in send command")
> Fixes: a4270ea4ff79 ("net/virtio: check head desc with correct wrap counter")
> 
> Signed-off-by: Tiwei Bie <tiwei....@intel.com>
> ---
>  drivers/net/virtio/virtio_ethdev.c | 62 ++++++++++++++----------------
>  drivers/net/virtio/virtqueue.h     | 12 +-----
>  2 files changed, 31 insertions(+), 43 deletions(-)
> 
> diff --git a/drivers/net/virtio/virtio_ethdev.c 
> b/drivers/net/virtio/virtio_ethdev.c
> index ee5a98b7c..a3fe65599 100644
> --- a/drivers/net/virtio/virtio_ethdev.c
> +++ b/drivers/net/virtio/virtio_ethdev.c
> @@ -142,16 +142,17 @@ static const struct rte_virtio_xstats_name_off 
> rte_virtio_txq_stat_strings[] = {
>  struct virtio_hw_internal virtio_hw_internal[RTE_MAX_ETHPORTS];
>  
>  static struct virtio_pmd_ctrl *
> -virtio_pq_send_command(struct virtnet_ctl *cvq, struct virtio_pmd_ctrl *ctrl,
> -                    int *dlen, int pkt_num)
> +virtio_send_command_packed(struct virtnet_ctl *cvq,
> +                        struct virtio_pmd_ctrl *ctrl,
> +                        int *dlen, int pkt_num)
>  {
>       struct virtqueue *vq = cvq->vq;
>       int head;
>       struct vring_packed_desc *desc = vq->ring_packed.desc_packed;
>       struct virtio_pmd_ctrl *result;
> -     bool avail_wrap_counter, used_wrap_counter;
> -     uint16_t flags;
> +     bool avail_wrap_counter;
>       int sum = 0;
> +     int nb_descs = 0;
>       int k;
>  
>       /*
> @@ -162,11 +163,10 @@ virtio_pq_send_command(struct virtnet_ctl *cvq, struct 
> virtio_pmd_ctrl *ctrl,
>        */
>       head = vq->vq_avail_idx;
>       avail_wrap_counter = vq->avail_wrap_counter;
> -     used_wrap_counter = vq->used_wrap_counter;
> -     desc[head].flags = VRING_DESC_F_NEXT;
>       desc[head].addr = cvq->virtio_net_hdr_mem;
>       desc[head].len = sizeof(struct virtio_net_ctrl_hdr);
>       vq->vq_free_cnt--;
> +     nb_descs++;
>       if (++vq->vq_avail_idx >= vq->vq_nentries) {
>               vq->vq_avail_idx -= vq->vq_nentries;
>               vq->avail_wrap_counter ^= 1;
> @@ -177,55 +177,51 @@ virtio_pq_send_command(struct virtnet_ctl *cvq, struct 
> virtio_pmd_ctrl *ctrl,
>                       + sizeof(struct virtio_net_ctrl_hdr)
>                       + sizeof(ctrl->status) + sizeof(uint8_t) * sum;
>               desc[vq->vq_avail_idx].len = dlen[k];
> -             flags = VRING_DESC_F_NEXT;

Looks like barriers was badly placed here before this patch.
Anyway, you need a write barrier here between {addr, len} and flags updates.

> +             desc[vq->vq_avail_idx].flags = VRING_DESC_F_NEXT |
> +                     VRING_DESC_F_AVAIL(vq->avail_wrap_counter) |
> +                     VRING_DESC_F_USED(!vq->avail_wrap_counter);
>               sum += dlen[k];
>               vq->vq_free_cnt--;
> -             flags |= VRING_DESC_F_AVAIL(vq->avail_wrap_counter) |
> -                      VRING_DESC_F_USED(!vq->avail_wrap_counter);
> -             desc[vq->vq_avail_idx].flags = flags;
> -             rte_smp_wmb();
> -             vq->vq_free_cnt--;
> +             nb_descs++;
>               if (++vq->vq_avail_idx >= vq->vq_nentries) {
>                       vq->vq_avail_idx -= vq->vq_nentries;
>                       vq->avail_wrap_counter ^= 1;
>               }
>       }
>  
> -
>       desc[vq->vq_avail_idx].addr = cvq->virtio_net_hdr_mem
>               + sizeof(struct virtio_net_ctrl_hdr);
>       desc[vq->vq_avail_idx].len = sizeof(ctrl->status);
> -     flags = VRING_DESC_F_WRITE;
> -     flags |= VRING_DESC_F_AVAIL(vq->avail_wrap_counter) |
> -              VRING_DESC_F_USED(!vq->avail_wrap_counter);
> -     desc[vq->vq_avail_idx].flags = flags;
> -     flags = VRING_DESC_F_NEXT;
> -     flags |= VRING_DESC_F_AVAIL(avail_wrap_counter) |
> -              VRING_DESC_F_USED(!avail_wrap_counter);
> -     desc[head].flags = flags;
> -     rte_smp_wmb();
> -

Same here. We need a write barrier to be sure that {addr, len} written
before updating flags.

Another way to avoid most of barriers is to work similar to
'flush_shadow_used_ring_packed',
i.e. update all the data in a loop - write barrier - update all the flags.

> +     desc[vq->vq_avail_idx].flags = VRING_DESC_F_WRITE |
> +             VRING_DESC_F_AVAIL(vq->avail_wrap_counter) |
> +             VRING_DESC_F_USED(!vq->avail_wrap_counter);
>       vq->vq_free_cnt--;
> +     nb_descs++;
>       if (++vq->vq_avail_idx >= vq->vq_nentries) {
>               vq->vq_avail_idx -= vq->vq_nentries;
>               vq->avail_wrap_counter ^= 1;
>       }
>  
> +     virtio_wmb(vq->hw->weak_barriers);
> +     desc[head].flags = VRING_DESC_F_NEXT |
> +             VRING_DESC_F_AVAIL(avail_wrap_counter) |
> +             VRING_DESC_F_USED(!avail_wrap_counter);
> +
> +     virtio_wmb(vq->hw->weak_barriers);
>       virtqueue_notify(vq);
>  
>       /* wait for used descriptors in virtqueue */
> -     do {
> -             rte_rmb();
> +     while (!desc_is_used(&desc[head], vq))
>               usleep(100);
> -     } while (!__desc_is_used(&desc[head], used_wrap_counter));
> +
> +     virtio_rmb(vq->hw->weak_barriers);
>  
>       /* now get used descriptors */
> -     while (desc_is_used(&desc[vq->vq_used_cons_idx], vq)) {
> -             vq->vq_free_cnt++;
> -             if (++vq->vq_used_cons_idx >= vq->vq_nentries) {
> -                     vq->vq_used_cons_idx -= vq->vq_nentries;
> -                     vq->used_wrap_counter ^= 1;
> -             }
> +     vq->vq_free_cnt += nb_descs;
> +     vq->vq_used_cons_idx += nb_descs;
> +     if (vq->vq_used_cons_idx >= vq->vq_nentries) {
> +             vq->vq_used_cons_idx -= vq->vq_nentries;
> +             vq->used_wrap_counter ^= 1;
>       }
>  
>       result = cvq->virtio_net_hdr_mz->addr;
> @@ -266,7 +262,7 @@ virtio_send_command(struct virtnet_ctl *cvq, struct 
> virtio_pmd_ctrl *ctrl,
>               sizeof(struct virtio_pmd_ctrl));
>  
>       if (vtpci_packed_queue(vq->hw)) {
> -             result = virtio_pq_send_command(cvq, ctrl, dlen, pkt_num);
> +             result = virtio_send_command_packed(cvq, ctrl, dlen, pkt_num);
>               goto out_unlock;
>       }
>  
> diff --git a/drivers/net/virtio/virtqueue.h b/drivers/net/virtio/virtqueue.h
> index 7fcde5643..ca9d8e6e3 100644
> --- a/drivers/net/virtio/virtqueue.h
> +++ b/drivers/net/virtio/virtqueue.h
> @@ -281,7 +281,7 @@ struct virtio_tx_region {
>  };
>  
>  static inline int
> -__desc_is_used(struct vring_packed_desc *desc, bool wrap_counter)
> +desc_is_used(struct vring_packed_desc *desc, struct virtqueue *vq)
>  {
>       uint16_t used, avail, flags;
>  
> @@ -289,16 +289,9 @@ __desc_is_used(struct vring_packed_desc *desc, bool 
> wrap_counter)
>       used = !!(flags & VRING_DESC_F_USED(1));
>       avail = !!(flags & VRING_DESC_F_AVAIL(1));
>  
> -     return avail == used && used == wrap_counter;
> +     return avail == used && used == vq->used_wrap_counter;
>  }
>  
> -static inline int
> -desc_is_used(struct vring_packed_desc *desc, struct virtqueue *vq)
> -{
> -     return __desc_is_used(desc, vq->used_wrap_counter);
> -}
> -
> -
>  static inline void
>  vring_desc_init_packed(struct virtqueue *vq, int n)
>  {
> @@ -354,7 +347,6 @@ virtqueue_enable_intr_packed(struct virtqueue *vq)
>  {
>       uint16_t *event_flags = &vq->ring_packed.driver_event->desc_event_flags;
>  
> -
>       if (vq->event_flags_shadow == RING_EVENT_FLAGS_DISABLE) {
>               virtio_wmb(vq->hw->weak_barriers);
>               vq->event_flags_shadow = RING_EVENT_FLAGS_ENABLE;
> 

Reply via email to