On Fri, Mar 28, 2025 at 6:10 AM Sahil Siddiq <icegambi...@gmail.com> wrote: > > Hi, > > On 3/26/25 5:32 PM, Eugenio Perez Martin wrote: > > On Mon, Mar 24, 2025 at 3:00 PM Sahil Siddiq <icegambi...@gmail.com> wrote: > >> > >> Implement the insertion of available buffers in the descriptor area of > >> packed shadow virtqueues. It takes into account descriptor chains, but > >> does not consider indirect descriptors. > >> > >> Enable the packed SVQ to forward the descriptors to the device. > >> > >> Signed-off-by: Sahil Siddiq <sahil...@proton.me> > >> --- > >> Changes from v4 -> v5: > >> - This was commit #2 in v4. This has been reordered to commit #3 > >> based on review comments. > >> - vhost-shadow-virtqueue.c: > >> (vhost_svq_valid_features): Move addition of enums to commit #6 > >> based on review comments. > >> (vhost_svq_add_packed): Set head_idx to buffer id instead of vring's > >> index. > >> (vhost_svq_kick): Split into vhost_svq_kick_split and > >> vhost_svq_kick_packed. > >> (vhost_svq_add): Use new vhost_svq_kick_* functions. > >> > >> hw/virtio/vhost-shadow-virtqueue.c | 117 +++++++++++++++++++++++++++-- > >> 1 file changed, 112 insertions(+), 5 deletions(-) > >> > >> diff --git a/hw/virtio/vhost-shadow-virtqueue.c > >> b/hw/virtio/vhost-shadow-virtqueue.c > >> index 4f74ad402a..6e16cd4bdf 100644 > >> --- a/hw/virtio/vhost-shadow-virtqueue.c > >> +++ b/hw/virtio/vhost-shadow-virtqueue.c > >> @@ -193,10 +193,83 @@ static void vhost_svq_add_split(VhostShadowVirtqueue > >> *svq, > >> /* Update the avail index after write the descriptor */ > >> smp_wmb(); > >> avail->idx = cpu_to_le16(svq->shadow_avail_idx); > >> +} > >> + > >> +/** > >> + * Write descriptors to SVQ packed vring > >> + * > >> + * @svq: The shadow virtqueue > >> + * @out_sg: The iovec to the guest > >> + * @out_num: Outgoing iovec length > >> + * @in_sg: The iovec from the guest > >> + * @in_num: Incoming iovec length > >> + * @sgs: Cache for hwaddr > >> + * @head: Saves current free_head > >> + */ > >> +static void vhost_svq_add_packed(VhostShadowVirtqueue *svq, > >> + const struct iovec *out_sg, size_t > >> out_num, > >> + const struct iovec *in_sg, size_t in_num, > >> + hwaddr *sgs, unsigned *head) > >> +{ > >> + uint16_t id, curr, i, head_flags = 0, head_idx; > >> + size_t num = out_num + in_num; > >> + unsigned n; > >> + > >> + struct vring_packed_desc *descs = svq->vring_packed.vring.desc; > >> + > >> + head_idx = svq->vring_packed.next_avail_idx; > >> + i = head_idx; > >> + id = svq->free_head; > >> + curr = id; > >> + *head = id; > >> + > >> + /* Write descriptors to SVQ packed vring */ > >> + for (n = 0; n < num; n++) { > >> + uint16_t flags = cpu_to_le16(svq->vring_packed.avail_used_flags | > >> + (n < out_num ? 0 : > >> VRING_DESC_F_WRITE) | > >> + (n + 1 == num ? 0 : > >> VRING_DESC_F_NEXT)); > >> + if (i == head_idx) { > >> + head_flags = flags; > >> + } else { > >> + descs[i].flags = flags; > >> + } > >> + > >> + descs[i].addr = cpu_to_le64(sgs[n]); > >> + descs[i].id = id; > >> + if (n < out_num) { > >> + descs[i].len = cpu_to_le32(out_sg[n].iov_len); > >> + } else { > >> + descs[i].len = cpu_to_le32(in_sg[n - out_num].iov_len); > >> + } > >> + > >> + curr = cpu_to_le16(svq->desc_next[curr]); > >> + > >> + if (++i >= svq->vring_packed.vring.num) { > >> + i = 0; > >> + svq->vring_packed.avail_used_flags ^= > >> + 1 << VRING_PACKED_DESC_F_AVAIL | > >> + 1 << VRING_PACKED_DESC_F_USED; > >> + } > >> + } > >> > >> + if (i <= head_idx) { > >> + svq->vring_packed.avail_wrap_counter ^= 1; > >> + } > >> + > >> + svq->vring_packed.next_avail_idx = i; > >> + svq->shadow_avail_idx = i; > >> + svq->free_head = curr; > >> + > >> + /* > >> + * A driver MUST NOT make the first descriptor in the list > >> + * available before all subsequent descriptors comprising > >> + * the list are made available. > >> + */ > >> + smp_wmb(); > >> + svq->vring_packed.vring.desc[head_idx].flags = head_flags; > >> } > >> > >> -static void vhost_svq_kick(VhostShadowVirtqueue *svq) > >> +static void vhost_svq_kick_split(VhostShadowVirtqueue *svq) > >> { > >> bool needs_kick; > >> > >> @@ -209,7 +282,8 @@ static void vhost_svq_kick(VhostShadowVirtqueue *svq) > >> if (virtio_vdev_has_feature(svq->vdev, VIRTIO_RING_F_EVENT_IDX)) { > >> uint16_t avail_event = le16_to_cpu( > >> *(uint16_t *)(&svq->vring.used->ring[svq->vring.num])); > >> - needs_kick = vring_need_event(avail_event, svq->shadow_avail_idx, > >> svq->shadow_avail_idx - 1); > >> + needs_kick = vring_need_event(avail_event, svq->shadow_avail_idx, > >> + svq->shadow_avail_idx - 1); > >> } else { > >> needs_kick = > >> !(svq->vring.used->flags & > >> cpu_to_le16(VRING_USED_F_NO_NOTIFY)); > >> @@ -222,6 +296,30 @@ static void vhost_svq_kick(VhostShadowVirtqueue *svq) > >> event_notifier_set(&svq->hdev_kick); > >> } > >> > >> +static void vhost_svq_kick_packed(VhostShadowVirtqueue *svq) > >> +{ > >> + bool needs_kick; > >> + > >> + /* > >> + * We need to expose the available array entries before checking > >> + * notification suppressions. > >> + */ > >> + smp_mb(); > >> + > >> + if (virtio_vdev_has_feature(svq->vdev, VIRTIO_RING_F_EVENT_IDX)) { > >> + return; > > > > It's weird SVQ does not need to kick if _F_EVENT_IDX. This should have > > code checking the device ring flags etc. > > > > Right, I haven't implemented this yet. Since the current implementation is > being tested with event_idx=off (points 3 and 4 of the roadmap [1]), I thought > I would leave this for later. > > Maybe I can add a comment in the implementation explaining this. >
Sure that's fine, and probably even better than trying to address everything in one shot :) Can you add a TODO in each place you identify so we're sure we don't miss any? > Thanks, > Sahil > > [1] https://wiki.qemu.org/Internships/ProjectIdeas/PackedShadowVirtqueue >