Hi Eric,

On Mon, Mar 16, 2020 at 1:15 PM Bharat Bhushan <bharatb.li...@gmail.com> wrote:
>
> Hi Eric,
>
> On Mon, Mar 16, 2020 at 1:02 PM Auger Eric <eric.au...@redhat.com> wrote:
> >
> > Hi Bharat,
> >
> > On 3/16/20 7:41 AM, Bharat Bhushan wrote:
> > > Hi Eric,
> > >
> > > On Fri, Mar 13, 2020 at 8:11 PM Auger Eric <eric.au...@redhat.com> wrote:
> > >>
> > >> Hi Bharat
> > >>
> > >> On 3/13/20 8:48 AM, Bharat Bhushan wrote:
> > >>> iommu-notifier are called when a device is attached
> > >> IOMMU notifiers
> > >>> or detached to as address-space.
> > >>> This is needed for VFIO.
> > >> and vhost for detach
> > >>>
> > >>> Signed-off-by: Bharat Bhushan <bbhush...@marvell.com>
> > >>> ---
> > >>>  hw/virtio/virtio-iommu.c | 47 ++++++++++++++++++++++++++++++++++++++++
> > >>>  1 file changed, 47 insertions(+)
> > >>>
> > >>> diff --git a/hw/virtio/virtio-iommu.c b/hw/virtio/virtio-iommu.c
> > >>> index e51344a53e..2006f72901 100644
> > >>> --- a/hw/virtio/virtio-iommu.c
> > >>> +++ b/hw/virtio/virtio-iommu.c
> > >>> @@ -49,6 +49,7 @@ typedef struct VirtIOIOMMUEndpoint {
> > >>>      uint32_t id;
> > >>>      VirtIOIOMMUDomain *domain;
> > >>>      QLIST_ENTRY(VirtIOIOMMUEndpoint) next;
> > >>> +    VirtIOIOMMU *viommu;
> > >> This needs specal care on post-load. When migrating the EPs, only the id
> > >> is migrated. On post-load you need to set viommu as it is done for
> > >> domain. migration is allowed with vhost.
> > >
> > > ok, I have not tried vhost/migration. Below change set viommu when
> > > reconstructing endpoint.
> >
> >
> > Yes I think this should be OK.
> >
> > By the end I did the series a try with vhost/vfio. with vhost it works
> > (not with recent kernel though, but the issue may be related to kernel).
> > With VFIO however it does not for me.
> >
> > First issue is: your guest can use 4K page and your host can use 64KB
> > pages. In that case VFIO_DMA_MAP will fail with -EINVAL. We must devise
> > a way to pass the host settings to the VIRTIO-IOMMU device.
> >
> > Even with 64KB pages, it did not work for me. I have obviously not the
> > storm of VFIO_DMA_MAP failures but I have some, most probably due to
> > some wrong notifications somewhere. I will try to investigate on my side.
> >
> > Did you test with VFIO on your side?
>
> I did not tried with different page sizes, only tested with 4K page size.
>
> Yes it works, I tested with two n/w device assigned to VM, both interfaces 
> works
>
> First I will try with 64k page size.

64K page size does not work for me as well,

I think we are not passing correct page_size_mask here
(config.page_size_mask is set to TARGET_PAGE_MASK ( which is
0xfffffffffffff000))

We need to set this correctly as per host page size, correct?

Thanks
-Bharat

>
> Thanks
> -Bharat
>
> >
> > Thanks
> >
> > Eric
> > >
> > > @@ -984,6 +973,7 @@ static gboolean reconstruct_endpoints(gpointer
> > > key, gpointer value,
> > >
> > >      QLIST_FOREACH(iter, &d->endpoint_list, next) {
> > >          iter->domain = d;
> > > +       iter->viommu = s;
> > >          g_tree_insert(s->endpoints, GUINT_TO_POINTER(iter->id), iter);
> > >      }
> > >      return false; /* continue the domain traversal */
> > >
> > >>>  } VirtIOIOMMUEndpoint;
> > >>>
> > >>>  typedef struct VirtIOIOMMUInterval {
> > >>> @@ -155,8 +156,44 @@ static void 
> > >>> virtio_iommu_notify_unmap(IOMMUMemoryRegion *mr, hwaddr iova,
> > >>>      memory_region_notify_iommu(mr, 0, entry);
> > >>>  }
> > >>>
> > >>> +static gboolean virtio_iommu_mapping_unmap(gpointer key, gpointer 
> > >>> value,
> > >>> +                                           gpointer data)
> > >>> +{
> > >>> +    VirtIOIOMMUInterval *interval = (VirtIOIOMMUInterval *) key;
> > >>> +    IOMMUMemoryRegion *mr = (IOMMUMemoryRegion *) data;
> > >>> +
> > >>> +    virtio_iommu_notify_unmap(mr, interval->low,
> > >>> +                              interval->high - interval->low + 1);
> > >>> +
> > >>> +    return false;
> > >>> +}
> > >>> +
> > >>> +static gboolean virtio_iommu_mapping_map(gpointer key, gpointer value,
> > >>> +                                         gpointer data)
> > >>> +{
> > >>> +    VirtIOIOMMUMapping *mapping = (VirtIOIOMMUMapping *) value;
> > >>> +    VirtIOIOMMUInterval *interval = (VirtIOIOMMUInterval *) key;
> > >>> +    IOMMUMemoryRegion *mr = (IOMMUMemoryRegion *) data;
> > >>> +
> > >>> +    virtio_iommu_notify_map(mr, interval->low, mapping->phys_addr,
> > >>> +                            interval->high - interval->low + 1);
> > >>> +
> > >>> +    return false;
> > >>> +}
> > >>> +
> > >>>  static void 
> > >>> virtio_iommu_detach_endpoint_from_domain(VirtIOIOMMUEndpoint *ep)
> > >>>  {
> > >>> +    VirtioIOMMUNotifierNode *node;
> > >>> +    VirtIOIOMMU *s = ep->viommu;
> > >>> +    VirtIOIOMMUDomain *domain = ep->domain;
> > >>> +
> > >>> +    QLIST_FOREACH(node, &s->notifiers_list, next) {
> > >>> +        if (ep->id == node->iommu_dev->devfn) {
> > >>> +            g_tree_foreach(domain->mappings, 
> > >>> virtio_iommu_mapping_unmap,
> > >>> +                           &node->iommu_dev->iommu_mr);
> > >> I understand this should fo the job for domain removal
> > >
> > > did not get the comment, are you saying we should do this on domain 
> > > removal?
> > see my reply on 2/5
> >
> > Note the above code should be moved after the check of !ep->domain below
>
> ohh yes, will move
>
> Thanks
> -Bharat
>
> > >
> > >>> +        }
> > >>> +    }
> > >>> +
> > >>>      if (!ep->domain) {
> > >>>          return;
> > >>>      }
> > >>> @@ -178,6 +215,7 @@ static VirtIOIOMMUEndpoint 
> > >>> *virtio_iommu_get_endpoint(VirtIOIOMMU *s,
> > >>>      }
> > >>>      ep = g_malloc0(sizeof(*ep));
> > >>>      ep->id = ep_id;
> > >>> +    ep->viommu = s;
> > >>>      trace_virtio_iommu_get_endpoint(ep_id);
> > >>>      g_tree_insert(s->endpoints, GUINT_TO_POINTER(ep_id), ep);
> > >>>      return ep;
> > >>> @@ -272,6 +310,7 @@ static int virtio_iommu_attach(VirtIOIOMMU *s,
> > >>>  {
> > >>>      uint32_t domain_id = le32_to_cpu(req->domain);
> > >>>      uint32_t ep_id = le32_to_cpu(req->endpoint);
> > >>> +    VirtioIOMMUNotifierNode *node;
> > >>>      VirtIOIOMMUDomain *domain;
> > >>>      VirtIOIOMMUEndpoint *ep;
> > >>>
> > >>> @@ -299,6 +338,14 @@ static int virtio_iommu_attach(VirtIOIOMMU *s,
> > >>>
> > >>>      ep->domain = domain;
> > >>>
> > >>> +    /* Replay existing address space mappings on the associated memory 
> > >>> region */
> > >> maybe use the "domain" terminology here.
> > >
> > > ok,
> > >
> > > Thanks
> > > -Bharat
> > >
> > >>> +    QLIST_FOREACH(node, &s->notifiers_list, next) {
> > >>> +        if (ep_id == node->iommu_dev->devfn) {
> > >>> +            g_tree_foreach(domain->mappings, virtio_iommu_mapping_map,
> > >>> +                           &node->iommu_dev->iommu_mr);
> > >>> +        }
> > >>> +    }
> > >>> +
> > >>>      return VIRTIO_IOMMU_S_OK;
> > >>>  }
> > >>>
> > >>>
> > >> Thanks
> > >>
> > >> Eric
> > >>
> > >
> >

Reply via email to