Then we can support mask bit operation of assigned devices now.

Signed-off-by: Sheng Yang <sh...@linux.intel.com>
---
 arch/x86/include/asm/kvm_host.h |    1 +
 arch/x86/kvm/Makefile           |    2 +-
 arch/x86/kvm/mmu.c              |    2 +
 arch/x86/kvm/x86.c              |   40 ++++-
 include/linux/kvm.h             |   28 ++++
 include/linux/kvm_host.h        |   34 +++++
 virt/kvm/assigned-dev.c         |   44 ++++++
 virt/kvm/kvm_main.c             |   38 +++++-
 virt/kvm/msix_mmio.c            |  302 +++++++++++++++++++++++++++++++++++++++
 virt/kvm/msix_mmio.h            |   25 ++++
 10 files changed, 503 insertions(+), 13 deletions(-)
 create mode 100644 virt/kvm/msix_mmio.c
 create mode 100644 virt/kvm/msix_mmio.h

diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h
index aa75f21..4a390a4 100644
--- a/arch/x86/include/asm/kvm_host.h
+++ b/arch/x86/include/asm/kvm_host.h
@@ -635,6 +635,7 @@ enum emulation_result {
        EMULATE_DONE,       /* no further processing */
        EMULATE_DO_MMIO,      /* kvm_run filled with mmio request */
        EMULATE_FAIL,         /* can't emulate this instruction */
+       EMULATE_USERSPACE_EXIT, /* we need exit to userspace */
 };
 
 #define EMULTYPE_NO_DECODE         (1 << 0)
diff --git a/arch/x86/kvm/Makefile b/arch/x86/kvm/Makefile
index f15501f..3a0d851 100644
--- a/arch/x86/kvm/Makefile
+++ b/arch/x86/kvm/Makefile
@@ -7,7 +7,7 @@ CFLAGS_vmx.o := -I.
 
 kvm-y                  += $(addprefix ../../../virt/kvm/, kvm_main.o ioapic.o \
                                coalesced_mmio.o irq_comm.o eventfd.o \
-                               assigned-dev.o)
+                               assigned-dev.o msix_mmio.o)
 kvm-$(CONFIG_IOMMU_API)        += $(addprefix ../../../virt/kvm/, iommu.o)
 kvm-$(CONFIG_KVM_ASYNC_PF)     += $(addprefix ../../../virt/kvm/, async_pf.o)
 
diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c
index 9cafbb4..912dca4 100644
--- a/arch/x86/kvm/mmu.c
+++ b/arch/x86/kvm/mmu.c
@@ -3358,6 +3358,8 @@ int kvm_mmu_page_fault(struct kvm_vcpu *vcpu, gva_t cr2, 
u32 error_code,
        case EMULATE_DO_MMIO:
                ++vcpu->stat.mmio_exits;
                /* fall through */
+       case EMULATE_USERSPACE_EXIT:
+               /* fall through */
        case EMULATE_FAIL:
                return 0;
        default:
diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
index 21b84e2..87308eb 100644
--- a/arch/x86/kvm/x86.c
+++ b/arch/x86/kvm/x86.c
@@ -1966,6 +1966,7 @@ int kvm_dev_ioctl_check_extension(long ext)
        case KVM_CAP_X86_ROBUST_SINGLESTEP:
        case KVM_CAP_XSAVE:
        case KVM_CAP_ASYNC_PF:
+       case KVM_CAP_MSIX_MMIO:
                r = 1;
                break;
        case KVM_CAP_COALESCED_MMIO:
@@ -3809,6 +3810,7 @@ static int emulator_write_emulated_onepage(unsigned long 
addr,
 {
        gpa_t                 gpa;
        struct kvm_io_ext_data ext_data;
+       int r;
 
        gpa = kvm_mmu_gva_to_gpa_write(vcpu, addr, exception);
 
@@ -3824,18 +3826,32 @@ static int emulator_write_emulated_onepage(unsigned 
long addr,
 
 mmio:
        trace_kvm_mmio(KVM_TRACE_MMIO_WRITE, bytes, gpa, *(u64 *)val);
+       r = vcpu_mmio_write(vcpu, gpa, bytes, val, &ext_data);
        /*
         * Is this MMIO handled locally?
         */
-       if (!vcpu_mmio_write(vcpu, gpa, bytes, val, &ext_data))
+       if (!r)
                return X86EMUL_CONTINUE;
 
-       vcpu->mmio_needed = 1;
-       vcpu->run->exit_reason = KVM_EXIT_MMIO;
-       vcpu->run->mmio.phys_addr = vcpu->mmio_phys_addr = gpa;
-       vcpu->run->mmio.len = vcpu->mmio_size = bytes;
-       vcpu->run->mmio.is_write = vcpu->mmio_is_write = 1;
-       memcpy(vcpu->run->mmio.data, val, bytes);
+       if (r == -ENOTSYNC) {
+               vcpu->userspace_exit_needed = 1;
+               vcpu->run->exit_reason = KVM_EXIT_MSIX_ROUTING_UPDATE;
+               vcpu->run->msix_routing.dev_id =
+                       ext_data.msix_routing.dev_id;
+               vcpu->run->msix_routing.type =
+                       ext_data.msix_routing.type;
+               vcpu->run->msix_routing.entry_idx =
+                       ext_data.msix_routing.entry_idx;
+               vcpu->run->msix_routing.flags =
+                       ext_data.msix_routing.flags;
+       } else  {
+               vcpu->mmio_needed = 1;
+               vcpu->run->exit_reason = KVM_EXIT_MMIO;
+               vcpu->run->mmio.phys_addr = vcpu->mmio_phys_addr = gpa;
+               vcpu->run->mmio.len = vcpu->mmio_size = bytes;
+               vcpu->run->mmio.is_write = vcpu->mmio_is_write = 1;
+               memcpy(vcpu->run->mmio.data, val, bytes);
+       }
 
        return X86EMUL_CONTINUE;
 }
@@ -4469,6 +4485,8 @@ done:
                r = EMULATE_DO_MMIO;
        } else if (r == EMULATION_RESTART)
                goto restart;
+       else if (vcpu->userspace_exit_needed)
+               r = EMULATE_USERSPACE_EXIT;
        else
                r = EMULATE_DONE;
 
@@ -5397,12 +5415,18 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, 
struct kvm_run *kvm_run)
                }
        }
 
-       if (vcpu->arch.pio.count || vcpu->mmio_needed) {
+       if (vcpu->arch.pio.count || vcpu->mmio_needed ||
+                       vcpu->userspace_exit_needed) {
                if (vcpu->mmio_needed) {
                        memcpy(vcpu->mmio_data, kvm_run->mmio.data, 8);
                        vcpu->mmio_read_completed = 1;
                        vcpu->mmio_needed = 0;
                }
+               if (vcpu->userspace_exit_needed) {
+                       vcpu->userspace_exit_needed = 0;
+                       r = 0;
+                       goto out;
+               }
                vcpu->srcu_idx = srcu_read_lock(&vcpu->kvm->srcu);
                r = emulate_instruction(vcpu, EMULTYPE_NO_DECODE);
                srcu_read_unlock(&vcpu->kvm->srcu, vcpu->srcu_idx);
diff --git a/include/linux/kvm.h b/include/linux/kvm.h
index ea2dc1a..4393e4e 100644
--- a/include/linux/kvm.h
+++ b/include/linux/kvm.h
@@ -161,6 +161,7 @@ struct kvm_pit_config {
 #define KVM_EXIT_NMI              16
 #define KVM_EXIT_INTERNAL_ERROR   17
 #define KVM_EXIT_OSI              18
+#define KVM_EXIT_MSIX_ROUTING_UPDATE 19
 
 /* For KVM_EXIT_INTERNAL_ERROR */
 #define KVM_INTERNAL_ERROR_EMULATION 1
@@ -264,6 +265,13 @@ struct kvm_run {
                struct {
                        __u64 gprs[32];
                } osi;
+               /* KVM_EXIT_MSIX_ROUTING_UPDATE*/
+               struct {
+                       __u32 dev_id;
+                       __u16 type;
+                       __u16 entry_idx;
+                       __u64 flags;
+               } msix_routing;
                /* Fix the size of the union. */
                char padding[256];
        };
@@ -541,6 +549,7 @@ struct kvm_ppc_pvinfo {
 #define KVM_CAP_PPC_GET_PVINFO 57
 #define KVM_CAP_PPC_IRQ_LEVEL 58
 #define KVM_CAP_ASYNC_PF 59
+#define KVM_CAP_MSIX_MMIO 60
 
 #ifdef KVM_CAP_IRQ_ROUTING
 
@@ -672,6 +681,9 @@ struct kvm_clock_data {
 #define KVM_XEN_HVM_CONFIG        _IOW(KVMIO,  0x7a, struct kvm_xen_hvm_config)
 #define KVM_SET_CLOCK             _IOW(KVMIO,  0x7b, struct kvm_clock_data)
 #define KVM_GET_CLOCK             _IOR(KVMIO,  0x7c, struct kvm_clock_data)
+/* Available with KVM_CAP_MSIX_MMIO */
+#define KVM_REGISTER_MSIX_MMIO    _IOW(KVMIO,  0x7d, struct kvm_msix_mmio_user)
+#define KVM_UNREGISTER_MSIX_MMIO  _IOW(KVMIO,  0x7e, struct kvm_msix_mmio_user)
 /* Available with KVM_CAP_PIT_STATE2 */
 #define KVM_GET_PIT2              _IOR(KVMIO,  0x9f, struct kvm_pit_state2)
 #define KVM_SET_PIT2              _IOW(KVMIO,  0xa0, struct kvm_pit_state2)
@@ -795,4 +807,20 @@ struct kvm_assigned_msix_entry {
        __u16 padding[3];
 };
 
+#define KVM_MSIX_MMIO_TYPE_ASSIGNED_DEV            (1 << 0)
+
+#define KVM_MSIX_MMIO_TYPE_BASE_TABLE      (1 << 8)
+
+#define KVM_MSIX_MMIO_TYPE_DEV_MASK        0x00ff
+#define KVM_MSIX_MMIO_TYPE_BASE_MASK       0xff00
+struct kvm_msix_mmio_user {
+       __u32 dev_id;
+       __u16 type;
+       __u16 max_entries_nr;
+       __u64 base_addr;
+       __u64 base_va;
+       __u64 flags;
+       __u64 reserved[4];
+};
+
 #endif /* __LINUX_KVM_H */
diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h
index 6bb211d..6aaf85e 100644
--- a/include/linux/kvm_host.h
+++ b/include/linux/kvm_host.h
@@ -68,9 +68,16 @@ enum kvm_bus {
        KVM_NR_BUSES
 };
 
+#define KVM_IO_EXT_DATA_TYPE_MSIX_ROUTING   1
 struct kvm_io_ext_data {
        int type;
        union {
+               struct {
+                       u32 dev_id;
+                       u16 type;
+                       u16 entry_idx;
+                       u64 flags;
+               } msix_routing;
                char padding[256];
        };
 };
@@ -168,6 +175,8 @@ struct kvm_vcpu {
        } async_pf;
 #endif
 
+       int userspace_exit_needed;
+
        struct kvm_vcpu_arch arch;
 };
 
@@ -241,6 +250,27 @@ struct kvm_memslots {
                                        KVM_PRIVATE_MEM_SLOTS];
 };
 
+#define KVM_MSIX_MMIO_MAX    32
+
+struct kvm_msix_mmio {
+       u32 dev_id;
+       u16 type;
+       u16 max_entries_nr;
+       u64 flags;
+       gpa_t table_base_addr;
+       hva_t table_base_va;
+       gpa_t pba_base_addr;
+       hva_t pba_base_va;
+};
+
+struct kvm_msix_mmio_dev {
+       struct kvm *kvm;
+       struct kvm_io_device table_dev;
+       int mmio_nr;
+       struct kvm_msix_mmio mmio[KVM_MSIX_MMIO_MAX];
+       struct mutex lock;
+};
+
 struct kvm {
        spinlock_t mmu_lock;
        raw_spinlock_t requests_lock;
@@ -289,6 +319,7 @@ struct kvm {
        long mmu_notifier_count;
 #endif
        long tlbs_dirty;
+       struct kvm_msix_mmio_dev msix_mmio_dev;
 };
 
 /* The guest did something we don't support. */
@@ -561,6 +592,9 @@ void kvm_unregister_irq_ack_notifier(struct kvm *kvm,
 int kvm_request_irq_source_id(struct kvm *kvm);
 void kvm_free_irq_source_id(struct kvm *kvm, int irq_source_id);
 
+int kvm_assigned_device_update_msix_mask_bit(struct kvm *kvm,
+                       int assigned_dev_id, int entry, bool mask);
+
 /* For vcpu->arch.iommu_flags */
 #define KVM_IOMMU_CACHE_COHERENCY      0x1
 
diff --git a/virt/kvm/assigned-dev.c b/virt/kvm/assigned-dev.c
index ae72ae6..d1598a6 100644
--- a/virt/kvm/assigned-dev.c
+++ b/virt/kvm/assigned-dev.c
@@ -18,6 +18,7 @@
 #include <linux/interrupt.h>
 #include <linux/slab.h>
 #include "irq.h"
+#include "msix_mmio.h"
 
 static struct kvm_assigned_dev_kernel *kvm_find_assigned_dev(struct list_head 
*head,
                                                      int assigned_dev_id)
@@ -191,12 +192,25 @@ static void kvm_free_assigned_irq(struct kvm *kvm,
        kvm_deassign_irq(kvm, assigned_dev, assigned_dev->irq_requested_type);
 }
 
+static void assigned_device_free_msix_mmio(struct kvm *kvm,
+                               struct kvm_assigned_dev_kernel *adev)
+{
+       struct kvm_msix_mmio mmio;
+
+       mmio.dev_id = adev->assigned_dev_id;
+       mmio.type = KVM_MSIX_MMIO_TYPE_ASSIGNED_DEV |
+                   KVM_MSIX_MMIO_TYPE_BASE_TABLE;
+       kvm_free_msix_mmio(kvm, &mmio);
+}
+
 static void kvm_free_assigned_device(struct kvm *kvm,
                                     struct kvm_assigned_dev_kernel
                                     *assigned_dev)
 {
        kvm_free_assigned_irq(kvm, assigned_dev);
 
+       assigned_device_free_msix_mmio(kvm, assigned_dev);
+
        __pci_reset_function(assigned_dev->dev);
        pci_restore_state(assigned_dev->dev);
 
@@ -785,3 +799,33 @@ out:
        return r;
 }
 
+/* The caller should hold kvm->lock */
+int kvm_assigned_device_update_msix_mask_bit(struct kvm *kvm,
+                               int assigned_dev_id, int entry, bool mask)
+{
+       int r = -EFAULT;
+       struct kvm_assigned_dev_kernel *adev;
+       int i;
+
+       if (!irqchip_in_kernel(kvm))
+               return r;
+
+       adev = kvm_find_assigned_dev(&kvm->arch.assigned_dev_head,
+                                     assigned_dev_id);
+       if (!adev)
+               goto out;
+
+       /* For non-MSIX enabled devices, entries_nr == 0 */
+       for (i = 0; i < adev->entries_nr; i++)
+               if (adev->host_msix_entries[i].entry == entry) {
+                       if (mask)
+                               disable_irq_nosync(
+                                       adev->host_msix_entries[i].vector);
+                       else
+                               enable_irq(adev->host_msix_entries[i].vector);
+                       r = 0;
+                       break;
+               }
+out:
+       return r;
+}
diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c
index a61f90e..f211e49 100644
--- a/virt/kvm/kvm_main.c
+++ b/virt/kvm/kvm_main.c
@@ -56,6 +56,7 @@
 
 #include "coalesced_mmio.h"
 #include "async_pf.h"
+#include "msix_mmio.h"
 
 #define CREATE_TRACE_POINTS
 #include <trace/events/kvm.h>
@@ -509,6 +510,7 @@ static void kvm_destroy_vm(struct kvm *kvm)
        struct mm_struct *mm = kvm->mm;
 
        kvm_arch_sync_events(kvm);
+       kvm_unregister_msix_mmio_dev(kvm);
        spin_lock(&kvm_lock);
        list_del(&kvm->vm_list);
        spin_unlock(&kvm_lock);
@@ -1877,6 +1879,24 @@ static long kvm_vm_ioctl(struct file *filp,
                mutex_unlock(&kvm->lock);
                break;
 #endif
+       case KVM_REGISTER_MSIX_MMIO: {
+               struct kvm_msix_mmio_user mmio_user;
+
+               r = -EFAULT;
+               if (copy_from_user(&mmio_user, argp, sizeof mmio_user))
+                       goto out;
+               r = kvm_vm_ioctl_register_msix_mmio(kvm, &mmio_user);
+               break;
+       }
+       case KVM_UNREGISTER_MSIX_MMIO: {
+               struct kvm_msix_mmio_user mmio_user;
+
+               r = -EFAULT;
+               if (copy_from_user(&mmio_user, argp, sizeof mmio_user))
+                       goto out;
+               r = kvm_vm_ioctl_unregister_msix_mmio(kvm, &mmio_user);
+               break;
+       }
        default:
                r = kvm_arch_vm_ioctl(filp, ioctl, arg);
                if (r == -ENOTTY)
@@ -1988,6 +2008,12 @@ static int kvm_dev_ioctl_create_vm(void)
                return r;
        }
 #endif
+       r = kvm_register_msix_mmio_dev(kvm);
+       if (r < 0) {
+               kvm_put_kvm(kvm);
+               return r;
+       }
+
        r = anon_inode_getfd("kvm-vm", &kvm_vm_fops, kvm, O_RDWR);
        if (r < 0)
                kvm_put_kvm(kvm);
@@ -2223,14 +2249,18 @@ static void kvm_io_bus_destroy(struct kvm_io_bus *bus)
 int kvm_io_bus_write(struct kvm *kvm, enum kvm_bus bus_idx, gpa_t addr,
                     int len, const void *val, struct kvm_io_ext_data *ext_data)
 {
-       int i;
+       int i, r = -EOPNOTSUPP;
        struct kvm_io_bus *bus;
 
        bus = srcu_dereference(kvm->buses[bus_idx], &kvm->srcu);
-       for (i = 0; i < bus->dev_count; i++)
-               if (!kvm_iodevice_write(bus->devs[i], addr, len, val, ext_data))
+       for (i = 0; i < bus->dev_count; i++) {
+               r = kvm_iodevice_write(bus->devs[i], addr, len, val, ext_data);
+               if (r == -ENOTSYNC)
+                       break;
+               else if (!r)
                        return 0;
-       return -EOPNOTSUPP;
+       }
+       return r;
 }
 
 /* kvm_io_bus_read - called under kvm->slots_lock */
diff --git a/virt/kvm/msix_mmio.c b/virt/kvm/msix_mmio.c
new file mode 100644
index 0000000..4a9cf15
--- /dev/null
+++ b/virt/kvm/msix_mmio.c
@@ -0,0 +1,302 @@
+/*
+ * MSI-X MMIO emulation
+ *
+ * Copyright (c) 2010 Intel Corporation
+ *
+ * This work is licensed under the terms of the GNU GPL, version 2.  See
+ * the COPYING file in the top-level directory.
+ *
+ * Author:
+ *   Sheng Yang <sheng.y...@intel.com>
+ */
+
+#include <linux/kvm_host.h>
+#include <linux/kvm.h>
+
+#include "msix_mmio.h"
+#include "iodev.h"
+
+static int update_msix_mask_bit(struct kvm *kvm, struct kvm_msix_mmio *mmio,
+                               int entry, u32 flag)
+{
+       if (mmio->type & KVM_MSIX_MMIO_TYPE_ASSIGNED_DEV)
+               return kvm_assigned_device_update_msix_mask_bit(kvm,
+                               mmio->dev_id, entry, flag);
+       return -EFAULT;
+}
+
+/* Caller must hold dev->lock */
+static int get_mmio_table_index(struct kvm_msix_mmio_dev *dev,
+                               gpa_t addr, int len)
+{
+       gpa_t start, end;
+       int i, r = -EINVAL;
+
+       for (i = 0; i < dev->mmio_nr; i++) {
+               start = dev->mmio[i].table_base_addr;
+               end = dev->mmio[i].table_base_addr + PCI_MSIX_ENTRY_SIZE *
+                       dev->mmio[i].max_entries_nr;
+               if (addr >= start && addr + len <= end) {
+                       r = i;
+                       break;
+               }
+       }
+
+       return r;
+}
+
+static int msix_table_mmio_read(struct kvm_io_device *this, gpa_t addr, int 
len,
+                               void *val)
+{
+       /*TODO: Add big endian support */
+       struct kvm_msix_mmio_dev *mmio_dev =
+               container_of(this, struct kvm_msix_mmio_dev, table_dev);
+       struct kvm_msix_mmio *mmio;
+       int idx, ret = 0, entry, offset, r;
+
+       mutex_lock(&mmio_dev->lock);
+       idx = get_mmio_table_index(mmio_dev, addr, len);
+       if (idx < 0) {
+               ret = -EOPNOTSUPP;
+               goto out;
+       }
+       if (!(len == 4 || len == 8) || addr & (len - 1))
+               goto out;
+
+       offset = addr % PCI_MSIX_ENTRY_SIZE;
+       mmio = &mmio_dev->mmio[idx];
+       entry = (addr - mmio->table_base_addr) / PCI_MSIX_ENTRY_SIZE;
+       r = copy_from_user(val, (void __user *)(mmio->table_base_va +
+                       entry * PCI_MSIX_ENTRY_SIZE + offset), len);
+       if (r)
+               goto out;
+out:
+       mutex_unlock(&mmio_dev->lock);
+       return ret;
+}
+
+static int msix_table_mmio_write(struct kvm_io_device *this, gpa_t addr,
+                               int len, const void *val,
+                               struct kvm_io_ext_data *ext_data)
+{
+       /*TODO: Add big endian support */
+       struct kvm_msix_mmio_dev *mmio_dev =
+               container_of(this, struct kvm_msix_mmio_dev, table_dev);
+       struct kvm_msix_mmio *mmio;
+       int idx, entry, offset, ret = 0, r = 0;
+       void __user *entry_base;
+       u32 __user *ctrl_pos;
+       u32 old_ctrl, new_ctrl;
+
+       mutex_lock(&mmio_dev->kvm->lock);
+       mutex_lock(&mmio_dev->lock);
+       idx = get_mmio_table_index(mmio_dev, addr, len);
+       if (idx < 0) {
+               ret = -EOPNOTSUPP;
+               goto out;
+       }
+       if (!(len == 4 || len == 8) || addr & (len - 1))
+               goto out;
+
+       offset = addr % PCI_MSIX_ENTRY_SIZE;
+
+       mmio = &mmio_dev->mmio[idx];
+       entry = (addr - mmio->table_base_addr) / PCI_MSIX_ENTRY_SIZE;
+       entry_base = (void __user *)(mmio->table_base_va +
+                       entry * PCI_MSIX_ENTRY_SIZE);
+       ctrl_pos = entry_base +
+                       PCI_MSIX_ENTRY_VECTOR_CTRL;
+
+       if (get_user(old_ctrl, ctrl_pos))
+               goto out;
+
+       /* Don't allow writing to other fields when entry is unmasked */
+       if (!(old_ctrl & PCI_MSIX_ENTRY_CTRL_MASKBIT) &&
+           offset != PCI_MSIX_ENTRY_VECTOR_CTRL)
+               goto out;
+
+       if (copy_to_user((void __user *)(entry_base + offset), val, len))
+               goto out;
+
+       ext_data->type = KVM_IO_EXT_DATA_TYPE_MSIX_ROUTING;
+       ext_data->msix_routing.dev_id = mmio->dev_id;
+       ext_data->msix_routing.type = mmio->type;
+       ext_data->msix_routing.entry_idx = entry;
+       ext_data->msix_routing.flags = 0;
+
+       if (offset + len < PCI_MSIX_ENTRY_VECTOR_CTRL) {
+               ret = -ENOTSYNC;
+               goto out;
+       }
+
+       if (get_user(new_ctrl, ctrl_pos))
+               goto out;
+
+       if (old_ctrl == new_ctrl) {
+               if (offset == PCI_MSIX_ENTRY_DATA && len == 8)
+                       ret = -ENOTSYNC;
+               goto out;
+       }
+       if ((old_ctrl & PCI_MSIX_ENTRY_CTRL_MASKBIT) ^
+                       (new_ctrl & PCI_MSIX_ENTRY_CTRL_MASKBIT))
+               r = update_msix_mask_bit(mmio_dev->kvm, mmio, entry,
+                               !!(new_ctrl & PCI_MSIX_ENTRY_CTRL_MASKBIT));
+       if (r)
+               ret = -ENOTSYNC;
+out:
+       mutex_unlock(&mmio_dev->lock);
+       mutex_unlock(&mmio_dev->kvm->lock);
+       return ret;
+}
+
+static const struct kvm_io_device_ops msix_mmio_table_ops = {
+       .read     = msix_table_mmio_read,
+       .write    = msix_table_mmio_write,
+};
+
+int kvm_register_msix_mmio_dev(struct kvm *kvm)
+{
+       int ret;
+
+       kvm_iodevice_init(&kvm->msix_mmio_dev.table_dev, &msix_mmio_table_ops);
+       mutex_init(&kvm->msix_mmio_dev.lock);
+       kvm->msix_mmio_dev.kvm = kvm;
+       mutex_lock(&kvm->slots_lock);
+       ret = kvm_io_bus_register_dev(kvm, KVM_MMIO_BUS,
+                                     &kvm->msix_mmio_dev.table_dev);
+       mutex_unlock(&kvm->slots_lock);
+       return ret;
+}
+
+int kvm_unregister_msix_mmio_dev(struct kvm *kvm)
+{
+       int ret;
+
+       mutex_lock(&kvm->slots_lock);
+       ret = kvm_io_bus_unregister_dev(kvm, KVM_MMIO_BUS,
+                                     &kvm->msix_mmio_dev.table_dev);
+       mutex_unlock(&kvm->slots_lock);
+       return ret;
+}
+
+int kvm_vm_ioctl_register_msix_mmio(struct kvm *kvm,
+                                   struct kvm_msix_mmio_user *mmio_user)
+{
+       struct kvm_msix_mmio_dev *mmio_dev = &kvm->msix_mmio_dev;
+       struct kvm_msix_mmio *mmio = NULL;
+       int r = 0, i;
+
+       mutex_lock(&mmio_dev->lock);
+       for (i = 0; i < mmio_dev->mmio_nr; i++) {
+               if (mmio_dev->mmio[i].dev_id == mmio_user->dev_id &&
+                   (mmio_dev->mmio[i].type & KVM_MSIX_MMIO_TYPE_DEV_MASK) ==
+                   (mmio_user->type & KVM_MSIX_MMIO_TYPE_DEV_MASK)) {
+                       mmio = &mmio_dev->mmio[i];
+                       if (mmio->max_entries_nr != mmio_user->max_entries_nr) {
+                               r = -EINVAL;
+                               goto out;
+                       }
+                       break;
+               }
+       }
+       if (mmio_user->max_entries_nr > KVM_MAX_MSIX_PER_DEV) {
+               r = -EINVAL;
+               goto out;
+       }
+       /* All reserved currently */
+       if (mmio_user->flags) {
+               r = -EINVAL;
+               goto out;
+       }
+
+       if ((mmio_user->type & KVM_MSIX_MMIO_TYPE_DEV_MASK) !=
+                       KVM_MSIX_MMIO_TYPE_ASSIGNED_DEV) {
+               r = -EINVAL;
+               goto out;
+       }
+       if ((mmio_user->type & KVM_MSIX_MMIO_TYPE_BASE_MASK) !=
+                       KVM_MSIX_MMIO_TYPE_BASE_TABLE) {
+               r = -EINVAL;
+               goto out;
+       }
+
+#ifndef CONFIG_64BIT
+       if (mmio_user->base_va >= 0xffffffff ||
+           mmio_user->base_addr >= 0xffffffff) {
+               r = -EINVAL;
+               goto out;
+       }
+#endif
+
+       /* Check alignment and accessibility */
+       if ((mmio_user->base_va % PCI_MSIX_ENTRY_SIZE) ||
+           !access_ok(VERIFY_WRITE, (void __user *)mmio_user->base_va,
+                       mmio_user->max_entries_nr * PCI_MSIX_ENTRY_SIZE)) {
+               r = -EINVAL;
+               goto out;
+       }
+       if (!mmio) {
+               if (mmio_dev->mmio_nr == KVM_MSIX_MMIO_MAX) {
+                       r = -ENOSPC;
+                       goto out;
+               }
+               mmio = &mmio_dev->mmio[mmio_dev->mmio_nr];
+               mmio_dev->mmio_nr++;
+       }
+
+       mmio->max_entries_nr = mmio_user->max_entries_nr;
+       mmio->dev_id = mmio_user->dev_id;
+       mmio->flags = mmio_user->flags;
+
+       if ((mmio_user->type & KVM_MSIX_MMIO_TYPE_DEV_MASK) ==
+                       KVM_MSIX_MMIO_TYPE_ASSIGNED_DEV)
+               mmio->type = KVM_MSIX_MMIO_TYPE_ASSIGNED_DEV;
+       if ((mmio_user->type & KVM_MSIX_MMIO_TYPE_BASE_MASK) ==
+                       KVM_MSIX_MMIO_TYPE_BASE_TABLE) {
+               mmio->type |= KVM_MSIX_MMIO_TYPE_BASE_TABLE;
+               mmio->table_base_addr = mmio_user->base_addr;
+               mmio->table_base_va = mmio_user->base_va;
+       }
+out:
+       mutex_unlock(&mmio_dev->lock);
+       return r;
+}
+
+int kvm_free_msix_mmio(struct kvm *kvm, struct kvm_msix_mmio *mmio)
+{
+       struct kvm_msix_mmio_dev *mmio_dev = &kvm->msix_mmio_dev;
+       int r = -EINVAL, i, j;
+
+       if (!mmio)
+               return 0;
+
+       mutex_lock(&mmio_dev->lock);
+       BUG_ON(mmio_dev->mmio_nr > KVM_MSIX_MMIO_MAX);
+       for (i = 0; i < mmio_dev->mmio_nr; i++) {
+               if (mmio_dev->mmio[i].dev_id == mmio->dev_id &&
+                   mmio_dev->mmio[i].type == mmio->type) {
+                       r = 0;
+                       for (j = i; j < mmio_dev->mmio_nr - 1; j++)
+                               mmio_dev->mmio[j] = mmio_dev->mmio[j + 1];
+                       mmio_dev->mmio[mmio_dev->mmio_nr].max_entries_nr = 0;
+                       mmio_dev->mmio[mmio_dev->mmio_nr].dev_id = 0;
+                       mmio_dev->mmio[mmio_dev->mmio_nr].type = 0;
+                       mmio_dev->mmio_nr--;
+                       break;
+               }
+       }
+       mutex_unlock(&mmio_dev->lock);
+       return r;
+}
+
+int kvm_vm_ioctl_unregister_msix_mmio(struct kvm *kvm,
+                                     struct kvm_msix_mmio_user *mmio_user)
+{
+       struct kvm_msix_mmio mmio;
+
+       mmio.dev_id = mmio_user->dev_id;
+       mmio.type = mmio_user->type;
+
+       return kvm_free_msix_mmio(kvm, &mmio);
+}
+
diff --git a/virt/kvm/msix_mmio.h b/virt/kvm/msix_mmio.h
new file mode 100644
index 0000000..01b6587
--- /dev/null
+++ b/virt/kvm/msix_mmio.h
@@ -0,0 +1,25 @@
+#ifndef __KVM_MSIX_MMIO_H__
+#define __KVM_MSIX_MMIO_H__
+/*
+ * MSI-X MMIO emulation
+ *
+ * Copyright (c) 2010 Intel Corporation
+ *
+ * This work is licensed under the terms of the GNU GPL, version 2.  See
+ * the COPYING file in the top-level directory.
+ *
+ * Author:
+ *   Sheng Yang <sheng.y...@intel.com>
+ */
+
+#include <linux/pci.h>
+
+int kvm_register_msix_mmio_dev(struct kvm *kvm);
+int kvm_unregister_msix_mmio_dev(struct kvm *kvm);
+int kvm_vm_ioctl_register_msix_mmio(struct kvm *kvm,
+                                   struct kvm_msix_mmio_user *mmio_user);
+int kvm_vm_ioctl_unregister_msix_mmio(struct kvm *kvm,
+                                     struct kvm_msix_mmio_user *mmio_user);
+int kvm_free_msix_mmio(struct kvm *kvm, struct kvm_msix_mmio *mmio_user);
+
+#endif
-- 
1.7.0.1

--
To unsubscribe from this list: send the line "unsubscribe kvm" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to