Marcelo Tosatti wrote: > Switch irq injection/acking to irq_lock, and change PIO/MMIO paths > so that the device search is protected by kvm->lock, but not the > read/write callbacks (which is responsability of the device). > > Fix for http://article.gmane.org/gmane.comp.emulators.kvm.devel/32286. > > Signed-off-by: Marcelo Tosatti <[email protected]> > > Index: kvm/arch/x86/kvm/i8254.c > =================================================================== > --- kvm.orig/arch/x86/kvm/i8254.c > +++ kvm/arch/x86/kvm/i8254.c > @@ -634,10 +634,10 @@ static void __inject_pit_timer_intr(stru > struct kvm_vcpu *vcpu; > int i; > > - mutex_lock(&kvm->lock); > + mutex_lock(&kvm->irq_lock); >
There would be advantages to having irq_lock be
interrupt/nonpreempt-friendly design, such as s/mutex/spinlock. For
instance, irqfd could inject the interrupt directly without deferring to
a workqueue. I'm not sure if this is possible/easy, but its something
to consider.
-Greg
> kvm_set_irq(kvm, kvm->arch.vpit->irq_source_id, 0, 1);
> kvm_set_irq(kvm, kvm->arch.vpit->irq_source_id, 0, 0);
> - mutex_unlock(&kvm->lock);
> + mutex_unlock(&kvm->irq_lock);
>
> /*
> * Provides NMI watchdog support via Virtual Wire mode.
> Index: kvm/arch/x86/kvm/x86.c
> ===================================================================
> --- kvm.orig/arch/x86/kvm/x86.c
> +++ kvm/arch/x86/kvm/x86.c
> @@ -1909,10 +1909,10 @@ long kvm_arch_vm_ioctl(struct file *filp
> goto out;
> if (irqchip_in_kernel(kvm)) {
> __s32 status;
> - mutex_lock(&kvm->lock);
> + mutex_lock(&kvm->irq_lock);
> status = kvm_set_irq(kvm, KVM_USERSPACE_IRQ_SOURCE_ID,
> irq_event.irq, irq_event.level);
> - mutex_unlock(&kvm->lock);
> + mutex_unlock(&kvm->irq_lock);
> if (ioctl == KVM_IRQ_LINE_STATUS) {
> irq_event.status = status;
> if (copy_to_user(argp, &irq_event,
> @@ -2158,12 +2158,11 @@ mmio:
> */
> mutex_lock(&vcpu->kvm->lock);
> mmio_dev = vcpu_find_mmio_dev(vcpu, gpa, bytes, 0);
> + mutex_unlock(&vcpu->kvm->lock);
> if (mmio_dev) {
> kvm_iodevice_read(mmio_dev, gpa, bytes, val);
> - mutex_unlock(&vcpu->kvm->lock);
> return X86EMUL_CONTINUE;
> }
> - mutex_unlock(&vcpu->kvm->lock);
>
> vcpu->mmio_needed = 1;
>
> vcpu->mmio_phys_addr = gpa;
> @@ -2541,7 +2540,6 @@ static void kernel_pio(struct kvm_io_dev
> {
> /* TODO: String I/O for in kernel device */
>
> - mutex_lock(&vcpu->kvm->lock);
> if (vcpu->arch.pio.in)
> kvm_iodevice_read(pio_dev, vcpu->arch.pio.port,
> vcpu->arch.pio.size,
> @@ -2550,7 +2548,6 @@ static void kernel_pio(struct kvm_io_dev
> kvm_iodevice_write(pio_dev, vcpu->arch.pio.port,
> vcpu->arch.pio.size,
> pd);
> - mutex_unlock(&vcpu->kvm->lock);
> }
>
> static void pio_string_write(struct kvm_io_device *pio_dev,
> @@ -2560,14 +2557,12 @@ static void pio_string_write(struct kvm_
> void *pd = vcpu->arch.pio_data;
> int i;
>
> - mutex_lock(&vcpu->kvm->lock);
> for (i = 0; i < io->cur_count; i++) {
> kvm_iodevice_write(pio_dev, io->port,
> io->size,
> pd);
> pd += io->size;
> }
> - mutex_unlock(&vcpu->kvm->lock);
> }
>
> static struct kvm_io_device *vcpu_find_pio_dev(struct kvm_vcpu *vcpu,
> @@ -2604,7 +2599,9 @@ int kvm_emulate_pio(struct kvm_vcpu *vcp
> val = kvm_register_read(vcpu, VCPU_REGS_RAX);
> memcpy(vcpu->arch.pio_data, &val, 4);
>
> + mutex_lock(&vcpu->kvm->lock);
> pio_dev = vcpu_find_pio_dev(vcpu, port, size, !in);
> + mutex_unlock(&vcpu->kvm->lock);
> if (pio_dev) {
> kernel_pio(pio_dev, vcpu, vcpu->arch.pio_data);
> complete_pio(vcpu);
> @@ -2668,9 +2665,11 @@ int kvm_emulate_pio_string(struct kvm_vc
>
> vcpu->arch.pio.guest_gva = address;
>
> + mutex_lock(&vcpu->kvm->lock);
> pio_dev = vcpu_find_pio_dev(vcpu, port,
> vcpu->arch.pio.cur_count,
> !vcpu->arch.pio.in);
> + mutex_unlock(&vcpu->kvm->lock);
> if (!vcpu->arch.pio.in) {
> /* string PIO write */
> ret = pio_copy_data(vcpu);
> Index: kvm/virt/kvm/kvm_main.c
> ===================================================================
> --- kvm.orig/virt/kvm/kvm_main.c
> +++ kvm/virt/kvm/kvm_main.c
> @@ -62,6 +62,12 @@
> MODULE_AUTHOR("Qumranet");
> MODULE_LICENSE("GPL");
>
> +/*
> + * Ordering of locks:
> + *
> + * kvm->lock --> kvm->irq_lock
> + */
> +
> DEFINE_SPINLOCK(kvm_lock);
> LIST_HEAD(vm_list);
>
> @@ -126,11 +132,7 @@ static void kvm_assigned_dev_interrupt_w
> interrupt_work);
> kvm = assigned_dev->kvm;
>
> - /* This is taken to safely inject irq inside the guest. When
> - * the interrupt injection (or the ioapic code) uses a
> - * finer-grained lock, update this
> - */
> - mutex_lock(&kvm->lock);
> + mutex_lock(&kvm->irq_lock);
> spin_lock_irq(&assigned_dev->assigned_dev_lock);
> if (assigned_dev->irq_requested_type & KVM_DEV_IRQ_HOST_MSIX) {
> struct kvm_guest_msix_entry *guest_entries =
> @@ -159,7 +161,7 @@ static void kvm_assigned_dev_interrupt_w
> }
>
> spin_unlock_irq(&assigned_dev->assigned_dev_lock);
> - mutex_unlock(&assigned_dev->kvm->lock);
> + mutex_unlock(&assigned_dev->kvm->irq_lock);
> }
>
> static irqreturn_t kvm_assigned_dev_intr(int irq, void *dev_id)
> @@ -215,7 +217,7 @@ static void kvm_assigned_dev_ack_irq(str
> static void deassign_guest_irq(struct kvm *kvm,
> struct kvm_assigned_dev_kernel *assigned_dev)
> {
> - kvm_unregister_irq_ack_notifier(&assigned_dev->ack_notifier);
> + kvm_unregister_irq_ack_notifier(kvm, &assigned_dev->ack_notifier);
> assigned_dev->ack_notifier.gsi = -1;
>
> if (assigned_dev->irq_source_id != -1)
> Index: kvm/virt/kvm/irq_comm.c
> ===================================================================
> --- kvm.orig/virt/kvm/irq_comm.c
> +++ kvm/virt/kvm/irq_comm.c
> @@ -62,6 +62,8 @@ int kvm_irq_delivery_to_apic(struct kvm
> int i, r = -1;
> struct kvm_vcpu *vcpu, *lowest = NULL;
>
> + WARN_ON(!mutex_is_locked(&kvm->irq_lock));
> +
> if (irq->dest_mode == 0 && irq->dest_id == 0xff &&
> kvm_is_dm_lowest_prio(irq))
> printk(KERN_INFO "kvm: apic: phys broadcast and lowest prio\n");
> @@ -113,7 +115,7 @@ static int kvm_set_msi(struct kvm_kernel
> return kvm_irq_delivery_to_apic(kvm, NULL, &irq);
> }
>
> -/* This should be called with the kvm->lock mutex held
> +/* This should be called with the kvm->irq_lock mutex held
> * Return value:
> * < 0 Interrupt was ignored (masked or not delivered for other reasons)
> * = 0 Interrupt was coalesced (previous irq is still pending)
> @@ -125,6 +127,8 @@ int kvm_set_irq(struct kvm *kvm, int irq
> unsigned long *irq_state, sig_level;
> int ret = -1;
>
> + WARN_ON(!mutex_is_locked(&kvm->irq_lock));
> +
> if (irq < KVM_IOAPIC_NUM_PINS) {
> irq_state = (unsigned long *)&kvm->arch.irq_states[irq];
>
> @@ -174,19 +178,26 @@ void kvm_notify_acked_irq(struct kvm *kv
> void kvm_register_irq_ack_notifier(struct kvm *kvm,
> struct kvm_irq_ack_notifier *kian)
> {
> + mutex_lock(&kvm->irq_lock);
> hlist_add_head(&kian->link, &kvm->arch.irq_ack_notifier_list);
> + mutex_unlock(&kvm->irq_lock);
> }
>
> -void kvm_unregister_irq_ack_notifier(struct kvm_irq_ack_notifier *kian)
> +void kvm_unregister_irq_ack_notifier(struct kvm *kvm,
> + struct kvm_irq_ack_notifier *kian)
> {
> + mutex_lock(&kvm->irq_lock);
> hlist_del_init(&kian->link);
> + mutex_unlock(&kvm->irq_lock);
> }
>
> -/* The caller must hold kvm->lock mutex */
> int kvm_request_irq_source_id(struct kvm *kvm)
> {
> unsigned long *bitmap = &kvm->arch.irq_sources_bitmap;
> - int irq_source_id = find_first_zero_bit(bitmap,
> + int irq_source_id;
> +
> + mutex_lock(&kvm->irq_lock);
> + irq_source_id = find_first_zero_bit(bitmap,
> sizeof(kvm->arch.irq_sources_bitmap));
>
> if (irq_source_id >= sizeof(kvm->arch.irq_sources_bitmap)) {
> @@ -196,6 +207,7 @@ int kvm_request_irq_source_id(struct kvm
>
> ASSERT(irq_source_id != KVM_USERSPACE_IRQ_SOURCE_ID);
> set_bit(irq_source_id, bitmap);
> + mutex_unlock(&kvm->irq_lock);
>
> return irq_source_id;
> }
> @@ -206,6 +218,7 @@ void kvm_free_irq_source_id(struct kvm *
>
> ASSERT(irq_source_id != KVM_USERSPACE_IRQ_SOURCE_ID);
>
> + mutex_lock(&kvm->irq_lock);
> if (irq_source_id < 0 ||
> irq_source_id >= sizeof(kvm->arch.irq_sources_bitmap)) {
> printk(KERN_ERR "kvm: IRQ source ID out of range!\n");
> @@ -214,19 +227,24 @@ void kvm_free_irq_source_id(struct kvm *
> for (i = 0; i < KVM_IOAPIC_NUM_PINS; i++)
> clear_bit(irq_source_id, &kvm->arch.irq_states[i]);
> clear_bit(irq_source_id, &kvm->arch.irq_sources_bitmap);
> + mutex_unlock(&kvm->irq_lock);
> }
>
> void kvm_register_irq_mask_notifier(struct kvm *kvm, int irq,
> struct kvm_irq_mask_notifier *kimn)
> {
> + mutex_lock(&kvm->irq_lock);
> kimn->irq = irq;
> hlist_add_head(&kimn->link, &kvm->mask_notifier_list);
> + mutex_unlock(&kvm->irq_lock);
> }
>
> void kvm_unregister_irq_mask_notifier(struct kvm *kvm, int irq,
> struct kvm_irq_mask_notifier *kimn)
> {
> + mutex_lock(&kvm->irq_lock);
> hlist_del(&kimn->link);
> + mutex_unlock(&kvm->irq_lock);
> }
>
> void kvm_fire_mask_notifiers(struct kvm *kvm, int irq, bool mask)
> @@ -234,6 +252,8 @@ void kvm_fire_mask_notifiers(struct kvm
> struct kvm_irq_mask_notifier *kimn;
> struct hlist_node *n;
>
> + WARN_ON(!mutex_is_locked(&kvm->lock));
> +
> hlist_for_each_entry(kimn, n, &kvm->mask_notifier_list, link)
> if (kimn->irq == irq)
> kimn->func(kimn, mask);
> @@ -249,7 +269,9 @@ static void __kvm_free_irq_routing(struc
>
> void kvm_free_irq_routing(struct kvm *kvm)
> {
> + mutex_lock(&kvm->irq_lock);
> __kvm_free_irq_routing(&kvm->irq_routing);
> + mutex_unlock(&kvm->irq_lock);
> }
>
> static int setup_routing_entry(struct kvm_kernel_irq_routing_entry *e,
> @@ -323,13 +345,13 @@ int kvm_set_irq_routing(struct kvm *kvm,
> e = NULL;
> }
>
> - mutex_lock(&kvm->lock);
> + mutex_lock(&kvm->irq_lock);
> list_splice(&kvm->irq_routing, &tmp);
> INIT_LIST_HEAD(&kvm->irq_routing);
> list_splice(&irq_list, &kvm->irq_routing);
> INIT_LIST_HEAD(&irq_list);
> list_splice(&tmp, &irq_list);
> - mutex_unlock(&kvm->lock);
> + mutex_unlock(&kvm->irq_lock);
>
> r = 0;
>
> Index: kvm/arch/x86/kvm/lapic.c
> ===================================================================
> --- kvm.orig/arch/x86/kvm/lapic.c
> +++ kvm/arch/x86/kvm/lapic.c
> @@ -425,7 +425,9 @@ static void apic_set_eoi(struct kvm_lapi
> trigger_mode = IOAPIC_LEVEL_TRIG;
> else
> trigger_mode = IOAPIC_EDGE_TRIG;
> + mutex_lock(&apic->vcpu->kvm->irq_lock);
> kvm_ioapic_update_eoi(apic->vcpu->kvm, vector, trigger_mode);
> + mutex_unlock(&apic->vcpu->kvm->irq_lock);
> }
>
> static void apic_send_ipi(struct kvm_lapic *apic)
> @@ -449,7 +451,9 @@ static void apic_send_ipi(struct kvm_lap
> irq.trig_mode, irq.level, irq.dest_mode, irq.delivery_mode,
> irq.vector);
>
> + mutex_lock(&apic->vcpu->kvm->irq_lock);
> kvm_irq_delivery_to_apic(apic->vcpu->kvm, apic, &irq);
> + mutex_unlock(&apic->vcpu->kvm->irq_lock);
> }
>
> static u32 apic_get_tmcct(struct kvm_lapic *apic)
> Index: kvm/include/linux/kvm_host.h
> ===================================================================
> --- kvm.orig/include/linux/kvm_host.h
> +++ kvm/include/linux/kvm_host.h
> @@ -375,7 +375,8 @@ int kvm_set_irq(struct kvm *kvm, int irq
> void kvm_notify_acked_irq(struct kvm *kvm, unsigned irqchip, unsigned pin);
> void kvm_register_irq_ack_notifier(struct kvm *kvm,
> struct kvm_irq_ack_notifier *kian);
> -void kvm_unregister_irq_ack_notifier(struct kvm_irq_ack_notifier *kian);
> +void kvm_unregister_irq_ack_notifier(struct kvm *kvm,
> + struct kvm_irq_ack_notifier *kian);
> int kvm_request_irq_source_id(struct kvm *kvm);
> void kvm_free_irq_source_id(struct kvm *kvm, int irq_source_id);
>
>
>
signature.asc
Description: OpenPGP digital signature
