Alright, I'll send that soon.  Sorry for the mixup.

On Wed, Apr 17, 2013 at 8:30 AM, Gleb Natapov <g...@redhat.com> wrote:
> On Wed, Apr 17, 2013 at 08:24:21AM -0700, Andrew Honig wrote:
>> I'm happy to not add a new API and use __kvm_set_memory_region to
>> unregister private memory regions, but I thought chaning the API was
>> the approach you asked for when I sent a previous patch.  See the end
>> of: http://article.gmane.org/gmane.comp.emulators.kvm.devel/107753
>>
>> Did I misunderstand your comment from 8 April?
>>
> Ugh, yes. My "Please send a second version" was in response to your
> question "or should I send a second version of this current patch?"
> Where second version is the same as the first one but uses
> __kvm_set_memory_region() for slot deletion. I am sorry about the
> confusion :(
>
>> On Wed, Apr 17, 2013 at 6:10 AM, Gleb Natapov <g...@redhat.com> wrote:
>> > On Mon, Apr 15, 2013 at 03:10:32PM -0700, Andrew Honig wrote:
>> >>
>> >> The motivation for this patch is to fix a 20KB leak of memory in vmx.c
>> >> when a VM is created and destroyed.
>> >>
>> >> On x86/vmx platforms KVM needs 5 pages of userspace memory per VM for
>> >> architecture specific reasons.  It currently allocates the pages on behalf
>> >> of user space, but has no way of cleanly freeing that memory while the
>> >> user space process is still running.  For user space processes that want
>> >> more control over that memory, this patch allows user space to provide the
>> >> memory that KVM uses.
>> >>
>> >> Signed-off-by: Andrew Honig <aho...@google.com>
>> > I thought we agreed on not adding new API and using
>> > __kvm_set_memory_region() to unregister private memory regions.
>> >
>> >
>> >> ---
>> >>  Documentation/virtual/kvm/api.txt |    8 +++++++
>> >>  arch/arm/kvm/arm.c                |    6 +++++
>> >>  arch/ia64/kvm/kvm-ia64.c          |    6 +++++
>> >>  arch/powerpc/kvm/powerpc.c        |    6 +++++
>> >>  arch/s390/kvm/kvm-s390.c          |    6 +++++
>> >>  arch/x86/include/asm/kvm_host.h   |    7 ++++++
>> >>  arch/x86/kvm/svm.c                |    8 +++++++
>> >>  arch/x86/kvm/vmx.c                |   47 
>> >> ++++++++++++++++++++++++++++++++++---
>> >>  arch/x86/kvm/x86.c                |   12 ++++++++--
>> >>  include/linux/kvm_host.h          |    2 ++
>> >>  virt/kvm/kvm_main.c               |    2 ++
>> >>  11 files changed, 105 insertions(+), 5 deletions(-)
>> >>
>> >> diff --git a/Documentation/virtual/kvm/api.txt 
>> >> b/Documentation/virtual/kvm/api.txt
>> >> index 119358d..aa18cac 100644
>> >> --- a/Documentation/virtual/kvm/api.txt
>> >> +++ b/Documentation/virtual/kvm/api.txt
>> >> @@ -879,6 +879,14 @@ It is recommended that the lower 21 bits of 
>> >> guest_phys_addr and userspace_addr
>> >>  be identical.  This allows large pages in the guest to be backed by large
>> >>  pages in the host.
>> >>
>> >> +On x86/vmx architectures KVM needs 5 pages of user space memory for 
>> >> architecture
>> >> +specific reasons.  Calling this ioctl with the special memslot
>> >> +KVM_PRIVATE_MEMORY_MEMSLOT will tell kvm which user space memory to use 
>> >> for
>> >> +that memory.  If that memslot is not set before creating VCPUs for a VM 
>> >> then
>> >> +kvm will allocate the memory on behalf of user space, but userspace will 
>> >> not
>> >> +be able to free that memory.  User space should treat this memory as 
>> >> opaque
>> >> +and not modify it.
>> >> +
>> >>  The flags field supports two flags: KVM_MEM_LOG_DIRTY_PAGES and
>> >>  KVM_MEM_READONLY.  The former can be set to instruct KVM to keep track of
>> >>  writes to memory within the slot.  See KVM_GET_DIRTY_LOG ioctl to know 
>> >> how to
>> >> diff --git a/arch/arm/kvm/arm.c b/arch/arm/kvm/arm.c
>> >> index 5a93698..ac52f14 100644
>> >> --- a/arch/arm/kvm/arm.c
>> >> +++ b/arch/arm/kvm/arm.c
>> >> @@ -228,6 +228,12 @@ int kvm_arch_set_memory_region(struct kvm *kvm,
>> >>       return 0;
>> >>  }
>> >>
>> >> +int kvm_arch_set_private_memory(struct kvm *kvm,
>> >> +                             struct kvm_userspace_memory_region *mem)
>> >> +{
>> >> +     return 0;
>> >> +}
>> >> +
>> >>  int kvm_arch_prepare_memory_region(struct kvm *kvm,
>> >>                                  struct kvm_memory_slot *memslot,
>> >>                                  struct kvm_memory_slot old,
>> >> diff --git a/arch/ia64/kvm/kvm-ia64.c b/arch/ia64/kvm/kvm-ia64.c
>> >> index ad3126a..570dd97 100644
>> >> --- a/arch/ia64/kvm/kvm-ia64.c
>> >> +++ b/arch/ia64/kvm/kvm-ia64.c
>> >> @@ -1576,6 +1576,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot 
>> >> *slot, unsigned long npages)
>> >>       return 0;
>> >>  }
>> >>
>> >> +int kvm_arch_set_private_memory(struct kvm *kvm,
>> >> +                             struct kvm_userspace_memory_region *mem)
>> >> +{
>> >> +     return 0;
>> >> +}
>> >> +
>> >>  int kvm_arch_prepare_memory_region(struct kvm *kvm,
>> >>               struct kvm_memory_slot *memslot,
>> >>               struct kvm_memory_slot old,
>> >> diff --git a/arch/powerpc/kvm/powerpc.c b/arch/powerpc/kvm/powerpc.c
>> >> index 934413c..6e3843b 100644
>> >> --- a/arch/powerpc/kvm/powerpc.c
>> >> +++ b/arch/powerpc/kvm/powerpc.c
>> >> @@ -410,6 +410,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot 
>> >> *slot, unsigned long npages)
>> >>       return kvmppc_core_create_memslot(slot, npages);
>> >>  }
>> >>
>> >> +int kvm_arch_set_private_memory(struct kvm *kvm,
>> >> +                             struct kvm_userspace_memory_region *mem)
>> >> +{
>> >> +     return 0;
>> >> +}
>> >> +
>> >>  int kvm_arch_prepare_memory_region(struct kvm *kvm,
>> >>                                     struct kvm_memory_slot *memslot,
>> >>                                     struct kvm_memory_slot old,
>> >> diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c
>> >> index 4cf35a0..a97f495 100644
>> >> --- a/arch/s390/kvm/kvm-s390.c
>> >> +++ b/arch/s390/kvm/kvm-s390.c
>> >> @@ -971,6 +971,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot 
>> >> *slot, unsigned long npages)
>> >>       return 0;
>> >>  }
>> >>
>> >> +int kvm_arch_set_private_memory(struct kvm *kvm,
>> >> +                             struct kvm_userspace_memory_region *mem)
>> >> +{
>> >> +     return 0;
>> >> +}
>> >> +
>> >>  /* Section: memory related */
>> >>  int kvm_arch_prepare_memory_region(struct kvm *kvm,
>> >>                                  struct kvm_memory_slot *memslot,
>> >> diff --git a/arch/x86/include/asm/kvm_host.h 
>> >> b/arch/x86/include/asm/kvm_host.h
>> >> index 4979778..7215817 100644
>> >> --- a/arch/x86/include/asm/kvm_host.h
>> >> +++ b/arch/x86/include/asm/kvm_host.h
>> >> @@ -37,6 +37,7 @@
>> >>  /* memory slots that are not exposed to userspace */
>> >>  #define KVM_PRIVATE_MEM_SLOTS 3
>> >>  #define KVM_MEM_SLOTS_NUM (KVM_USER_MEM_SLOTS + KVM_PRIVATE_MEM_SLOTS)
>> >> +#define KVM_PRIVATE_MEMORY_MEMSLOT 0x80000001
>> >>
>> >>  #define KVM_MMIO_SIZE 16
>> >>
>> >> @@ -553,6 +554,9 @@ struct kvm_arch {
>> >>       struct page *ept_identity_pagetable;
>> >>       bool ept_identity_pagetable_done;
>> >>       gpa_t ept_identity_map_addr;
>> >> +     unsigned long ept_ptr;
>> >> +     unsigned long apic_ptr;
>> >> +     unsigned long tss_ptr;
>> >>
>> >>       unsigned long irq_sources_bitmap;
>> >>       s64 kvmclock_offset;
>> >> @@ -640,6 +644,9 @@ struct kvm_x86_ops {
>> >>       bool (*cpu_has_accelerated_tpr)(void);
>> >>       void (*cpuid_update)(struct kvm_vcpu *vcpu);
>> >>
>> >> +     int (*set_private_memory)(struct kvm *kvm,
>> >> +                               struct kvm_userspace_memory_region *mem);
>> >> +
>> >>       /* Create, but do not attach this VCPU */
>> >>       struct kvm_vcpu *(*vcpu_create)(struct kvm *kvm, unsigned id);
>> >>       void (*vcpu_free)(struct kvm_vcpu *vcpu);
>> >> diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c
>> >> index e1b1ce2..3cc4e56 100644
>> >> --- a/arch/x86/kvm/svm.c
>> >> +++ b/arch/x86/kvm/svm.c
>> >> @@ -1211,6 +1211,12 @@ static int svm_vcpu_reset(struct kvm_vcpu *vcpu)
>> >>       return 0;
>> >>  }
>> >>
>> >> +static int svm_set_private_memory(struct kvm *kvm,
>> >> +                               struct kvm_userspace_memory_region *mem)
>> >> +{
>> >> +     return 0;
>> >> +}
>> >> +
>> >>  static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id)
>> >>  {
>> >>       struct vcpu_svm *svm;
>> >> @@ -4257,6 +4263,8 @@ static struct kvm_x86_ops svm_x86_ops = {
>> >>       .hardware_disable = svm_hardware_disable,
>> >>       .cpu_has_accelerated_tpr = svm_cpu_has_accelerated_tpr,
>> >>
>> >> +     .set_private_memory = svm_set_private_memory,
>> >> +
>> >>       .vcpu_create = svm_create_vcpu,
>> >>       .vcpu_free = svm_free_vcpu,
>> >>       .vcpu_reset = svm_vcpu_reset,
>> >> diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c
>> >> index 6667042..796ac07 100644
>> >> --- a/arch/x86/kvm/vmx.c
>> >> +++ b/arch/x86/kvm/vmx.c
>> >> @@ -3692,7 +3692,13 @@ static int alloc_apic_access_page(struct kvm *kvm)
>> >>       kvm_userspace_mem.flags = 0;
>> >>       kvm_userspace_mem.guest_phys_addr = 0xfee00000ULL;
>> >>       kvm_userspace_mem.memory_size = PAGE_SIZE;
>> >> -     r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false);
>> >> +     if (kvm->arch.apic_ptr) {
>> >> +             kvm_userspace_mem.userspace_addr = kvm->arch.apic_ptr;
>> >> +             r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true);
>> >> +     } else {
>> >> +             r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false);
>> >> +     }
>> >> +
>> >>       if (r)
>> >>               goto out;
>> >>
>> >> @@ -3722,7 +3728,13 @@ static int alloc_identity_pagetable(struct kvm 
>> >> *kvm)
>> >>       kvm_userspace_mem.guest_phys_addr =
>> >>               kvm->arch.ept_identity_map_addr;
>> >>       kvm_userspace_mem.memory_size = PAGE_SIZE;
>> >> -     r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false);
>> >> +     if (kvm->arch.ept_ptr) {
>> >> +             kvm_userspace_mem.userspace_addr = kvm->arch.ept_ptr;
>> >> +             r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true);
>> >> +     } else {
>> >> +             r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false);
>> >> +     }
>> >> +
>> >>       if (r)
>> >>               goto out;
>> >>
>> >> @@ -4362,7 +4374,13 @@ static int vmx_set_tss_addr(struct kvm *kvm, 
>> >> unsigned int addr)
>> >>               .flags = 0,
>> >>       };
>> >>
>> >> -     ret = kvm_set_memory_region(kvm, &tss_mem, false);
>> >> +     if (kvm->arch.tss_ptr) {
>> >> +             tss_mem.userspace_addr = kvm->arch.tss_ptr;
>> >> +             ret = kvm_set_memory_region(kvm, &tss_mem, true);
>> >> +     } else {
>> >> +             ret = kvm_set_memory_region(kvm, &tss_mem, false);
>> >> +     }
>> >> +
>> >>       if (ret)
>> >>               return ret;
>> >>       kvm->arch.tss_addr = addr;
>> >> @@ -6683,6 +6701,27 @@ static void __noclone vmx_vcpu_run(struct kvm_vcpu 
>> >> *vcpu)
>> >>       vmx_complete_interrupts(vmx);
>> >>  }
>> >>
>> >> +static int vmx_set_private_memory(struct kvm *kvm,
>> >> +                               struct kvm_userspace_memory_region *mem)
>> >> +{
>> >> +     /*
>> >> +      * Early sanity checking so userspace gets an error message during
>> >> +      * memory setup and not when trying to use this memory.
>> >> +      * Checks to see if the memory is valid are performed later when
>> >> +      * the memory is used.
>> >> +      */
>> >> +     if (!mem->userspace_addr || mem->userspace_addr & (PAGE_SIZE - 1) ||
>> >> +                     mem->memory_size & (PAGE_SIZE - 1) ||
>> >> +                     mem->memory_size < PAGE_SIZE * 5)
>> >> +             return -EINVAL;
>> >> +
>> >> +     kvm->arch.ept_ptr = mem->userspace_addr;
>> >> +     kvm->arch.apic_ptr = mem->userspace_addr + PAGE_SIZE;
>> >> +     kvm->arch.tss_ptr = mem->userspace_addr + PAGE_SIZE * 2;
>> >> +
>> >> +     return 0;
>> >> +}
>> >> +
>> >>  static void vmx_free_vcpu(struct kvm_vcpu *vcpu)
>> >>  {
>> >>       struct vcpu_vmx *vmx = to_vmx(vcpu);
>> >> @@ -7532,6 +7571,8 @@ static struct kvm_x86_ops vmx_x86_ops = {
>> >>       .hardware_disable = hardware_disable,
>> >>       .cpu_has_accelerated_tpr = report_flexpriority,
>> >>
>> >> +     .set_private_memory = vmx_set_private_memory,
>> >> +
>> >>       .vcpu_create = vmx_create_vcpu,
>> >>       .vcpu_free = vmx_free_vcpu,
>> >>       .vcpu_reset = vmx_vcpu_reset,
>> >> diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
>> >> index e172132..7045d0a 100644
>> >> --- a/arch/x86/kvm/x86.c
>> >> +++ b/arch/x86/kvm/x86.c
>> >> @@ -6809,6 +6809,12 @@ void kvm_arch_sync_events(struct kvm *kvm)
>> >>       kvm_free_pit(kvm);
>> >>  }
>> >>
>> >> +int kvm_arch_set_private_memory(struct kvm *kvm,
>> >> +                         struct kvm_userspace_memory_region *mem)
>> >> +{
>> >> +     return kvm_x86_ops->set_private_memory(kvm, mem);
>> >> +}
>> >> +
>> >>  void kvm_arch_destroy_vm(struct kvm *kvm)
>> >>  {
>> >>       kvm_iommu_unmap_guest(kvm);
>> >> @@ -6913,7 +6919,8 @@ int kvm_arch_prepare_memory_region(struct kvm *kvm,
>> >>        * Only private memory slots need to be mapped here since
>> >>        * KVM_SET_MEMORY_REGION ioctl is no longer supported.
>> >>        */
>> >> -     if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages) {
>> >> +     if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages &&
>> >> +                                             !user_alloc) {
>> >>               unsigned long userspace_addr;
>> >>
>> >>               /*
>> >> @@ -6941,7 +6948,8 @@ void kvm_arch_commit_memory_region(struct kvm *kvm,
>> >>
>> >>       int nr_mmu_pages = 0, npages = mem->memory_size >> PAGE_SHIFT;
>> >>
>> >> -     if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages) {
>> >> +     if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages &&
>> >> +                                             !user_alloc) {
>> >>               int ret;
>> >>
>> >>               ret = vm_munmap(old.userspace_addr,
>> >> diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h
>> >> index c139582..f441d1f 100644
>> >> --- a/include/linux/kvm_host.h
>> >> +++ b/include/linux/kvm_host.h
>> >> @@ -461,6 +461,8 @@ int __kvm_set_memory_region(struct kvm *kvm,
>> >>  void kvm_arch_free_memslot(struct kvm_memory_slot *free,
>> >>                          struct kvm_memory_slot *dont);
>> >>  int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long 
>> >> npages);
>> >> +int kvm_arch_set_private_memory(struct kvm *kvm,
>> >> +                             struct kvm_userspace_memory_region *mem);
>> >>  int kvm_arch_prepare_memory_region(struct kvm *kvm,
>> >>                               struct kvm_memory_slot *memslot,
>> >>                               struct kvm_memory_slot old,
>> >> diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c
>> >> index f18013f..5372225 100644
>> >> --- a/virt/kvm/kvm_main.c
>> >> +++ b/virt/kvm/kvm_main.c
>> >> @@ -949,6 +949,8 @@ int kvm_vm_ioctl_set_memory_region(struct kvm *kvm,
>> >>                                  kvm_userspace_memory_region *mem,
>> >>                                  bool user_alloc)
>> >>  {
>> >> +     if (mem->slot == KVM_PRIVATE_MEMORY_MEMSLOT)
>> >> +             return kvm_arch_set_private_memory(kvm, mem);
>> >>       if (mem->slot >= KVM_USER_MEM_SLOTS)
>> >>               return -EINVAL;
>> >>       return kvm_set_memory_region(kvm, mem, user_alloc);
>> >> --
>> >> 1.7.10.4
>> >>
>> >> --
>> >> To unsubscribe from this list: send the line "unsubscribe kvm" in
>> >> the body of a message to majord...@vger.kernel.org
>> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> >
>> > --
>> >                         Gleb.
>
> --
>                         Gleb.
--
To unsubscribe from this list: send the line "unsubscribe kvm" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to