mmu_shrink() should attempt to free @nr_to_scan entries.

Signed-off-by: Lai Jiangshan <[email protected]>
---
diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c
index 9c69725..1034373 100644
--- a/arch/x86/kvm/mmu.c
+++ b/arch/x86/kvm/mmu.c
@@ -3138,37 +3138,51 @@ static int mmu_shrink(struct shrinker *shrink, int 
nr_to_scan, gfp_t gfp_mask)
 {
        struct kvm *kvm;
        struct kvm *kvm_freed = NULL;
+       struct kvm *kvm_last;
        int cache_count = 0;
 
        spin_lock(&kvm_lock);
 
-       list_for_each_entry(kvm, &vm_list, vm_list) {
+       if (list_empty(&vm_list))
+               goto out;
+
+       kvm_last = list_entry(vm_list.prev, struct kvm, vm_list);
+
+       for (;;) {
                int npages, idx, freed_pages;
                LIST_HEAD(invalid_list);
 
+               kvm = list_first_entry(&vm_list, struct kvm, vm_list);
                idx = srcu_read_lock(&kvm->srcu);
                spin_lock(&kvm->mmu_lock);
                npages = kvm->arch.n_alloc_mmu_pages -
                         kvm->arch.n_free_mmu_pages;
-               cache_count += npages;
-               if (!kvm_freed && nr_to_scan > 0 && npages > 0) {
+               if (kvm_last)
+                       cache_count += npages;
+               if (nr_to_scan > 0 && npages > 0) {
                        freed_pages = kvm_mmu_remove_some_alloc_mmu_pages(kvm,
                                                          &invalid_list);
+                       kvm_mmu_commit_zap_page(kvm, &invalid_list);
                        cache_count -= freed_pages;
                        kvm_freed = kvm;
-               }
-               nr_to_scan--;
+                       nr_to_scan -= freed_pages;
+               } else if (kvm == kvm_freed)
+                       nr_to_scan = 0; /* no more page to be freed, break */
 
-               kvm_mmu_commit_zap_page(kvm, &invalid_list);
                spin_unlock(&kvm->mmu_lock);
                srcu_read_unlock(&kvm->srcu, idx);
-       }
-       if (kvm_freed)
                list_move_tail(&kvm_freed->vm_list, &vm_list);
 
+               if (kvm == kvm_last) /* just scaned all vms */
+                       kvm_last = NULL;
+               if (!kvm_last && (nr_to_scan <= 0 || !kvm_freed))
+                       break;
+       }
+
+out:
        spin_unlock(&kvm_lock);
 
-       return cache_count;
+       return cache_count < 0 ? 0 : cache_count;
 }
 
 static struct shrinker mmu_shrinker = {
--
To unsubscribe from this list: send the line "unsubscribe kvm" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to