Move VM-level TDP MMU state to struct kvm so it can be accessed by
common code in a future commit.

No functional change intended.

Signed-off-by: David Matlack <dmatl...@google.com>
---
 arch/x86/include/asm/kvm_host.h | 39 --------------------------------
 arch/x86/kvm/mmu/tdp_mmu.c      | 40 ++++++++++++++++-----------------
 arch/x86/kvm/mmu/tdp_pgtable.c  |  8 +++----
 include/kvm/mmu_types.h         | 40 +++++++++++++++++++++++++++++++++
 include/linux/kvm_host.h        |  8 +++++++
 5 files changed, 72 insertions(+), 63 deletions(-)

diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h
index 9cf8f956bac3..95c731028452 100644
--- a/arch/x86/include/asm/kvm_host.h
+++ b/arch/x86/include/asm/kvm_host.h
@@ -1272,45 +1272,6 @@ struct kvm_arch {
        struct kvm_pmu_event_filter __rcu *pmu_event_filter;
        struct task_struct *nx_huge_page_recovery_thread;
 
-#ifdef CONFIG_X86_64
-       /* The number of TDP MMU pages across all roots. */
-       atomic64_t tdp_mmu_pages;
-
-       /*
-        * List of struct kvm_mmu_pages being used as roots.
-        * All struct kvm_mmu_pages in the list should have
-        * tdp_mmu_page set.
-        *
-        * For reads, this list is protected by:
-        *      the MMU lock in read mode + RCU or
-        *      the MMU lock in write mode
-        *
-        * For writes, this list is protected by:
-        *      the MMU lock in read mode + the tdp_mmu_pages_lock or
-        *      the MMU lock in write mode
-        *
-        * Roots will remain in the list until their tdp_mmu_root_count
-        * drops to zero, at which point the thread that decremented the
-        * count to zero should removed the root from the list and clean
-        * it up, freeing the root after an RCU grace period.
-        */
-       struct list_head tdp_mmu_roots;
-
-       /*
-        * Protects accesses to the following fields when the MMU lock
-        * is held in read mode:
-        *  - tdp_mmu_roots (above)
-        *  - the link field of kvm_mmu_page structs used by the TDP MMU
-        *  - possible_nx_huge_pages;
-        *  - the possible_nx_huge_page_link field of kvm_mmu_page structs used
-        *    by the TDP MMU
-        * It is acceptable, but not necessary, to acquire this lock when
-        * the thread holds the MMU lock in write mode.
-        */
-       spinlock_t tdp_mmu_pages_lock;
-       struct workqueue_struct *tdp_mmu_zap_wq;
-#endif /* CONFIG_X86_64 */
-
        /*
         * If set, at least one shadow root has been allocated. This flag
         * is used as one input when determining whether certain memslot
diff --git a/arch/x86/kvm/mmu/tdp_mmu.c b/arch/x86/kvm/mmu/tdp_mmu.c
index 24d1dbd0a1ec..b997f84c0ea7 100644
--- a/arch/x86/kvm/mmu/tdp_mmu.c
+++ b/arch/x86/kvm/mmu/tdp_mmu.c
@@ -21,9 +21,9 @@ int kvm_mmu_init_tdp_mmu(struct kvm *kvm)
        if (!wq)
                return -ENOMEM;
 
-       INIT_LIST_HEAD(&kvm->arch.tdp_mmu_roots);
-       spin_lock_init(&kvm->arch.tdp_mmu_pages_lock);
-       kvm->arch.tdp_mmu_zap_wq = wq;
+       INIT_LIST_HEAD(&kvm->tdp_mmu.roots);
+       spin_lock_init(&kvm->tdp_mmu.pages_lock);
+       kvm->tdp_mmu.zap_wq = wq;
        return 1;
 }
 
@@ -42,10 +42,10 @@ static __always_inline bool 
kvm_lockdep_assert_mmu_lock_held(struct kvm *kvm,
 void kvm_mmu_uninit_tdp_mmu(struct kvm *kvm)
 {
        /* Also waits for any queued work items.  */
-       destroy_workqueue(kvm->arch.tdp_mmu_zap_wq);
+       destroy_workqueue(kvm->tdp_mmu.zap_wq);
 
-       WARN_ON(atomic64_read(&kvm->arch.tdp_mmu_pages));
-       WARN_ON(!list_empty(&kvm->arch.tdp_mmu_roots));
+       WARN_ON(atomic64_read(&kvm->tdp_mmu.pages));
+       WARN_ON(!list_empty(&kvm->tdp_mmu.roots));
 
        /*
         * Ensure that all the outstanding RCU callbacks to free shadow pages
@@ -114,7 +114,7 @@ static void tdp_mmu_schedule_zap_root(struct kvm *kvm, 
struct kvm_mmu_page *root
 {
        root->tdp_mmu_async_data = kvm;
        INIT_WORK(&root->tdp_mmu_async_work, tdp_mmu_zap_root_work);
-       queue_work(kvm->arch.tdp_mmu_zap_wq, &root->tdp_mmu_async_work);
+       queue_work(kvm->tdp_mmu.zap_wq, &root->tdp_mmu_async_work);
 }
 
 static inline bool kvm_tdp_root_mark_invalid(struct kvm_mmu_page *page)
@@ -173,9 +173,9 @@ void kvm_tdp_mmu_put_root(struct kvm *kvm, struct 
kvm_mmu_page *root,
                return;
        }
 
-       spin_lock(&kvm->arch.tdp_mmu_pages_lock);
+       spin_lock(&kvm->tdp_mmu.pages_lock);
        list_del_rcu(&root->link);
-       spin_unlock(&kvm->arch.tdp_mmu_pages_lock);
+       spin_unlock(&kvm->tdp_mmu.pages_lock);
        call_rcu(&root->rcu_head, tdp_mmu_free_sp_rcu_callback);
 }
 
@@ -198,11 +198,11 @@ static struct kvm_mmu_page *tdp_mmu_next_root(struct kvm 
*kvm,
        rcu_read_lock();
 
        if (prev_root)
-               next_root = list_next_or_null_rcu(&kvm->arch.tdp_mmu_roots,
+               next_root = list_next_or_null_rcu(&kvm->tdp_mmu.roots,
                                                  &prev_root->link,
                                                  typeof(*prev_root), link);
        else
-               next_root = list_first_or_null_rcu(&kvm->arch.tdp_mmu_roots,
+               next_root = list_first_or_null_rcu(&kvm->tdp_mmu.roots,
                                                   typeof(*next_root), link);
 
        while (next_root) {
@@ -210,7 +210,7 @@ static struct kvm_mmu_page *tdp_mmu_next_root(struct kvm 
*kvm,
                    kvm_tdp_mmu_get_root(next_root))
                        break;
 
-               next_root = list_next_or_null_rcu(&kvm->arch.tdp_mmu_roots,
+               next_root = list_next_or_null_rcu(&kvm->tdp_mmu.roots,
                                &next_root->link, typeof(*next_root), link);
        }
 
@@ -254,7 +254,7 @@ static struct kvm_mmu_page *tdp_mmu_next_root(struct kvm 
*kvm,
  * is guaranteed to be stable.
  */
 #define for_each_tdp_mmu_root(_kvm, _root, _as_id)                     \
-       list_for_each_entry(_root, &_kvm->arch.tdp_mmu_roots, link)     \
+       list_for_each_entry(_root, &_kvm->tdp_mmu.roots, link)  \
                if (kvm_lockdep_assert_mmu_lock_held(_kvm, false) &&    \
                    _root->role.as_id != _as_id) {              \
                } else
@@ -324,9 +324,9 @@ hpa_t kvm_tdp_mmu_get_vcpu_root_hpa(struct kvm_vcpu *vcpu)
 
        refcount_set(&root->root_refcount, 1);
 
-       spin_lock(&kvm->arch.tdp_mmu_pages_lock);
-       list_add_rcu(&root->link, &kvm->arch.tdp_mmu_roots);
-       spin_unlock(&kvm->arch.tdp_mmu_pages_lock);
+       spin_lock(&kvm->tdp_mmu.pages_lock);
+       list_add_rcu(&root->link, &kvm->tdp_mmu.roots);
+       spin_unlock(&kvm->tdp_mmu.pages_lock);
 
 out:
        return __pa(root->spt);
@@ -368,13 +368,13 @@ static void handle_changed_spte_dirty_log(struct kvm 
*kvm, int as_id, gfn_t gfn,
 static void tdp_account_mmu_page(struct kvm *kvm, struct kvm_mmu_page *sp)
 {
        kvm_account_pgtable_pages((void *)sp->spt, +1);
-       atomic64_inc(&kvm->arch.tdp_mmu_pages);
+       atomic64_inc(&kvm->tdp_mmu.pages);
 }
 
 static void tdp_unaccount_mmu_page(struct kvm *kvm, struct kvm_mmu_page *sp)
 {
        kvm_account_pgtable_pages((void *)sp->spt, -1);
-       atomic64_dec(&kvm->arch.tdp_mmu_pages);
+       atomic64_dec(&kvm->tdp_mmu.pages);
 }
 
 __weak void tdp_mmu_arch_unlink_sp(struct kvm *kvm, struct kvm_mmu_page *sp,
@@ -1010,7 +1010,7 @@ void kvm_tdp_mmu_zap_all(struct kvm *kvm)
  */
 void kvm_tdp_mmu_zap_invalidated_roots(struct kvm *kvm)
 {
-       flush_workqueue(kvm->arch.tdp_mmu_zap_wq);
+       flush_workqueue(kvm->tdp_mmu.zap_wq);
 }
 
 /*
@@ -1035,7 +1035,7 @@ void kvm_tdp_mmu_invalidate_all_roots(struct kvm *kvm)
        struct kvm_mmu_page *root;
 
        lockdep_assert_held_write(&kvm->mmu_lock);
-       list_for_each_entry(root, &kvm->arch.tdp_mmu_roots, link) {
+       list_for_each_entry(root, &kvm->tdp_mmu.roots, link) {
                if (!root->role.invalid &&
                    !WARN_ON_ONCE(!kvm_tdp_mmu_get_root(root))) {
                        root->role.invalid = true;
diff --git a/arch/x86/kvm/mmu/tdp_pgtable.c b/arch/x86/kvm/mmu/tdp_pgtable.c
index 840d063c45b8..cc7b10f703e1 100644
--- a/arch/x86/kvm/mmu/tdp_pgtable.c
+++ b/arch/x86/kvm/mmu/tdp_pgtable.c
@@ -141,9 +141,9 @@ void tdp_mmu_arch_post_link_sp(struct kvm *kvm,
        if (fault->req_level < sp->role.level)
                return;
 
-       spin_lock(&kvm->arch.tdp_mmu_pages_lock);
+       spin_lock(&kvm->tdp_mmu.pages_lock);
        track_possible_nx_huge_page(kvm, sp);
-       spin_unlock(&kvm->arch.tdp_mmu_pages_lock);
+       spin_unlock(&kvm->tdp_mmu.pages_lock);
 }
 
 void tdp_mmu_arch_unlink_sp(struct kvm *kvm, struct kvm_mmu_page *sp,
@@ -153,7 +153,7 @@ void tdp_mmu_arch_unlink_sp(struct kvm *kvm, struct 
kvm_mmu_page *sp,
                return;
 
        if (shared)
-               spin_lock(&kvm->arch.tdp_mmu_pages_lock);
+               spin_lock(&kvm->tdp_mmu.pages_lock);
        else
                lockdep_assert_held_write(&kvm->mmu_lock);
 
@@ -161,7 +161,7 @@ void tdp_mmu_arch_unlink_sp(struct kvm *kvm, struct 
kvm_mmu_page *sp,
        untrack_possible_nx_huge_page(kvm, sp);
 
        if (shared)
-               spin_unlock(&kvm->arch.tdp_mmu_pages_lock);
+               spin_unlock(&kvm->tdp_mmu.pages_lock);
 }
 
 int tdp_mmu_max_mapping_level(struct kvm *kvm,
diff --git a/include/kvm/mmu_types.h b/include/kvm/mmu_types.h
index 07c9962f9aea..8ccc48a1cd4c 100644
--- a/include/kvm/mmu_types.h
+++ b/include/kvm/mmu_types.h
@@ -136,4 +136,44 @@ enum {
        RET_PF_SPURIOUS,
 };
 
+struct tdp_mmu {
+       /* The number of TDP MMU pages across all roots. */
+       atomic64_t pages;
+
+       /*
+        * List of kvm_mmu_page structs being used as roots.
+        * All kvm_mmu_page structs in the list should have
+        * tdp_mmu_page set.
+        *
+        * For reads, this list is protected by:
+        *      the MMU lock in read mode + RCU or
+        *      the MMU lock in write mode
+        *
+        * For writes, this list is protected by:
+        *      the MMU lock in read mode + the tdp_mmu_pages_lock or
+        *      the MMU lock in write mode
+        *
+        * Roots will remain in the list until their tdp_mmu_root_count
+        * drops to zero, at which point the thread that decremented the
+        * count to zero should removed the root from the list and clean
+        * it up, freeing the root after an RCU grace period.
+        */
+       struct list_head roots;
+
+       /*
+        * Protects accesses to the following fields when the MMU lock
+        * is held in read mode:
+        *  - roots (above)
+        *  - the link field of kvm_mmu_page structs used by the TDP MMU
+        *  - (x86-only) possible_nx_huge_pages;
+        *  - (x86-only) the arch.possible_nx_huge_page_link field of
+        *    kvm_mmu_page structs used by the TDP MMU
+        * It is acceptable, but not necessary, to acquire this lock when
+        * the thread holds the MMU lock in write mode.
+        */
+       spinlock_t pages_lock;
+
+       struct workqueue_struct *zap_wq;
+};
+
 #endif /* !__KVM_MMU_TYPES_H */
diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h
index 469ff4202a0d..242eaed55320 100644
--- a/include/linux/kvm_host.h
+++ b/include/linux/kvm_host.h
@@ -45,6 +45,10 @@
 #include <asm/kvm_host.h>
 #include <linux/kvm_dirty_ring.h>
 
+#ifdef CONFIG_HAVE_TDP_MMU
+#include <kvm/mmu_types.h>
+#endif
+
 #ifndef KVM_MAX_VCPU_IDS
 #define KVM_MAX_VCPU_IDS KVM_MAX_VCPUS
 #endif
@@ -797,6 +801,10 @@ struct kvm {
        struct notifier_block pm_notifier;
 #endif
        char stats_id[KVM_STATS_NAME_SIZE];
+
+#ifdef CONFIG_HAVE_TDP_MMU
+       struct tdp_mmu tdp_mmu;
+#endif
 };
 
 #define kvm_err(fmt, ...) \
-- 
2.39.0.rc1.256.g54fd8350bd-goog

_______________________________________________
kvmarm mailing list
kvmarm@lists.cs.columbia.edu
https://lists.cs.columbia.edu/mailman/listinfo/kvmarm

Reply via email to