@@ -1272,45 +1272,6 @@ struct kvm_arch {
struct kvm_pmu_event_filter __rcu *pmu_event_filter;
struct task_struct *nx_huge_page_recovery_thread;
-#ifdef CONFIG_X86_64
- /* The number of TDP MMU pages across all roots. */
- atomic64_t tdp_mmu_pages;
-
- /*
- * List of struct kvm_mmu_pages being used as roots.
- * All struct kvm_mmu_pages in the list should have
- * tdp_mmu_page set.
- *
- * For reads, this list is protected by:
- * the MMU lock in read mode + RCU or
- * the MMU lock in write mode
- *
- * For writes, this list is protected by:
- * the MMU lock in read mode + the tdp_mmu_pages_lock or
- * the MMU lock in write mode
- *
- * Roots will remain in the list until their tdp_mmu_root_count
- * drops to zero, at which point the thread that decremented the
- * count to zero should removed the root from the list and clean
- * it up, freeing the root after an RCU grace period.
- */
- struct list_head tdp_mmu_roots;
-
- /*
- * Protects accesses to the following fields when the MMU lock
- * is held in read mode:
- * - tdp_mmu_roots (above)
- * - the link field of kvm_mmu_page structs used by the TDP MMU
- * - possible_nx_huge_pages;
- * - the possible_nx_huge_page_link field of kvm_mmu_page structs used
- * by the TDP MMU
- * It is acceptable, but not necessary, to acquire this lock when
- * the thread holds the MMU lock in write mode.
- */
- spinlock_t tdp_mmu_pages_lock;
- struct workqueue_struct *tdp_mmu_zap_wq;
-#endif /* CONFIG_X86_64 */
-
/*
* If set, at least one shadow root has been allocated. This flag
* is used as one input when determining whether certain memslot
@@ -21,9 +21,9 @@ int kvm_mmu_init_tdp_mmu(struct kvm *kvm)
if (!wq)
return -ENOMEM;
- INIT_LIST_HEAD(&kvm->arch.tdp_mmu_roots);
- spin_lock_init(&kvm->arch.tdp_mmu_pages_lock);
- kvm->arch.tdp_mmu_zap_wq = wq;
+ INIT_LIST_HEAD(&kvm->tdp_mmu.roots);
+ spin_lock_init(&kvm->tdp_mmu.pages_lock);
+ kvm->tdp_mmu.zap_wq = wq;
return 1;
}
@@ -42,10 +42,10 @@ static __always_inline bool kvm_lockdep_assert_mmu_lock_held(struct kvm *kvm,
void kvm_mmu_uninit_tdp_mmu(struct kvm *kvm)
{
/* Also waits for any queued work items. */
- destroy_workqueue(kvm->arch.tdp_mmu_zap_wq);
+ destroy_workqueue(kvm->tdp_mmu.zap_wq);
- WARN_ON(atomic64_read(&kvm->arch.tdp_mmu_pages));
- WARN_ON(!list_empty(&kvm->arch.tdp_mmu_roots));
+ WARN_ON(atomic64_read(&kvm->tdp_mmu.pages));
+ WARN_ON(!list_empty(&kvm->tdp_mmu.roots));
/*
* Ensure that all the outstanding RCU callbacks to free shadow pages
@@ -114,7 +114,7 @@ static void tdp_mmu_schedule_zap_root(struct kvm *kvm, struct kvm_mmu_page *root
{
root->tdp_mmu_async_data = kvm;
INIT_WORK(&root->tdp_mmu_async_work, tdp_mmu_zap_root_work);
- queue_work(kvm->arch.tdp_mmu_zap_wq, &root->tdp_mmu_async_work);
+ queue_work(kvm->tdp_mmu.zap_wq, &root->tdp_mmu_async_work);
}
static inline bool kvm_tdp_root_mark_invalid(struct kvm_mmu_page *page)
@@ -173,9 +173,9 @@ void kvm_tdp_mmu_put_root(struct kvm *kvm, struct kvm_mmu_page *root,
return;
}
- spin_lock(&kvm->arch.tdp_mmu_pages_lock);
+ spin_lock(&kvm->tdp_mmu.pages_lock);
list_del_rcu(&root->link);
- spin_unlock(&kvm->arch.tdp_mmu_pages_lock);
+ spin_unlock(&kvm->tdp_mmu.pages_lock);
call_rcu(&root->rcu_head, tdp_mmu_free_sp_rcu_callback);
}
@@ -198,11 +198,11 @@ static struct kvm_mmu_page *tdp_mmu_next_root(struct kvm *kvm,
rcu_read_lock();
if (prev_root)
- next_root = list_next_or_null_rcu(&kvm->arch.tdp_mmu_roots,
+ next_root = list_next_or_null_rcu(&kvm->tdp_mmu.roots,
&prev_root->link,
typeof(*prev_root), link);
else
- next_root = list_first_or_null_rcu(&kvm->arch.tdp_mmu_roots,
+ next_root = list_first_or_null_rcu(&kvm->tdp_mmu.roots,
typeof(*next_root), link);
while (next_root) {
@@ -210,7 +210,7 @@ static struct kvm_mmu_page *tdp_mmu_next_root(struct kvm *kvm,
kvm_tdp_mmu_get_root(next_root))
break;
- next_root = list_next_or_null_rcu(&kvm->arch.tdp_mmu_roots,
+ next_root = list_next_or_null_rcu(&kvm->tdp_mmu.roots,
&next_root->link, typeof(*next_root), link);
}
@@ -254,7 +254,7 @@ static struct kvm_mmu_page *tdp_mmu_next_root(struct kvm *kvm,
* is guaranteed to be stable.
*/
#define for_each_tdp_mmu_root(_kvm, _root, _as_id) \
- list_for_each_entry(_root, &_kvm->arch.tdp_mmu_roots, link) \
+ list_for_each_entry(_root, &_kvm->tdp_mmu.roots, link) \
if (kvm_lockdep_assert_mmu_lock_held(_kvm, false) && \
_root->role.as_id != _as_id) { \
} else
@@ -324,9 +324,9 @@ hpa_t kvm_tdp_mmu_get_vcpu_root_hpa(struct kvm_vcpu *vcpu)
refcount_set(&root->root_refcount, 1);
- spin_lock(&kvm->arch.tdp_mmu_pages_lock);
- list_add_rcu(&root->link, &kvm->arch.tdp_mmu_roots);
- spin_unlock(&kvm->arch.tdp_mmu_pages_lock);
+ spin_lock(&kvm->tdp_mmu.pages_lock);
+ list_add_rcu(&root->link, &kvm->tdp_mmu.roots);
+ spin_unlock(&kvm->tdp_mmu.pages_lock);
out:
return __pa(root->spt);
@@ -368,13 +368,13 @@ static void handle_changed_spte_dirty_log(struct kvm *kvm, int as_id, gfn_t gfn,
static void tdp_account_mmu_page(struct kvm *kvm, struct kvm_mmu_page *sp)
{
kvm_account_pgtable_pages((void *)sp->spt, +1);
- atomic64_inc(&kvm->arch.tdp_mmu_pages);
+ atomic64_inc(&kvm->tdp_mmu.pages);
}
static void tdp_unaccount_mmu_page(struct kvm *kvm, struct kvm_mmu_page *sp)
{
kvm_account_pgtable_pages((void *)sp->spt, -1);
- atomic64_dec(&kvm->arch.tdp_mmu_pages);
+ atomic64_dec(&kvm->tdp_mmu.pages);
}
__weak void tdp_mmu_arch_unlink_sp(struct kvm *kvm, struct kvm_mmu_page *sp,
@@ -1010,7 +1010,7 @@ void kvm_tdp_mmu_zap_all(struct kvm *kvm)
*/
void kvm_tdp_mmu_zap_invalidated_roots(struct kvm *kvm)
{
- flush_workqueue(kvm->arch.tdp_mmu_zap_wq);
+ flush_workqueue(kvm->tdp_mmu.zap_wq);
}
/*
@@ -1035,7 +1035,7 @@ void kvm_tdp_mmu_invalidate_all_roots(struct kvm *kvm)
struct kvm_mmu_page *root;
lockdep_assert_held_write(&kvm->mmu_lock);
- list_for_each_entry(root, &kvm->arch.tdp_mmu_roots, link) {
+ list_for_each_entry(root, &kvm->tdp_mmu.roots, link) {
if (!root->role.invalid &&
!WARN_ON_ONCE(!kvm_tdp_mmu_get_root(root))) {
root->role.invalid = true;
@@ -141,9 +141,9 @@ void tdp_mmu_arch_post_link_sp(struct kvm *kvm,
if (fault->req_level < sp->role.level)
return;
- spin_lock(&kvm->arch.tdp_mmu_pages_lock);
+ spin_lock(&kvm->tdp_mmu.pages_lock);
track_possible_nx_huge_page(kvm, sp);
- spin_unlock(&kvm->arch.tdp_mmu_pages_lock);
+ spin_unlock(&kvm->tdp_mmu.pages_lock);
}
void tdp_mmu_arch_unlink_sp(struct kvm *kvm, struct kvm_mmu_page *sp,
@@ -153,7 +153,7 @@ void tdp_mmu_arch_unlink_sp(struct kvm *kvm, struct kvm_mmu_page *sp,
return;
if (shared)
- spin_lock(&kvm->arch.tdp_mmu_pages_lock);
+ spin_lock(&kvm->tdp_mmu.pages_lock);
else
lockdep_assert_held_write(&kvm->mmu_lock);
@@ -161,7 +161,7 @@ void tdp_mmu_arch_unlink_sp(struct kvm *kvm, struct kvm_mmu_page *sp,
untrack_possible_nx_huge_page(kvm, sp);
if (shared)
- spin_unlock(&kvm->arch.tdp_mmu_pages_lock);
+ spin_unlock(&kvm->tdp_mmu.pages_lock);
}
int tdp_mmu_max_mapping_level(struct kvm *kvm,
@@ -136,4 +136,44 @@ enum {
RET_PF_SPURIOUS,
};
+struct tdp_mmu {
+ /* The number of TDP MMU pages across all roots. */
+ atomic64_t pages;
+
+ /*
+ * List of kvm_mmu_page structs being used as roots.
+ * All kvm_mmu_page structs in the list should have
+ * tdp_mmu_page set.
+ *
+ * For reads, this list is protected by:
+ * the MMU lock in read mode + RCU or
+ * the MMU lock in write mode
+ *
+ * For writes, this list is protected by:
+ * the MMU lock in read mode + the tdp_mmu_pages_lock or
+ * the MMU lock in write mode
+ *
+ * Roots will remain in the list until their tdp_mmu_root_count
+ * drops to zero, at which point the thread that decremented the
+ * count to zero should removed the root from the list and clean
+ * it up, freeing the root after an RCU grace period.
+ */
+ struct list_head roots;
+
+ /*
+ * Protects accesses to the following fields when the MMU lock
+ * is held in read mode:
+ * - roots (above)
+ * - the link field of kvm_mmu_page structs used by the TDP MMU
+ * - (x86-only) possible_nx_huge_pages;
+ * - (x86-only) the arch.possible_nx_huge_page_link field of
+ * kvm_mmu_page structs used by the TDP MMU
+ * It is acceptable, but not necessary, to acquire this lock when
+ * the thread holds the MMU lock in write mode.
+ */
+ spinlock_t pages_lock;
+
+ struct workqueue_struct *zap_wq;
+};
+
#endif /* !__KVM_MMU_TYPES_H */
@@ -45,6 +45,10 @@
#include <asm/kvm_host.h>
#include <linux/kvm_dirty_ring.h>
+#ifdef CONFIG_HAVE_TDP_MMU
+#include <kvm/mmu_types.h>
+#endif
+
#ifndef KVM_MAX_VCPU_IDS
#define KVM_MAX_VCPU_IDS KVM_MAX_VCPUS
#endif
@@ -797,6 +801,10 @@ struct kvm {
struct notifier_block pm_notifier;
#endif
char stats_id[KVM_STATS_NAME_SIZE];
+
+#ifdef CONFIG_HAVE_TDP_MMU
+ struct tdp_mmu tdp_mmu;
+#endif
};
#define kvm_err(fmt, ...) \
Move VM-level TDP MMU state to struct kvm so it can be accessed by common code in a future commit. No functional change intended. Signed-off-by: David Matlack <dmatlack@google.com> --- arch/x86/include/asm/kvm_host.h | 39 -------------------------------- arch/x86/kvm/mmu/tdp_mmu.c | 40 ++++++++++++++++----------------- arch/x86/kvm/mmu/tdp_pgtable.c | 8 +++---- include/kvm/mmu_types.h | 40 +++++++++++++++++++++++++++++++++ include/linux/kvm_host.h | 8 +++++++ 5 files changed, 72 insertions(+), 63 deletions(-)