]> git.baikalelectronics.ru Git - kernel.git/commitdiff
KVM: x86/mmu: Rename pte_list_{destroy,remove}() to show they zap SPTEs
authorSean Christopherson <seanjc@google.com>
Fri, 15 Jul 2022 22:42:25 +0000 (22:42 +0000)
committerPaolo Bonzini <pbonzini@redhat.com>
Thu, 28 Jul 2022 17:22:18 +0000 (13:22 -0400)
Rename pte_list_remove() and pte_list_destroy() to kvm_zap_one_rmap_spte()
and kvm_zap_all_rmap_sptes() respectively to document that (a) they zap
SPTEs and (b) to better document how they differ (remove vs. destroy does
not exactly scream "one vs. all").

No functional change intended.

Signed-off-by: Sean Christopherson <seanjc@google.com>
Message-Id: <20220715224226.3749507-7-seanjc@google.com>
Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
arch/x86/kvm/mmu/mmu.c

index 00be88e0a5f7f5282aecd62e16b33addee8b60c8..282e7e2ab446759f0edc08628fb88ac2bd234599 100644 (file)
@@ -957,15 +957,16 @@ static void __pte_list_remove(u64 *spte, struct kvm_rmap_head *rmap_head)
        }
 }
 
-static void pte_list_remove(struct kvm *kvm, struct kvm_rmap_head *rmap_head,
-                           u64 *sptep)
+static void kvm_zap_one_rmap_spte(struct kvm *kvm,
+                                 struct kvm_rmap_head *rmap_head, u64 *sptep)
 {
        mmu_spte_clear_track_bits(kvm, sptep);
        __pte_list_remove(sptep, rmap_head);
 }
 
-/* Return true if rmap existed, false otherwise */
-static bool pte_list_destroy(struct kvm *kvm, struct kvm_rmap_head *rmap_head)
+/* Return true if at least one SPTE was zapped, false otherwise */
+static bool kvm_zap_all_rmap_sptes(struct kvm *kvm,
+                                  struct kvm_rmap_head *rmap_head)
 {
        struct pte_list_desc *desc, *next;
        int i;
@@ -1386,7 +1387,7 @@ static bool kvm_vcpu_write_protect_gfn(struct kvm_vcpu *vcpu, u64 gfn)
 static bool __kvm_zap_rmap(struct kvm *kvm, struct kvm_rmap_head *rmap_head,
                           const struct kvm_memory_slot *slot)
 {
-       return pte_list_destroy(kvm, rmap_head);
+       return kvm_zap_all_rmap_sptes(kvm, rmap_head);
 }
 
 static bool kvm_zap_rmap(struct kvm *kvm, struct kvm_rmap_head *rmap_head,
@@ -1417,7 +1418,7 @@ restart:
                need_flush = true;
 
                if (pte_write(pte)) {
-                       pte_list_remove(kvm, rmap_head, sptep);
+                       kvm_zap_one_rmap_spte(kvm, rmap_head, sptep);
                        goto restart;
                } else {
                        new_spte = kvm_mmu_changed_pte_notifier_make_spte(
@@ -1596,7 +1597,7 @@ static void __rmap_add(struct kvm *kvm,
        rmap_count = pte_list_add(cache, spte, rmap_head);
 
        if (rmap_count > RMAP_RECYCLE_THRESHOLD) {
-               pte_list_destroy(kvm, rmap_head);
+               kvm_zap_all_rmap_sptes(kvm, rmap_head);
                kvm_flush_remote_tlbs_with_address(
                                kvm, sp->gfn, KVM_PAGES_PER_HPAGE(sp->role.level));
        }
@@ -6406,7 +6407,7 @@ restart:
                if (sp->role.direct &&
                    sp->role.level < kvm_mmu_max_mapping_level(kvm, slot, sp->gfn,
                                                               pfn, PG_LEVEL_NUM)) {
-                       pte_list_remove(kvm, rmap_head, sptep);
+                       kvm_zap_one_rmap_spte(kvm, rmap_head, sptep);
 
                        if (kvm_available_flush_tlb_with_range())
                                kvm_flush_remote_tlbs_with_address(kvm, sp->gfn,