KVM: x86/mmu: Move flush logic from mmu_page_zap_pte() to FNAME(invlpg)
authorSean Christopherson <sean.j.christopherson@intel.com>
Wed, 23 Sep 2020 22:14:05 +0000 (15:14 -0700)
committerPaolo Bonzini <pbonzini@redhat.com>
Mon, 28 Sep 2020 11:57:34 +0000 (07:57 -0400)
Move the logic that controls whether or not FNAME(invlpg) needs to flush
fully into FNAME(invlpg) so that mmu_page_zap_pte() doesn't return a
value.  This allows a future patch to redefine the return semantics for
mmu_page_zap_pte() so that it can recursively zap orphaned child shadow
pages for nested TDP MMUs.

No functional change intended.

Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com>
Message-Id: <20200923221406.16297-2-sean.j.christopherson@intel.com>
Reviewed-by: Ben Gardon <bgardon@google.com>
Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
arch/x86/kvm/mmu/mmu.c
arch/x86/kvm/mmu/paging_tmpl.h

index 5950dc92b583d5e44e06bf0535983b343f9bb217..0122e68ac0abf9b6ac5b797fe89b2cc72a1baf24 100644 (file)
@@ -2615,7 +2615,7 @@ static void validate_direct_spte(struct kvm_vcpu *vcpu, u64 *sptep,
        }
 }
 
-static bool mmu_page_zap_pte(struct kvm *kvm, struct kvm_mmu_page *sp,
+static void mmu_page_zap_pte(struct kvm *kvm, struct kvm_mmu_page *sp,
                             u64 *spte)
 {
        u64 pte;
@@ -2631,13 +2631,9 @@ static bool mmu_page_zap_pte(struct kvm *kvm, struct kvm_mmu_page *sp,
                        child = to_shadow_page(pte & PT64_BASE_ADDR_MASK);
                        drop_parent_pte(child, spte);
                }
-               return true;
-       }
-
-       if (is_mmio_spte(pte))
+       } else if (is_mmio_spte(pte)) {
                mmu_spte_clear_no_track(spte);
-
-       return false;
+       }
 }
 
 static void kvm_mmu_page_unlink_children(struct kvm *kvm,
index 4dd6b1e5b8cf7238c038c2842b2a0fe4edcb18dc..3bb624a3dda92d5015b90909de061ffde82fbcdb 100644 (file)
@@ -895,6 +895,7 @@ static void FNAME(invlpg)(struct kvm_vcpu *vcpu, gva_t gva, hpa_t root_hpa)
 {
        struct kvm_shadow_walk_iterator iterator;
        struct kvm_mmu_page *sp;
+       u64 old_spte;
        int level;
        u64 *sptep;
 
@@ -917,7 +918,8 @@ static void FNAME(invlpg)(struct kvm_vcpu *vcpu, gva_t gva, hpa_t root_hpa)
                sptep = iterator.sptep;
 
                sp = sptep_to_sp(sptep);
-               if (is_last_spte(*sptep, level)) {
+               old_spte = *sptep;
+               if (is_last_spte(old_spte, level)) {
                        pt_element_t gpte;
                        gpa_t pte_gpa;
 
@@ -927,7 +929,8 @@ static void FNAME(invlpg)(struct kvm_vcpu *vcpu, gva_t gva, hpa_t root_hpa)
                        pte_gpa = FNAME(get_level1_sp_gpa)(sp);
                        pte_gpa += (sptep - sp->spt) * sizeof(pt_element_t);
 
-                       if (mmu_page_zap_pte(vcpu->kvm, sp, sptep))
+                       mmu_page_zap_pte(vcpu->kvm, sp, sptep);
+                       if (is_shadow_present_pte(old_spte))
                                kvm_flush_remote_tlbs_with_address(vcpu->kvm,
                                        sp->gfn, KVM_PAGES_PER_HPAGE(sp->role.level));