Loading arch/x86/kvm/mmu.c +8 −9 Original line number Diff line number Diff line Loading @@ -679,7 +679,7 @@ static void rmap_remove(struct kvm *kvm, u64 *spte) } } static void drop_spte(struct kvm *kvm, u64 *sptep, u64 new_spte) static void set_spte_track_bits(u64 *sptep, u64 new_spte) { pfn_t pfn; u64 old_spte; Loading @@ -692,6 +692,11 @@ static void drop_spte(struct kvm *kvm, u64 *sptep, u64 new_spte) kvm_set_pfn_accessed(pfn); if (is_writable_pte(old_spte)) kvm_set_pfn_dirty(pfn); } static void drop_spte(struct kvm *kvm, u64 *sptep, u64 new_spte) { set_spte_track_bits(sptep, new_spte); rmap_remove(kvm, sptep); } Loading Loading @@ -791,7 +796,7 @@ static int kvm_set_pte_rmapp(struct kvm *kvm, unsigned long *rmapp, unsigned long data) { int need_flush = 0; u64 *spte, new_spte, old_spte; u64 *spte, new_spte; pte_t *ptep = (pte_t *)data; pfn_t new_pfn; Loading @@ -812,13 +817,7 @@ static int kvm_set_pte_rmapp(struct kvm *kvm, unsigned long *rmapp, new_spte &= ~PT_WRITABLE_MASK; new_spte &= ~SPTE_HOST_WRITEABLE; new_spte &= ~shadow_accessed_mask; if (is_writable_pte(*spte)) kvm_set_pfn_dirty(spte_to_pfn(*spte)); old_spte = __xchg_spte(spte, new_spte); if (is_shadow_present_pte(old_spte) && (!shadow_accessed_mask || old_spte & shadow_accessed_mask)) mark_page_accessed(pfn_to_page(spte_to_pfn(old_spte))); set_spte_track_bits(spte, new_spte); spte = rmap_next(kvm, rmapp, spte); } } Loading Loading
arch/x86/kvm/mmu.c +8 −9 Original line number Diff line number Diff line Loading @@ -679,7 +679,7 @@ static void rmap_remove(struct kvm *kvm, u64 *spte) } } static void drop_spte(struct kvm *kvm, u64 *sptep, u64 new_spte) static void set_spte_track_bits(u64 *sptep, u64 new_spte) { pfn_t pfn; u64 old_spte; Loading @@ -692,6 +692,11 @@ static void drop_spte(struct kvm *kvm, u64 *sptep, u64 new_spte) kvm_set_pfn_accessed(pfn); if (is_writable_pte(old_spte)) kvm_set_pfn_dirty(pfn); } static void drop_spte(struct kvm *kvm, u64 *sptep, u64 new_spte) { set_spte_track_bits(sptep, new_spte); rmap_remove(kvm, sptep); } Loading Loading @@ -791,7 +796,7 @@ static int kvm_set_pte_rmapp(struct kvm *kvm, unsigned long *rmapp, unsigned long data) { int need_flush = 0; u64 *spte, new_spte, old_spte; u64 *spte, new_spte; pte_t *ptep = (pte_t *)data; pfn_t new_pfn; Loading @@ -812,13 +817,7 @@ static int kvm_set_pte_rmapp(struct kvm *kvm, unsigned long *rmapp, new_spte &= ~PT_WRITABLE_MASK; new_spte &= ~SPTE_HOST_WRITEABLE; new_spte &= ~shadow_accessed_mask; if (is_writable_pte(*spte)) kvm_set_pfn_dirty(spte_to_pfn(*spte)); old_spte = __xchg_spte(spte, new_spte); if (is_shadow_present_pte(old_spte) && (!shadow_accessed_mask || old_spte & shadow_accessed_mask)) mark_page_accessed(pfn_to_page(spte_to_pfn(old_spte))); set_spte_track_bits(spte, new_spte); spte = rmap_next(kvm, rmapp, spte); } } Loading