Loading arch/x86/kvm/mmu.c +2 −2 Original line number Diff line number Diff line Loading @@ -2781,7 +2781,7 @@ static bool handle_abnormal_pfn(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn, return ret; } static bool page_fault_can_be_fast(struct kvm_vcpu *vcpu, u32 error_code) static bool page_fault_can_be_fast(u32 error_code) { /* * Do not fix the mmio spte with invalid generation number which Loading Loading @@ -2834,7 +2834,7 @@ static bool fast_page_fault(struct kvm_vcpu *vcpu, gva_t gva, int level, bool ret = false; u64 spte = 0ull; if (!page_fault_can_be_fast(vcpu, error_code)) if (!page_fault_can_be_fast(error_code)) return false; walk_shadow_page_lockless_begin(vcpu); Loading Loading
arch/x86/kvm/mmu.c +2 −2 Original line number Diff line number Diff line Loading @@ -2781,7 +2781,7 @@ static bool handle_abnormal_pfn(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn, return ret; } static bool page_fault_can_be_fast(struct kvm_vcpu *vcpu, u32 error_code) static bool page_fault_can_be_fast(u32 error_code) { /* * Do not fix the mmio spte with invalid generation number which Loading Loading @@ -2834,7 +2834,7 @@ static bool fast_page_fault(struct kvm_vcpu *vcpu, gva_t gva, int level, bool ret = false; u64 spte = 0ull; if (!page_fault_can_be_fast(vcpu, error_code)) if (!page_fault_can_be_fast(error_code)) return false; walk_shadow_page_lockless_begin(vcpu); Loading