Loading arch/ia64/kvm/kvm-ia64.c +5 −0 Original line number Diff line number Diff line Loading @@ -1330,6 +1330,11 @@ int kvm_arch_vcpu_setup(struct kvm_vcpu *vcpu) return 0; } int kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu) { return 0; } int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu) { return -EINVAL; Loading arch/powerpc/kvm/powerpc.c +5 −0 Original line number Diff line number Diff line Loading @@ -439,6 +439,11 @@ struct kvm_vcpu *kvm_arch_vcpu_create(struct kvm *kvm, unsigned int id) return vcpu; } int kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu) { return 0; } void kvm_arch_vcpu_free(struct kvm_vcpu *vcpu) { /* Make sure we're not using the vcpu anymore */ Loading arch/s390/kvm/kvm-s390.c +5 −0 Original line number Diff line number Diff line Loading @@ -355,6 +355,11 @@ static void kvm_s390_vcpu_initial_reset(struct kvm_vcpu *vcpu) atomic_set_mask(CPUSTAT_STOPPED, &vcpu->arch.sie_block->cpuflags); } int kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu) { return 0; } int kvm_arch_vcpu_setup(struct kvm_vcpu *vcpu) { atomic_set(&vcpu->arch.sie_block->cpuflags, CPUSTAT_ZARCH | Loading arch/x86/kvm/svm.c +0 −1 Original line number Diff line number Diff line Loading @@ -1251,7 +1251,6 @@ static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id) svm->vmcb_pa = page_to_pfn(page) << PAGE_SHIFT; svm->asid_generation = 0; init_vmcb(svm); kvm_write_tsc(&svm->vcpu, 0); err = fx_init(&svm->vcpu); if (err) Loading arch/x86/kvm/vmx.c +0 −2 Original line number Diff line number Diff line Loading @@ -3896,8 +3896,6 @@ static int vmx_vcpu_setup(struct vcpu_vmx *vmx) vmcs_writel(CR0_GUEST_HOST_MASK, ~0UL); set_cr4_guest_host_mask(vmx); kvm_write_tsc(&vmx->vcpu, 0); return 0; } Loading Loading
arch/ia64/kvm/kvm-ia64.c +5 −0 Original line number Diff line number Diff line Loading @@ -1330,6 +1330,11 @@ int kvm_arch_vcpu_setup(struct kvm_vcpu *vcpu) return 0; } int kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu) { return 0; } int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu) { return -EINVAL; Loading
arch/powerpc/kvm/powerpc.c +5 −0 Original line number Diff line number Diff line Loading @@ -439,6 +439,11 @@ struct kvm_vcpu *kvm_arch_vcpu_create(struct kvm *kvm, unsigned int id) return vcpu; } int kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu) { return 0; } void kvm_arch_vcpu_free(struct kvm_vcpu *vcpu) { /* Make sure we're not using the vcpu anymore */ Loading
arch/s390/kvm/kvm-s390.c +5 −0 Original line number Diff line number Diff line Loading @@ -355,6 +355,11 @@ static void kvm_s390_vcpu_initial_reset(struct kvm_vcpu *vcpu) atomic_set_mask(CPUSTAT_STOPPED, &vcpu->arch.sie_block->cpuflags); } int kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu) { return 0; } int kvm_arch_vcpu_setup(struct kvm_vcpu *vcpu) { atomic_set(&vcpu->arch.sie_block->cpuflags, CPUSTAT_ZARCH | Loading
arch/x86/kvm/svm.c +0 −1 Original line number Diff line number Diff line Loading @@ -1251,7 +1251,6 @@ static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id) svm->vmcb_pa = page_to_pfn(page) << PAGE_SHIFT; svm->asid_generation = 0; init_vmcb(svm); kvm_write_tsc(&svm->vcpu, 0); err = fx_init(&svm->vcpu); if (err) Loading
arch/x86/kvm/vmx.c +0 −2 Original line number Diff line number Diff line Loading @@ -3896,8 +3896,6 @@ static int vmx_vcpu_setup(struct vcpu_vmx *vmx) vmcs_writel(CR0_GUEST_HOST_MASK, ~0UL); set_cr4_guest_host_mask(vmx); kvm_write_tsc(&vmx->vcpu, 0); return 0; } Loading