]> git.ipfire.org Git - thirdparty/kernel/stable.git/commitdiff
KVM: VMX: Disable preemption when probing user return MSRs
authorSean Christopherson <seanjc@google.com>
Tue, 4 May 2021 17:17:24 +0000 (10:17 -0700)
committerGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Wed, 19 May 2021 08:29:56 +0000 (10:29 +0200)
commit 5104d7ffcf24749939bea7fdb5378d186473f890 upstream.

Disable preemption when probing a user return MSR via RDSMR/WRMSR.  If
the MSR holds a different value per logical CPU, the WRMSR could corrupt
the host's value if KVM is preempted between the RDMSR and WRMSR, and
then rescheduled on a different CPU.

Opportunistically land the helper in common x86, SVM will use the helper
in a future commit.

Fixes: 4be534102624 ("KVM: VMX: Initialize vmx->guest_msrs[] right after allocation")
Cc: stable@vger.kernel.org
Cc: Xiaoyao Li <xiaoyao.li@intel.com>
Signed-off-by: Sean Christopherson <seanjc@google.com>
Message-Id: <20210504171734.1434054-6-seanjc@google.com>
Reviewed-by: Jim Mattson <jmattson@google.com>
Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
arch/x86/include/asm/kvm_host.h
arch/x86/kvm/vmx/vmx.c
arch/x86/kvm/x86.c

index 90ebe458ab7bbb2deb56af859665b9025d5e5df4..d5b365e670ac005813ee98c9db88fd8afcf2dbb4 100644 (file)
@@ -1694,6 +1694,7 @@ int kvm_pv_send_ipi(struct kvm *kvm, unsigned long ipi_bitmap_low,
                    unsigned long icr, int op_64_bit);
 
 void kvm_define_user_return_msr(unsigned index, u32 msr);
+int kvm_probe_user_return_msr(u32 msr);
 int kvm_set_user_return_msr(unsigned index, u64 val, u64 mask);
 
 u64 kvm_scale_tsc(struct kvm_vcpu *vcpu, u64 tsc);
index 40e7a91eebac78505e81c3e7ed544bc549c24c03..d3ec6ba3acb5cf22026f7e002629f76c26cd3347 100644 (file)
@@ -6850,12 +6850,9 @@ static int vmx_create_vcpu(struct kvm_vcpu *vcpu)
 
        for (i = 0; i < ARRAY_SIZE(vmx_uret_msrs_list); ++i) {
                u32 index = vmx_uret_msrs_list[i];
-               u32 data_low, data_high;
                int j = vmx->nr_uret_msrs;
 
-               if (rdmsr_safe(index, &data_low, &data_high) < 0)
-                       continue;
-               if (wrmsr_safe(index, data_low, data_high) < 0)
+               if (kvm_probe_user_return_msr(index))
                        continue;
 
                vmx->guest_uret_msrs[j].slot = i;
index 8105e9ae1ff8186285e04d76ad76ae8a6fd6f9de..95e28358f443a783a6834556520a41dace190dc0 100644 (file)
@@ -323,6 +323,22 @@ static void kvm_on_user_return(struct user_return_notifier *urn)
        }
 }
 
+int kvm_probe_user_return_msr(u32 msr)
+{
+       u64 val;
+       int ret;
+
+       preempt_disable();
+       ret = rdmsrl_safe(msr, &val);
+       if (ret)
+               goto out;
+       ret = wrmsrl_safe(msr, val);
+out:
+       preempt_enable();
+       return ret;
+}
+EXPORT_SYMBOL_GPL(kvm_probe_user_return_msr);
+
 void kvm_define_user_return_msr(unsigned slot, u32 msr)
 {
        BUG_ON(slot >= KVM_MAX_NR_USER_RETURN_MSRS);