--- a/arch/x86/include/asm/msr-index.h
+++ b/arch/x86/include/asm/msr-index.h
-@@ -389,6 +389,11 @@
+@@ -399,6 +399,11 @@
#define MSR_AMD64_OSVW_STATUS 0xc0010141
#define MSR_AMD64_LS_CFG 0xc0011020
#define MSR_AMD64_DC_CFG 0xc0011022
#define MSR_AMD64_BU_CFG2 0xc001102a
#define MSR_AMD64_IBSFETCHCTL 0xc0011030
#define MSR_AMD64_IBSFETCHLINAD 0xc0011031
-@@ -457,9 +462,6 @@
+@@ -467,9 +472,6 @@
#define FAM10H_MMIO_CONF_BASE_MASK 0xfffffffULL
#define FAM10H_MMIO_CONF_BASE_SHIFT 20
#define MSR_FAM10H_NODE_ID 0xc001100c
static void init_amd_ln(struct cpuinfo_x86 *c)
{
/*
-@@ -951,16 +949,16 @@ static void init_amd(struct cpuinfo_x86
+@@ -960,16 +958,16 @@ static void init_amd(struct cpuinfo_x86
* msr_set_bit() uses the safe accessors, too, even if the MSR
* is not present.
*/
} else {
--- a/arch/x86/kvm/svm.c
+++ b/arch/x86/kvm/svm.c
-@@ -4154,9 +4154,9 @@ static int svm_get_msr_feature(struct kv
+@@ -4155,9 +4155,9 @@ static int svm_get_msr_feature(struct kv
msr->data = 0;
switch (msr->index) {
break;
default:
return 1;
-@@ -4258,7 +4258,7 @@ static int svm_get_msr(struct kvm_vcpu *
+@@ -4259,7 +4259,7 @@ static int svm_get_msr(struct kvm_vcpu *
msr_info->data = 0x1E;
}
break;
msr_info->data = svm->msr_decfg;
break;
default:
-@@ -4445,7 +4445,7 @@ static int svm_set_msr(struct kvm_vcpu *
+@@ -4446,7 +4446,7 @@ static int svm_set_msr(struct kvm_vcpu *
case MSR_VM_IGNNE:
vcpu_unimpl(vcpu, "unimplemented wrmsr: 0x%x data 0x%llx\n", ecx, data);
break;