From: Greg Kroah-Hartman Date: Thu, 14 Jul 2022 06:44:41 +0000 (+0200) Subject: 5.10-stable patches X-Git-Tag: v5.18.12~16 X-Git-Url: http://git.ipfire.org/?a=commitdiff_plain;h=a0ac8f6ea344fdf826a53248493af3e686e61caa;p=thirdparty%2Fkernel%2Fstable-queue.git 5.10-stable patches added patches: x86-kvm-use-proper-asm-macros-for-kvm_vcpu_is_preempted.patch --- diff --git a/queue-5.10/series b/queue-5.10/series index 84af897df89..57fa6e8ec08 100644 --- a/queue-5.10/series +++ b/queue-5.10/series @@ -129,3 +129,4 @@ x86-kexec-disable-ret-on-kexec.patch x86-speculation-disable-rrsba-behavior.patch x86-static_call-serialize-__static_call_fixup-properly.patch tools-insn-restore-the-relative-include-paths-for-cross-building.patch +x86-kvm-use-proper-asm-macros-for-kvm_vcpu_is_preempted.patch diff --git a/queue-5.10/x86-kvm-use-proper-asm-macros-for-kvm_vcpu_is_preempted.patch b/queue-5.10/x86-kvm-use-proper-asm-macros-for-kvm_vcpu_is_preempted.patch new file mode 100644 index 00000000000..3750ee432b3 --- /dev/null +++ b/queue-5.10/x86-kvm-use-proper-asm-macros-for-kvm_vcpu_is_preempted.patch @@ -0,0 +1,36 @@ +From edbaf6e5e93acda96aae23ba134ef3c1466da3b5 Mon Sep 17 00:00:00 2001 +From: Greg Kroah-Hartman +Date: Thu, 30 Jun 2022 12:19:47 +0200 +Subject: x86, kvm: use proper ASM macros for kvm_vcpu_is_preempted + +From: Greg Kroah-Hartman + +commit edbaf6e5e93acda96aae23ba134ef3c1466da3b5 upstream. + +The build rightfully complains about: + arch/x86/kernel/kvm.o: warning: objtool: __raw_callee_save___kvm_vcpu_is_preempted()+0x12: missing int3 after ret + +because the ASM_RET call is not being used correctly in kvm_vcpu_is_preempted(). + +This was hand-fixed-up in the kvm merge commit a4cfff3f0f8c ("Merge branch +'kvm-older-features' into HEAD") which of course can not be backported to +stable kernels, so just fix this up directly instead. + +Cc: Paolo Bonzini +Cc: Ben Hutchings +Signed-off-by: Greg Kroah-Hartman +--- + arch/x86/kernel/kvm.c | 2 +- + 1 file changed, 1 insertion(+), 1 deletion(-) + +--- a/arch/x86/kernel/kvm.c ++++ b/arch/x86/kernel/kvm.c +@@ -953,7 +953,7 @@ asm( + "movq __per_cpu_offset(,%rdi,8), %rax;" + "cmpb $0, " __stringify(KVM_STEAL_TIME_preempted) "+steal_time(%rax);" + "setne %al;" +-"ret;" ++ASM_RET + ".size __raw_callee_save___kvm_vcpu_is_preempted, .-__raw_callee_save___kvm_vcpu_is_preempted;" + ".popsection"); +