From: Greg Kroah-Hartman Date: Tue, 29 Sep 2020 10:40:55 +0000 (+0200) Subject: 4.19-stable patches X-Git-Tag: v4.4.238~3 X-Git-Url: http://git.ipfire.org/?a=commitdiff_plain;h=61e5fe8b66b732df300ff0573092eeef6b74a59d;p=thirdparty%2Fkernel%2Fstable-queue.git 4.19-stable patches added patches: kvm-arm64-assume-write-fault-on-s1ptw-permission-fault-on-instruction-fetch.patch --- diff --git a/queue-4.19/kvm-arm64-assume-write-fault-on-s1ptw-permission-fault-on-instruction-fetch.patch b/queue-4.19/kvm-arm64-assume-write-fault-on-s1ptw-permission-fault-on-instruction-fetch.patch new file mode 100644 index 00000000000..17e8d5177e3 --- /dev/null +++ b/queue-4.19/kvm-arm64-assume-write-fault-on-s1ptw-permission-fault-on-instruction-fetch.patch @@ -0,0 +1,153 @@ +From c4ad98e4b72cb5be30ea282fce935248f2300e62 Mon Sep 17 00:00:00 2001 +From: Marc Zyngier +Date: Tue, 15 Sep 2020 11:42:17 +0100 +Subject: KVM: arm64: Assume write fault on S1PTW permission fault on instruction fetch + +From: Marc Zyngier + +commit c4ad98e4b72cb5be30ea282fce935248f2300e62 upstream. + +KVM currently assumes that an instruction abort can never be a write. +This is in general true, except when the abort is triggered by +a S1PTW on instruction fetch that tries to update the S1 page tables +(to set AF, for example). + +This can happen if the page tables have been paged out and brought +back in without seeing a direct write to them (they are thus marked +read only), and the fault handling code will make the PT executable(!) +instead of writable. The guest gets stuck forever. + +In these conditions, the permission fault must be considered as +a write so that the Stage-1 update can take place. This is essentially +the I-side equivalent of the problem fixed by 60e21a0ef54c ("arm64: KVM: +Take S1 walks into account when determining S2 write faults"). + +Update kvm_is_write_fault() to return true on IABT+S1PTW, and introduce +kvm_vcpu_trap_is_exec_fault() that only return true when no faulting +on a S1 fault. Additionally, kvm_vcpu_dabt_iss1tw() is renamed to +kvm_vcpu_abt_iss1tw(), as the above makes it plain that it isn't +specific to data abort. + +Signed-off-by: Marc Zyngier +Reviewed-by: Will Deacon +Cc: stable@vger.kernel.org +Link: https://lore.kernel.org/r/20200915104218.1284701-2-maz@kernel.org +Signed-off-by: Greg Kroah-Hartman +--- + arch/arm/include/asm/kvm_emulate.h | 11 ++++++++--- + arch/arm64/include/asm/kvm_emulate.h | 9 +++++++-- + arch/arm64/kvm/hyp/switch.c | 2 +- + virt/kvm/arm/mmio.c | 2 +- + virt/kvm/arm/mmu.c | 5 ++++- + 5 files changed, 21 insertions(+), 8 deletions(-) + +--- a/arch/arm/include/asm/kvm_emulate.h ++++ b/arch/arm/include/asm/kvm_emulate.h +@@ -216,7 +216,7 @@ static inline int kvm_vcpu_dabt_get_rd(s + return (kvm_vcpu_get_hsr(vcpu) & HSR_SRT_MASK) >> HSR_SRT_SHIFT; + } + +-static inline bool kvm_vcpu_dabt_iss1tw(struct kvm_vcpu *vcpu) ++static inline bool kvm_vcpu_abt_iss1tw(const struct kvm_vcpu *vcpu) + { + return kvm_vcpu_get_hsr(vcpu) & HSR_DABT_S1PTW; + } +@@ -248,16 +248,21 @@ static inline bool kvm_vcpu_trap_il_is32 + return kvm_vcpu_get_hsr(vcpu) & HSR_IL; + } + +-static inline u8 kvm_vcpu_trap_get_class(struct kvm_vcpu *vcpu) ++static inline u8 kvm_vcpu_trap_get_class(const struct kvm_vcpu *vcpu) + { + return kvm_vcpu_get_hsr(vcpu) >> HSR_EC_SHIFT; + } + +-static inline bool kvm_vcpu_trap_is_iabt(struct kvm_vcpu *vcpu) ++static inline bool kvm_vcpu_trap_is_iabt(const struct kvm_vcpu *vcpu) + { + return kvm_vcpu_trap_get_class(vcpu) == HSR_EC_IABT; + } + ++static inline bool kvm_vcpu_trap_is_exec_fault(const struct kvm_vcpu *vcpu) ++{ ++ return kvm_vcpu_trap_is_iabt(vcpu) && !kvm_vcpu_abt_iss1tw(vcpu); ++} ++ + static inline u8 kvm_vcpu_trap_get_fault(struct kvm_vcpu *vcpu) + { + return kvm_vcpu_get_hsr(vcpu) & HSR_FSC; +--- a/arch/arm64/include/asm/kvm_emulate.h ++++ b/arch/arm64/include/asm/kvm_emulate.h +@@ -303,7 +303,7 @@ static inline int kvm_vcpu_dabt_get_rd(c + return (kvm_vcpu_get_hsr(vcpu) & ESR_ELx_SRT_MASK) >> ESR_ELx_SRT_SHIFT; + } + +-static inline bool kvm_vcpu_dabt_iss1tw(const struct kvm_vcpu *vcpu) ++static inline bool kvm_vcpu_abt_iss1tw(const struct kvm_vcpu *vcpu) + { + return !!(kvm_vcpu_get_hsr(vcpu) & ESR_ELx_S1PTW); + } +@@ -311,7 +311,7 @@ static inline bool kvm_vcpu_dabt_iss1tw( + static inline bool kvm_vcpu_dabt_iswrite(const struct kvm_vcpu *vcpu) + { + return !!(kvm_vcpu_get_hsr(vcpu) & ESR_ELx_WNR) || +- kvm_vcpu_dabt_iss1tw(vcpu); /* AF/DBM update */ ++ kvm_vcpu_abt_iss1tw(vcpu); /* AF/DBM update */ + } + + static inline bool kvm_vcpu_dabt_is_cm(const struct kvm_vcpu *vcpu) +@@ -340,6 +340,11 @@ static inline bool kvm_vcpu_trap_is_iabt + return kvm_vcpu_trap_get_class(vcpu) == ESR_ELx_EC_IABT_LOW; + } + ++static inline bool kvm_vcpu_trap_is_exec_fault(const struct kvm_vcpu *vcpu) ++{ ++ return kvm_vcpu_trap_is_iabt(vcpu) && !kvm_vcpu_abt_iss1tw(vcpu); ++} ++ + static inline u8 kvm_vcpu_trap_get_fault(const struct kvm_vcpu *vcpu) + { + return kvm_vcpu_get_hsr(vcpu) & ESR_ELx_FSC; +--- a/arch/arm64/kvm/hyp/switch.c ++++ b/arch/arm64/kvm/hyp/switch.c +@@ -430,7 +430,7 @@ static bool __hyp_text fixup_guest_exit( + kvm_vcpu_trap_get_fault_type(vcpu) == FSC_FAULT && + kvm_vcpu_dabt_isvalid(vcpu) && + !kvm_vcpu_dabt_isextabt(vcpu) && +- !kvm_vcpu_dabt_iss1tw(vcpu); ++ !kvm_vcpu_abt_iss1tw(vcpu); + + if (valid) { + int ret = __vgic_v2_perform_cpuif_access(vcpu); +--- a/virt/kvm/arm/mmio.c ++++ b/virt/kvm/arm/mmio.c +@@ -142,7 +142,7 @@ static int decode_hsr(struct kvm_vcpu *v + bool sign_extend; + bool sixty_four; + +- if (kvm_vcpu_dabt_iss1tw(vcpu)) { ++ if (kvm_vcpu_abt_iss1tw(vcpu)) { + /* page table accesses IO mem: tell guest to fix its TTBR */ + kvm_inject_dabt(vcpu, kvm_vcpu_get_hfar(vcpu)); + return 1; +--- a/virt/kvm/arm/mmu.c ++++ b/virt/kvm/arm/mmu.c +@@ -1282,6 +1282,9 @@ static bool transparent_hugepage_adjust( + + static bool kvm_is_write_fault(struct kvm_vcpu *vcpu) + { ++ if (kvm_vcpu_abt_iss1tw(vcpu)) ++ return true; ++ + if (kvm_vcpu_trap_is_iabt(vcpu)) + return false; + +@@ -1496,7 +1499,7 @@ static int user_mem_abort(struct kvm_vcp + unsigned long flags = 0; + + write_fault = kvm_is_write_fault(vcpu); +- exec_fault = kvm_vcpu_trap_is_iabt(vcpu); ++ exec_fault = kvm_vcpu_trap_is_exec_fault(vcpu); + VM_BUG_ON(write_fault && exec_fault); + + if (fault_status == FSC_PERM && !write_fault && !exec_fault) { diff --git a/queue-4.19/series b/queue-4.19/series index e3be6a38a89..4d0798ad909 100644 --- a/queue-4.19/series +++ b/queue-4.19/series @@ -242,3 +242,4 @@ kprobes-fix-compiler-warning-for-config_kprobes_on_ftrace.patch ata-define-ac_err_ok.patch ata-make-qc_prep-return-ata_completion_errors.patch ata-sata_mv-avoid-trigerrable-bug_on.patch +kvm-arm64-assume-write-fault-on-s1ptw-permission-fault-on-instruction-fetch.patch