]> git.ipfire.org Git - thirdparty/kernel/stable.git/commitdiff
KVM: arm64: Eagerly switch ZCR_EL{1,2}
authorMark Rutland <mark.rutland@arm.com>
Fri, 21 Mar 2025 00:16:08 +0000 (00:16 +0000)
committerGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Fri, 28 Mar 2025 20:59:56 +0000 (21:59 +0100)
[ Upstream commit 59419f10045bc955d2229819c7cf7a8b0b9c5b59 ]

In non-protected KVM modes, while the guest FPSIMD/SVE/SME state is live on the
CPU, the host's active SVE VL may differ from the guest's maximum SVE VL:

* For VHE hosts, when a VM uses NV, ZCR_EL2 contains a value constrained
  by the guest hypervisor, which may be less than or equal to that
  guest's maximum VL.

  Note: in this case the value of ZCR_EL1 is immaterial due to E2H.

* For nVHE/hVHE hosts, ZCR_EL1 contains a value written by the guest,
  which may be less than or greater than the guest's maximum VL.

  Note: in this case hyp code traps host SVE usage and lazily restores
  ZCR_EL2 to the host's maximum VL, which may be greater than the
  guest's maximum VL.

This can be the case between exiting a guest and kvm_arch_vcpu_put_fp().
If a softirq is taken during this period and the softirq handler tries
to use kernel-mode NEON, then the kernel will fail to save the guest's
FPSIMD/SVE state, and will pend a SIGKILL for the current thread.

This happens because kvm_arch_vcpu_ctxsync_fp() binds the guest's live
FPSIMD/SVE state with the guest's maximum SVE VL, and
fpsimd_save_user_state() verifies that the live SVE VL is as expected
before attempting to save the register state:

| if (WARN_ON(sve_get_vl() != vl)) {
|         force_signal_inject(SIGKILL, SI_KERNEL, 0, 0);
|         return;
| }

Fix this and make this a bit easier to reason about by always eagerly
switching ZCR_EL{1,2} at hyp during guest<->host transitions. With this
happening, there's no need to trap host SVE usage, and the nVHE/nVHE
__deactivate_cptr_traps() logic can be simplified to enable host access
to all present FPSIMD/SVE/SME features.

In protected nVHE/hVHE modes, the host's state is always saved/restored
by hyp, and the guest's state is saved prior to exit to the host, so
from the host's PoV the guest never has live FPSIMD/SVE/SME state, and
the host's ZCR_EL1 is never clobbered by hyp.

Fixes: 8c8010d69c132273 ("KVM: arm64: Save/restore SVE state for nVHE")
Fixes: 2e3cf82063a00ea0 ("KVM: arm64: nv: Ensure correct VL is loaded before saving SVE state")
Signed-off-by: Mark Rutland <mark.rutland@arm.com>
Reviewed-by: Mark Brown <broonie@kernel.org>
Tested-by: Mark Brown <broonie@kernel.org>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: Fuad Tabba <tabba@google.com>
Cc: Marc Zyngier <maz@kernel.org>
Cc: Oliver Upton <oliver.upton@linux.dev>
Cc: Will Deacon <will@kernel.org>
Reviewed-by: Oliver Upton <oliver.upton@linux.dev>
Link: https://lore.kernel.org/r/20250210195226.1215254-9-mark.rutland@arm.com
Signed-off-by: Marc Zyngier <maz@kernel.org>
[ v6.6 lacks pKVM saving of host SVE state, pull in discovery of maximum
  host VL separately -- broonie ]
Signed-off-by: Mark Brown <broonie@kernel.org>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
arch/arm64/include/asm/kvm_host.h
arch/arm64/include/asm/kvm_hyp.h
arch/arm64/kvm/fpsimd.c
arch/arm64/kvm/hyp/entry.S
arch/arm64/kvm/hyp/include/hyp/switch.h
arch/arm64/kvm/hyp/nvhe/hyp-main.c
arch/arm64/kvm/hyp/nvhe/pkvm.c
arch/arm64/kvm/hyp/nvhe/switch.c
arch/arm64/kvm/hyp/vhe/switch.c
arch/arm64/kvm/reset.c

index 27c53b603d693d1cfe8d81f94a59d3afb665be8e..df73e452d2cc0e5fd815d398a932763a5aa7607a 100644 (file)
@@ -73,6 +73,7 @@ static inline enum kvm_mode kvm_get_mode(void) { return KVM_MODE_NONE; };
 #endif
 
 extern unsigned int __ro_after_init kvm_sve_max_vl;
+extern unsigned int __ro_after_init kvm_host_sve_max_vl;
 int __init kvm_arm_init_sve(void);
 
 u32 __attribute_const__ kvm_target_cpu(void);
index 66efd67ea7e8d6dd98dc6088a6004fdb646df2a3..51cd106cd840a006264d9a316f2dc6332e8327ad 100644 (file)
@@ -145,5 +145,6 @@ extern u64 kvm_nvhe_sym(id_aa64smfr0_el1_sys_val);
 
 extern unsigned long kvm_nvhe_sym(__icache_flags);
 extern unsigned int kvm_nvhe_sym(kvm_arm_vmid_bits);
+extern unsigned int kvm_nvhe_sym(kvm_host_sve_max_vl);
 
 #endif /* __ARM64_KVM_HYP_H__ */
index 0492a1fc718130934f8bada836fe42bea6db7abb..e57db49ea468e41e411fd896321c3b1b87a098ec 100644 (file)
@@ -152,15 +152,16 @@ void kvm_arch_vcpu_put_fp(struct kvm_vcpu *vcpu)
        local_irq_save(flags);
 
        if (vcpu->arch.fp_state == FP_STATE_GUEST_OWNED) {
-               if (vcpu_has_sve(vcpu)) {
-                       __vcpu_sys_reg(vcpu, ZCR_EL1) = read_sysreg_el1(SYS_ZCR);
-
-                       /* Restore the VL that was saved when bound to the CPU */
-                       if (!has_vhe())
-                               sve_cond_update_zcr_vq(vcpu_sve_max_vq(vcpu) - 1,
-                                                      SYS_ZCR_EL1);
-               }
-
+               /*
+                * Flush (save and invalidate) the fpsimd/sve state so that if
+                * the host tries to use fpsimd/sve, it's not using stale data
+                * from the guest.
+                *
+                * Flushing the state sets the TIF_FOREIGN_FPSTATE bit for the
+                * context unconditionally, in both nVHE and VHE. This allows
+                * the kernel to restore the fpsimd/sve state, including ZCR_EL1
+                * when needed.
+                */
                fpsimd_save_and_flush_cpu_state();
        }
 
index f3aa7738b477d6a83e41db42cc161fc6f9aa6291..f02d5701fc21c685122aa5ba92309f7f4def3ae3 100644 (file)
@@ -44,6 +44,11 @@ alternative_if ARM64_HAS_RAS_EXTN
 alternative_else_nop_endif
        mrs     x1, isr_el1
        cbz     x1,  1f
+
+       // Ensure that __guest_enter() always provides a context
+       // synchronization event so that callers don't need ISBs for anything
+       // that would usually be synchonized by the ERET.
+       isb
        mov     x0, #ARM_EXCEPTION_IRQ
        ret
 
index 77a9f68ca4904a2b39ca59b7ee7cf54b9fe11f66..526085401f66387137a0d78d1d520af8b2306e43 100644 (file)
@@ -273,6 +273,61 @@ static inline void __hyp_sve_restore_guest(struct kvm_vcpu *vcpu)
        write_sysreg_el1(__vcpu_sys_reg(vcpu, ZCR_EL1), SYS_ZCR);
 }
 
+static inline void fpsimd_lazy_switch_to_guest(struct kvm_vcpu *vcpu)
+{
+       u64 zcr_el1, zcr_el2;
+
+       if (!guest_owns_fp_regs(vcpu))
+               return;
+
+       if (vcpu_has_sve(vcpu)) {
+               zcr_el2 = vcpu_sve_max_vq(vcpu) - 1;
+
+               write_sysreg_el2(zcr_el2, SYS_ZCR);
+
+               zcr_el1 = __vcpu_sys_reg(vcpu, ZCR_EL1);
+               write_sysreg_el1(zcr_el1, SYS_ZCR);
+       }
+}
+
+static inline void fpsimd_lazy_switch_to_host(struct kvm_vcpu *vcpu)
+{
+       u64 zcr_el1, zcr_el2;
+
+       if (!guest_owns_fp_regs(vcpu))
+               return;
+
+       /*
+        * When the guest owns the FP regs, we know that guest+hyp traps for
+        * any FPSIMD/SVE/SME features exposed to the guest have been disabled
+        * by either fpsimd_lazy_switch_to_guest() or kvm_hyp_handle_fpsimd()
+        * prior to __guest_entry(). As __guest_entry() guarantees a context
+        * synchronization event, we don't need an ISB here to avoid taking
+        * traps for anything that was exposed to the guest.
+        */
+       if (vcpu_has_sve(vcpu)) {
+               zcr_el1 = read_sysreg_el1(SYS_ZCR);
+               __vcpu_sys_reg(vcpu, ZCR_EL1) = zcr_el1;
+
+               /*
+                * The guest's state is always saved using the guest's max VL.
+                * Ensure that the host has the guest's max VL active such that
+                * the host can save the guest's state lazily, but don't
+                * artificially restrict the host to the guest's max VL.
+                */
+               if (has_vhe()) {
+                       zcr_el2 = vcpu_sve_max_vq(vcpu) - 1;
+                       write_sysreg_el2(zcr_el2, SYS_ZCR);
+               } else {
+                       zcr_el2 = sve_vq_from_vl(kvm_host_sve_max_vl) - 1;
+                       write_sysreg_el2(zcr_el2, SYS_ZCR);
+
+                       zcr_el1 = vcpu_sve_max_vq(vcpu) - 1;
+                       write_sysreg_el1(zcr_el1, SYS_ZCR);
+               }
+       }
+}
+
 /*
  * We trap the first access to the FP/SIMD to save the host context and
  * restore the guest context lazily.
index 8390173f7f5b38e01c04ae6563853df11a1c3df2..350d1775a5ce880ac23bb7adddbed644e45d0e40 100644 (file)
@@ -5,6 +5,7 @@
  */
 
 #include <hyp/adjust_pc.h>
+#include <hyp/switch.h>
 
 #include <asm/pgtable-types.h>
 #include <asm/kvm_asm.h>
@@ -95,7 +96,9 @@ static void handle___kvm_vcpu_run(struct kvm_cpu_context *host_ctxt)
                pkvm_put_hyp_vcpu(hyp_vcpu);
        } else {
                /* The host is fully trusted, run its vCPU directly. */
+               fpsimd_lazy_switch_to_guest(host_vcpu);
                ret = __kvm_vcpu_run(host_vcpu);
+               fpsimd_lazy_switch_to_host(host_vcpu);
        }
 
 out:
@@ -416,15 +419,6 @@ void handle_trap(struct kvm_cpu_context *host_ctxt)
        case ESR_ELx_EC_SMC64:
                handle_host_smc(host_ctxt);
                break;
-       case ESR_ELx_EC_SVE:
-               if (has_hvhe())
-                       sysreg_clear_set(cpacr_el1, 0, (CPACR_EL1_ZEN_EL1EN |
-                                                       CPACR_EL1_ZEN_EL0EN));
-               else
-                       sysreg_clear_set(cptr_el2, CPTR_EL2_TZ, 0);
-               isb();
-               sve_cond_update_zcr_vq(ZCR_ELx_LEN_MASK, SYS_ZCR_EL2);
-               break;
        case ESR_ELx_EC_IABT_LOW:
        case ESR_ELx_EC_DABT_LOW:
                handle_host_mem_abort(host_ctxt);
index 9e7612343ad66d3dc13b8ee858a4e61e917901c9..03acc8343c5d1b23ce0ea1fa2fcabaf2a93902c5 100644 (file)
@@ -18,6 +18,8 @@ unsigned long __icache_flags;
 /* Used by kvm_get_vttbr(). */
 unsigned int kvm_arm_vmid_bits;
 
+unsigned int kvm_host_sve_max_vl;
+
 /*
  * Set trap register values based on features in ID_AA64PFR0.
  */
index aa827e2a2b1e5b2c67304221706f406b42ceafb2..27ad6933bec003d722d3d66519b6225cd207cad7 100644 (file)
@@ -40,6 +40,9 @@ static void __activate_cptr_traps(struct kvm_vcpu *vcpu)
 {
        u64 val = CPTR_EL2_TAM; /* Same bit irrespective of E2H */
 
+       if (!guest_owns_fp_regs(vcpu))
+               __activate_traps_fpsimd32(vcpu);
+
        if (has_hvhe()) {
                val |= CPACR_ELx_TTA;
 
@@ -48,6 +51,8 @@ static void __activate_cptr_traps(struct kvm_vcpu *vcpu)
                        if (vcpu_has_sve(vcpu))
                                val |= CPACR_ELx_ZEN;
                }
+
+               write_sysreg(val, cpacr_el1);
        } else {
                val |= CPTR_EL2_TTA | CPTR_NVHE_EL2_RES1;
 
@@ -62,12 +67,32 @@ static void __activate_cptr_traps(struct kvm_vcpu *vcpu)
 
                if (!guest_owns_fp_regs(vcpu))
                        val |= CPTR_EL2_TFP;
+
+               write_sysreg(val, cptr_el2);
        }
+}
 
-       if (!guest_owns_fp_regs(vcpu))
-               __activate_traps_fpsimd32(vcpu);
+static void __deactivate_cptr_traps(struct kvm_vcpu *vcpu)
+{
+       if (has_hvhe()) {
+               u64 val = CPACR_ELx_FPEN;
+
+               if (cpus_have_final_cap(ARM64_SVE))
+                       val |= CPACR_ELx_ZEN;
+               if (cpus_have_final_cap(ARM64_SME))
+                       val |= CPACR_ELx_SMEN;
+
+               write_sysreg(val, cpacr_el1);
+       } else {
+               u64 val = CPTR_NVHE_EL2_RES1;
+
+               if (!cpus_have_final_cap(ARM64_SVE))
+                       val |= CPTR_EL2_TZ;
+               if (!cpus_have_final_cap(ARM64_SME))
+                       val |= CPTR_EL2_TSM;
 
-       kvm_write_cptr_el2(val);
+               write_sysreg(val, cptr_el2);
+       }
 }
 
 static void __activate_traps(struct kvm_vcpu *vcpu)
@@ -120,7 +145,7 @@ static void __deactivate_traps(struct kvm_vcpu *vcpu)
 
        write_sysreg(this_cpu_ptr(&kvm_init_params)->hcr_el2, hcr_el2);
 
-       kvm_reset_cptr_el2(vcpu);
+       __deactivate_cptr_traps(vcpu);
        write_sysreg(__kvm_hyp_host_vector, vbar_el2);
 }
 
index d0c7725879a7b38cea53f1dfb111933e0ad89845..31304da0d617313244caff582853965a7a206f72 100644 (file)
@@ -212,6 +212,8 @@ static int __kvm_vcpu_run_vhe(struct kvm_vcpu *vcpu)
 
        sysreg_save_host_state_vhe(host_ctxt);
 
+       fpsimd_lazy_switch_to_guest(vcpu);
+
        /*
         * ARM erratum 1165522 requires us to configure both stage 1 and
         * stage 2 translation for the guest context before we clear
@@ -247,6 +249,8 @@ static int __kvm_vcpu_run_vhe(struct kvm_vcpu *vcpu)
 
        __deactivate_traps(vcpu);
 
+       fpsimd_lazy_switch_to_host(vcpu);
+
        sysreg_restore_host_state_vhe(host_ctxt);
 
        if (vcpu->arch.fp_state == FP_STATE_GUEST_OWNED)
index 7a65a35ee4ac4165b4147c724d28b59f8150a0d2..43a53a403f5102c81bba0d928c0f78331f800269 100644 (file)
@@ -46,11 +46,14 @@ static u32 __ro_after_init kvm_ipa_limit;
                                 PSR_AA32_I_BIT | PSR_AA32_F_BIT)
 
 unsigned int __ro_after_init kvm_sve_max_vl;
+unsigned int __ro_after_init kvm_host_sve_max_vl;
 
 int __init kvm_arm_init_sve(void)
 {
        if (system_supports_sve()) {
                kvm_sve_max_vl = sve_max_virtualisable_vl();
+               kvm_host_sve_max_vl = sve_max_vl();
+               kvm_nvhe_sym(kvm_host_sve_max_vl) = kvm_host_sve_max_vl;
 
                /*
                 * The get_sve_reg()/set_sve_reg() ioctl interface will need