]> git.ipfire.org Git - thirdparty/kernel/stable.git/commitdiff
KVM: arm64: Disable SPE Profiling Buffer when running in guest context
authorWill Deacon <will@kernel.org>
Fri, 27 Mar 2026 13:00:45 +0000 (13:00 +0000)
committerMarc Zyngier <maz@kernel.org>
Sat, 28 Mar 2026 17:07:49 +0000 (17:07 +0000)
The nVHE world-switch code relies on zeroing PMSCR_EL1 to disable
profiling data generation in guest context when SPE is in use by the
host.

Unfortunately, this may leave PMBLIMITR_EL1.E set and consequently we
can end up running in guest/hypervisor context with the Profiling Buffer
enabled. The current "known issues" document for Rev M.a of the Arm ARM
states that this can lead to speculative, out-of-context translations:

  | 2.18 D23136:
  |
  | When the Profiling Buffer is enabled, profiling is not stopped, and
  | Discard mode is not enabled, the Statistical Profiling Unit might
  | cause speculative translations for the owning translation regime,
  | including when the owning translation regime is out-of-context.

In a similar fashion to TRBE, ensure that the Profiling Buffer is
disabled during the nVHE world switch before we start messing with the
stage-2 MMU and trap configuration.

Cc: Marc Zyngier <maz@kernel.org>
Cc: Oliver Upton <oupton@kernel.org>
Cc: James Clark <james.clark@linaro.org>
Cc: Leo Yan <leo.yan@arm.com>
Cc: Suzuki K Poulose <suzuki.poulose@arm.com>
Cc: Fuad Tabba <tabba@google.com>
Cc: Alexandru Elisei <alexandru.elisei@arm.com>
Reviewed-by: Alexandru Elisei <alexandru.elisei@arm.com>
Reviewed-by: Fuad Tabba <tabba@google.com>
Tested-by: Alexandru Elisei <alexandru.elisei@arm.com>
Tested-by: Fuad Tabba <tabba@google.com>
Fixes: f85279b4bd48 ("arm64: KVM: Save/restore the host SPE state when entering/leaving a VM")
Signed-off-by: Will Deacon <will@kernel.org>
Link: https://patch.msgid.link/20260327130047.21065-3-will@kernel.org
Signed-off-by: Marc Zyngier <maz@kernel.org>
arch/arm64/include/asm/kvm_host.h
arch/arm64/kvm/hyp/nvhe/debug-sr.c
arch/arm64/kvm/hyp/nvhe/switch.c

index b1335c55dbef50194ad3bef672a5208a58464c75..fe588760fe62b1ebcf8efb7d7ae5954353c9c587 100644 (file)
@@ -768,6 +768,7 @@ struct kvm_host_data {
                struct kvm_guest_debug_arch regs;
                /* Statistical profiling extension */
                u64 pmscr_el1;
+               u64 pmblimitr_el1;
                /* Self-hosted trace */
                u64 trfcr_el1;
                u64 trblimitr_el1;
index 0955af771ad1d8ec96e9f20f2d5ad89f36a7a39f..84bc80f4e36b24a5875fb3ffe94fab1d36cb8cb7 100644 (file)
 #include <asm/kvm_hyp.h>
 #include <asm/kvm_mmu.h>
 
-static void __debug_save_spe(u64 *pmscr_el1)
+static void __debug_save_spe(void)
 {
-       u64 reg;
+       u64 *pmscr_el1, *pmblimitr_el1;
 
-       /* Clear pmscr in case of early return */
-       *pmscr_el1 = 0;
+       pmscr_el1 = host_data_ptr(host_debug_state.pmscr_el1);
+       pmblimitr_el1 = host_data_ptr(host_debug_state.pmblimitr_el1);
 
        /*
         * At this point, we know that this CPU implements
         * SPE and is available to the host.
         * Check if the host is actually using it ?
         */
-       reg = read_sysreg_s(SYS_PMBLIMITR_EL1);
-       if (!(reg & BIT(PMBLIMITR_EL1_E_SHIFT)))
+       *pmblimitr_el1 = read_sysreg_s(SYS_PMBLIMITR_EL1);
+       if (!(*pmblimitr_el1 & BIT(PMBLIMITR_EL1_E_SHIFT)))
                return;
 
        /* Yes; save the control register and disable data generation */
@@ -37,18 +37,29 @@ static void __debug_save_spe(u64 *pmscr_el1)
 
        /* Now drain all buffered data to memory */
        psb_csync();
+       dsb(nsh);
+
+       /* And disable the profiling buffer */
+       write_sysreg_s(0, SYS_PMBLIMITR_EL1);
+       isb();
 }
 
-static void __debug_restore_spe(u64 pmscr_el1)
+static void __debug_restore_spe(void)
 {
-       if (!pmscr_el1)
+       u64 pmblimitr_el1 = *host_data_ptr(host_debug_state.pmblimitr_el1);
+
+       if (!(pmblimitr_el1 & BIT(PMBLIMITR_EL1_E_SHIFT)))
                return;
 
        /* The host page table is installed, but not yet synchronised */
        isb();
 
+       /* Re-enable the profiling buffer. */
+       write_sysreg_s(pmblimitr_el1, SYS_PMBLIMITR_EL1);
+       isb();
+
        /* Re-enable data generation */
-       write_sysreg_el1(pmscr_el1, SYS_PMSCR);
+       write_sysreg_el1(*host_data_ptr(host_debug_state.pmscr_el1), SYS_PMSCR);
 }
 
 static void __trace_do_switch(u64 *saved_trfcr, u64 new_trfcr)
@@ -175,7 +186,7 @@ void __debug_save_host_buffers_nvhe(struct kvm_vcpu *vcpu)
 {
        /* Disable and flush SPE data generation */
        if (host_data_test_flag(HAS_SPE))
-               __debug_save_spe(host_data_ptr(host_debug_state.pmscr_el1));
+               __debug_save_spe();
 
        /* Disable BRBE branch records */
        if (host_data_test_flag(HAS_BRBE))
@@ -193,7 +204,7 @@ void __debug_switch_to_guest(struct kvm_vcpu *vcpu)
 void __debug_restore_host_buffers_nvhe(struct kvm_vcpu *vcpu)
 {
        if (host_data_test_flag(HAS_SPE))
-               __debug_restore_spe(*host_data_ptr(host_debug_state.pmscr_el1));
+               __debug_restore_spe();
        if (host_data_test_flag(HAS_BRBE))
                __debug_restore_brbe(*host_data_ptr(host_debug_state.brbcr_el1));
        if (__trace_needs_switch())
index f00688e69d88cd65597ea11dfe4c644d78ef6213..9b6e87dac3b95a9b2add34daad246d89f75a7ba0 100644 (file)
@@ -278,7 +278,7 @@ int __kvm_vcpu_run(struct kvm_vcpu *vcpu)
         * We're about to restore some new MMU state. Make sure
         * ongoing page-table walks that have started before we
         * trapped to EL2 have completed. This also synchronises the
-        * above disabling of BRBE and SPE.
+        * above disabling of BRBE.
         *
         * See DDI0487I.a D8.1.5 "Out-of-context translation regimes",
         * rule R_LFHQG and subsequent information statements.