]> git.ipfire.org Git - thirdparty/linux.git/commitdiff
KVM: arm64: nv: Describe trap behaviour of MDCR_EL2.HPMN
authorOliver Upton <oliver.upton@linux.dev>
Fri, 25 Oct 2024 18:23:45 +0000 (18:23 +0000)
committerOliver Upton <oliver.upton@linux.dev>
Thu, 31 Oct 2024 19:00:40 +0000 (19:00 +0000)
MDCR_EL2.HPMN splits the PMU event counters into two ranges: the first
range is accessible from all ELs, and the second range is accessible
only to EL2/3. Supposing the guest hypervisor allows direct access to
the PMU counters from the L2, KVM needs to locally handle those
accesses.

Add a new complex trap configuration for HPMN that checks if the counter
index is accessible to the current context. As written, the architecture
suggests HPMN only causes PMEVCNTR<n>_EL0 to trap, though intuition (and
the pseudocode) suggest that the trap applies to PMEVTYPER<n>_EL0 as
well.

Reviewed-by: Marc Zyngier <maz@kernel.org>
Link: https://lore.kernel.org/r/20241025182354.3364124-11-oliver.upton@linux.dev
Signed-off-by: Oliver Upton <oliver.upton@linux.dev>
arch/arm64/kvm/emulate-nested.c
arch/arm64/kvm/pmu-emul.c
include/kvm/arm_pmu.h

index 4f14fa47df8e9a9dadd37edbaf25028ac4d47068..897ea81ed2b2990f8051f36483e622d1b16b441b 100644 (file)
@@ -110,6 +110,7 @@ enum cgt_group_id {
        CGT_HCR_TPU_TOCU,
        CGT_HCR_NV1_nNV2_ENSCXT,
        CGT_MDCR_TPM_TPMCR,
+       CGT_MDCR_TPM_HPMN,
        CGT_MDCR_TDE_TDA,
        CGT_MDCR_TDE_TDOSA,
        CGT_MDCR_TDE_TDRA,
@@ -126,6 +127,7 @@ enum cgt_group_id {
        CGT_CNTHCTL_EL1PTEN,
 
        CGT_CPTR_TTA,
+       CGT_MDCR_HPMN,
 
        /* Must be last */
        __NR_CGT_GROUP_IDS__
@@ -441,6 +443,7 @@ static const enum cgt_group_id *coarse_control_combo[] = {
        MCB(CGT_HCR_TPU_TOCU,           CGT_HCR_TPU, CGT_HCR_TOCU),
        MCB(CGT_HCR_NV1_nNV2_ENSCXT,    CGT_HCR_NV1_nNV2, CGT_HCR_ENSCXT),
        MCB(CGT_MDCR_TPM_TPMCR,         CGT_MDCR_TPM, CGT_MDCR_TPMCR),
+       MCB(CGT_MDCR_TPM_HPMN,          CGT_MDCR_TPM, CGT_MDCR_HPMN),
        MCB(CGT_MDCR_TDE_TDA,           CGT_MDCR_TDE, CGT_MDCR_TDA),
        MCB(CGT_MDCR_TDE_TDOSA,         CGT_MDCR_TDE, CGT_MDCR_TDOSA),
        MCB(CGT_MDCR_TDE_TDRA,          CGT_MDCR_TDE, CGT_MDCR_TDRA),
@@ -504,6 +507,34 @@ static enum trap_behaviour check_cptr_tta(struct kvm_vcpu *vcpu)
        return BEHAVE_HANDLE_LOCALLY;
 }
 
+static enum trap_behaviour check_mdcr_hpmn(struct kvm_vcpu *vcpu)
+{
+       u32 sysreg = esr_sys64_to_sysreg(kvm_vcpu_get_esr(vcpu));
+       unsigned int idx;
+
+
+       switch (sysreg) {
+       case SYS_PMEVTYPERn_EL0(0) ... SYS_PMEVTYPERn_EL0(30):
+       case SYS_PMEVCNTRn_EL0(0) ... SYS_PMEVCNTRn_EL0(30):
+               idx = (sys_reg_CRm(sysreg) & 0x3) << 3 | sys_reg_Op2(sysreg);
+               break;
+       case SYS_PMXEVTYPER_EL0:
+       case SYS_PMXEVCNTR_EL0:
+               idx = SYS_FIELD_GET(PMSELR_EL0, SEL,
+                                   __vcpu_sys_reg(vcpu, PMSELR_EL0));
+               break;
+       default:
+               /* Someone used this trap helper for something else... */
+               KVM_BUG_ON(1, vcpu->kvm);
+               return BEHAVE_HANDLE_LOCALLY;
+       }
+
+       if (kvm_pmu_counter_is_hyp(vcpu, idx))
+               return BEHAVE_FORWARD_RW | BEHAVE_FORWARD_IN_HOST_EL0;
+
+       return BEHAVE_HANDLE_LOCALLY;
+}
+
 #define CCC(id, fn)                            \
        [id - __COMPLEX_CONDITIONS__] = fn
 
@@ -511,6 +542,7 @@ static const complex_condition_check ccc[] = {
        CCC(CGT_CNTHCTL_EL1PCTEN, check_cnthctl_el1pcten),
        CCC(CGT_CNTHCTL_EL1PTEN, check_cnthctl_el1pten),
        CCC(CGT_CPTR_TTA, check_cptr_tta),
+       CCC(CGT_MDCR_HPMN, check_mdcr_hpmn),
 };
 
 /*
@@ -925,77 +957,77 @@ static const struct encoding_to_trap_config encoding_to_cgt[] __initconst = {
        SR_TRAP(SYS_PMOVSCLR_EL0,       CGT_MDCR_TPM),
        SR_TRAP(SYS_PMCEID0_EL0,        CGT_MDCR_TPM),
        SR_TRAP(SYS_PMCEID1_EL0,        CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMXEVTYPER_EL0,     CGT_MDCR_TPM),
+       SR_TRAP(SYS_PMXEVTYPER_EL0,     CGT_MDCR_TPM_HPMN),
        SR_TRAP(SYS_PMSWINC_EL0,        CGT_MDCR_TPM),
        SR_TRAP(SYS_PMSELR_EL0,         CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMXEVCNTR_EL0,      CGT_MDCR_TPM),
+       SR_TRAP(SYS_PMXEVCNTR_EL0,      CGT_MDCR_TPM_HPMN),
        SR_TRAP(SYS_PMCCNTR_EL0,        CGT_MDCR_TPM),
        SR_TRAP(SYS_PMUSERENR_EL0,      CGT_MDCR_TPM),
        SR_TRAP(SYS_PMINTENSET_EL1,     CGT_MDCR_TPM),
        SR_TRAP(SYS_PMINTENCLR_EL1,     CGT_MDCR_TPM),
        SR_TRAP(SYS_PMMIR_EL1,          CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(0),   CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(1),   CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(2),   CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(3),   CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(4),   CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(5),   CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(6),   CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(7),   CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(8),   CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(9),   CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(10),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(11),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(12),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(13),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(14),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(15),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(16),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(17),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(18),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(19),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(20),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(21),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(22),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(23),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(24),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(25),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(26),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(27),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(28),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(29),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVCNTRn_EL0(30),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(0),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(1),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(2),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(3),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(4),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(5),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(6),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(7),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(8),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(9),  CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(10), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(11), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(12), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(13), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(14), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(15), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(16), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(17), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(18), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(19), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(20), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(21), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(22), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(23), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(24), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(25), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(26), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(27), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(28), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(29), CGT_MDCR_TPM),
-       SR_TRAP(SYS_PMEVTYPERn_EL0(30), CGT_MDCR_TPM),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(0),   CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(1),   CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(2),   CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(3),   CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(4),   CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(5),   CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(6),   CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(7),   CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(8),   CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(9),   CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(10),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(11),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(12),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(13),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(14),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(15),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(16),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(17),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(18),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(19),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(20),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(21),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(22),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(23),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(24),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(25),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(26),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(27),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(28),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(29),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVCNTRn_EL0(30),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(0),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(1),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(2),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(3),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(4),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(5),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(6),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(7),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(8),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(9),  CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(10), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(11), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(12), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(13), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(14), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(15), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(16), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(17), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(18), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(19), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(20), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(21), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(22), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(23), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(24), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(25), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(26), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(27), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(28), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(29), CGT_MDCR_TPM_HPMN),
+       SR_TRAP(SYS_PMEVTYPERn_EL0(30), CGT_MDCR_TPM_HPMN),
        SR_TRAP(SYS_PMCCFILTR_EL0,      CGT_MDCR_TPM),
        SR_TRAP(SYS_MDCCSR_EL0,         CGT_MDCR_TDCC_TDE_TDA),
        SR_TRAP(SYS_MDCCINT_EL1,        CGT_MDCR_TDCC_TDE_TDA),
index ac36c438b8c18c9d4a06781707dc73ea202c948e..28f938f145ac722b96a15d5e6db9433bbd36bc28 100644 (file)
@@ -265,6 +265,24 @@ void kvm_pmu_vcpu_destroy(struct kvm_vcpu *vcpu)
        irq_work_sync(&vcpu->arch.pmu.overflow_work);
 }
 
+bool kvm_pmu_counter_is_hyp(struct kvm_vcpu *vcpu, unsigned int idx)
+{
+       unsigned int hpmn;
+
+       if (!vcpu_has_nv(vcpu) || idx == ARMV8_PMU_CYCLE_IDX)
+               return false;
+
+       /*
+        * Programming HPMN=0 is CONSTRAINED UNPREDICTABLE if FEAT_HPMN0 isn't
+        * implemented. Since KVM's ability to emulate HPMN=0 does not directly
+        * depend on hardware (all PMU registers are trapped), make the
+        * implementation choice that all counters are included in the second
+        * range reserved for EL2/EL3.
+        */
+       hpmn = SYS_FIELD_GET(MDCR_EL2, HPMN, __vcpu_sys_reg(vcpu, MDCR_EL2));
+       return idx >= hpmn;
+}
+
 u64 kvm_pmu_valid_counter_mask(struct kvm_vcpu *vcpu)
 {
        u64 val = FIELD_GET(ARMV8_PMU_PMCR_N, kvm_vcpu_read_pmcr(vcpu));
index e08aeec5d936fb1cee2900e6996196854afb2094..feb5d1d35f0fde43f49901c6b010a4f6fd2d7d04 100644 (file)
@@ -96,6 +96,7 @@ int kvm_arm_set_default_pmu(struct kvm *kvm);
 u8 kvm_arm_pmu_get_max_counters(struct kvm *kvm);
 
 u64 kvm_vcpu_read_pmcr(struct kvm_vcpu *vcpu);
+bool kvm_pmu_counter_is_hyp(struct kvm_vcpu *vcpu, unsigned int idx);
 #else
 struct kvm_pmu {
 };
@@ -187,6 +188,11 @@ static inline u64 kvm_vcpu_read_pmcr(struct kvm_vcpu *vcpu)
        return 0;
 }
 
+static inline bool kvm_pmu_counter_is_hyp(struct kvm_vcpu *vcpu, unsigned int idx)
+{
+       return false;
+}
+
 #endif
 
 #endif