]> git.ipfire.org Git - thirdparty/kernel/linux.git/commitdiff
KVM: x86/pmu: Manipulate FIXED_CTR_CTRL MSR with macros
authorDapeng Mi <dapeng1.mi@linux.intel.com>
Tue, 30 Apr 2024 00:52:39 +0000 (08:52 +0800)
committerSean Christopherson <seanjc@google.com>
Mon, 3 Jun 2024 21:25:22 +0000 (14:25 -0700)
Magic numbers are used to manipulate the bit fields of
FIXED_CTR_CTRL MSR. This makes reading code become difficult, so use
pre-defined macros to replace these magic numbers.

Signed-off-by: Dapeng Mi <dapeng1.mi@linux.intel.com>
Link: https://lore.kernel.org/r/20240430005239.13527-3-dapeng1.mi@linux.intel.com
[sean: drop unnecessary curly braces]
Signed-off-by: Sean Christopherson <seanjc@google.com>
arch/x86/kvm/pmu.c
arch/x86/kvm/pmu.h
arch/x86/kvm/vmx/pmu_intel.c

index cc909a28e92f7dbb66801cebda453eb475885e68..7d414363ae7ba0e7e6fe984aa4127b43d29280f4 100644 (file)
@@ -469,11 +469,11 @@ static int reprogram_counter(struct kvm_pmc *pmc)
        if (pmc_is_fixed(pmc)) {
                fixed_ctr_ctrl = fixed_ctrl_field(pmu->fixed_ctr_ctrl,
                                                  pmc->idx - KVM_FIXED_PMC_BASE_IDX);
-               if (fixed_ctr_ctrl & 0x1)
+               if (fixed_ctr_ctrl & INTEL_FIXED_0_KERNEL)
                        eventsel |= ARCH_PERFMON_EVENTSEL_OS;
-               if (fixed_ctr_ctrl & 0x2)
+               if (fixed_ctr_ctrl & INTEL_FIXED_0_USER)
                        eventsel |= ARCH_PERFMON_EVENTSEL_USR;
-               if (fixed_ctr_ctrl & 0x8)
+               if (fixed_ctr_ctrl & INTEL_FIXED_0_ENABLE_PMI)
                        eventsel |= ARCH_PERFMON_EVENTSEL_INT;
                new_config = (u64)fixed_ctr_ctrl;
        }
@@ -846,8 +846,8 @@ static inline bool cpl_is_matched(struct kvm_pmc *pmc)
        } else {
                config = fixed_ctrl_field(pmc_to_pmu(pmc)->fixed_ctr_ctrl,
                                          pmc->idx - KVM_FIXED_PMC_BASE_IDX);
-               select_os = config & 0x1;
-               select_user = config & 0x2;
+               select_os = config & INTEL_FIXED_0_KERNEL;
+               select_user = config & INTEL_FIXED_0_USER;
        }
 
        /*
index 2eab8ea610db28a2065d58b78eadd1df8ad0944a..d54741fe4bdda991dcb2e68a13158f95d095e841 100644 (file)
@@ -14,7 +14,8 @@
                                          MSR_IA32_MISC_ENABLE_BTS_UNAVAIL)
 
 /* retrieve the 4 bits for EN and PMI out of IA32_FIXED_CTR_CTRL */
-#define fixed_ctrl_field(ctrl_reg, idx) (((ctrl_reg) >> ((idx)*4)) & 0xf)
+#define fixed_ctrl_field(ctrl_reg, idx) \
+       (((ctrl_reg) >> ((idx) * INTEL_FIXED_BITS_STRIDE)) & INTEL_FIXED_BITS_MASK)
 
 #define VMWARE_BACKDOOR_PMC_HOST_TSC           0x10000
 #define VMWARE_BACKDOOR_PMC_REAL_TIME          0x10001
@@ -170,7 +171,8 @@ static inline bool pmc_speculative_in_use(struct kvm_pmc *pmc)
 
        if (pmc_is_fixed(pmc))
                return fixed_ctrl_field(pmu->fixed_ctr_ctrl,
-                                       pmc->idx - KVM_FIXED_PMC_BASE_IDX) & 0x3;
+                                       pmc->idx - KVM_FIXED_PMC_BASE_IDX) &
+                                       (INTEL_FIXED_0_KERNEL | INTEL_FIXED_0_USER);
 
        return pmc->eventsel & ARCH_PERFMON_EVENTSEL_ENABLE;
 }
index eaee9a08952e07cf10af0cadba3ccfeec8e0e8a1..e01c8798192736a373a9ed8debb9c54e3fcb3332 100644 (file)
@@ -502,7 +502,12 @@ static void intel_pmu_refresh(struct kvm_vcpu *vcpu)
        }
 
        for (i = 0; i < pmu->nr_arch_fixed_counters; i++)
-               pmu->fixed_ctr_ctrl_rsvd &= ~(0xbull << (i * 4));
+               pmu->fixed_ctr_ctrl_rsvd &=
+                        ~intel_fixed_bits_by_idx(i,
+                                                 INTEL_FIXED_0_KERNEL |
+                                                 INTEL_FIXED_0_USER |
+                                                 INTEL_FIXED_0_ENABLE_PMI);
+
        counter_rsvd = ~(((1ull << pmu->nr_arch_gp_counters) - 1) |
                (((1ull << pmu->nr_arch_fixed_counters) - 1) << KVM_FIXED_PMC_BASE_IDX));
        pmu->global_ctrl_rsvd = counter_rsvd;
@@ -546,10 +551,9 @@ static void intel_pmu_refresh(struct kvm_vcpu *vcpu)
                if (perf_capabilities & PERF_CAP_PEBS_BASELINE) {
                        pmu->pebs_enable_rsvd = counter_rsvd;
                        pmu->reserved_bits &= ~ICL_EVENTSEL_ADAPTIVE;
-                       for (i = 0; i < pmu->nr_arch_fixed_counters; i++) {
+                       for (i = 0; i < pmu->nr_arch_fixed_counters; i++)
                                pmu->fixed_ctr_ctrl_rsvd &=
-                                       ~(1ULL << (KVM_FIXED_PMC_BASE_IDX + i * 4));
-                       }
+                                       ~intel_fixed_bits_by_idx(i, ICL_FIXED_0_ADAPTIVE);
                        pmu->pebs_data_cfg_rsvd = ~0xff00000full;
                } else {
                        pmu->pebs_enable_rsvd =