arm64: KVM: Add helper to handle PMCR register bits
authorShannon Zhao <shannon.zhao@linaro.org>
Wed, 28 Oct 2015 04:10:30 +0000 (12:10 +0800)
committerMarc Zyngier <marc.zyngier@arm.com>
Mon, 29 Feb 2016 18:34:21 +0000 (18:34 +0000)
According to ARMv8 spec, when writing 1 to PMCR.E, all counters are
enabled by PMCNTENSET, while writing 0 to PMCR.E, all counters are
disabled. When writing 1 to PMCR.P, reset all event counters, not
including PMCCNTR, to zero. When writing 1 to PMCR.C, reset PMCCNTR to
zero.

Signed-off-by: Shannon Zhao <shannon.zhao@linaro.org>
Reviewed-by: Marc Zyngier <marc.zyngier@arm.com>
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
arch/arm64/include/asm/kvm_perf_event.h
arch/arm64/kvm/sys_regs.c
include/kvm/arm_pmu.h
virt/kvm/arm/pmu.c

index 62fa60fbc0b3294058fd7965965f8ea152b763b3..6d080c07873b2a12c86f5d7083663fec2ba9cec2 100644 (file)
 #define ARMV8_PMU_PMCR_D       (1 << 3) /* CCNT counts every 64th cpu cycle */
 #define ARMV8_PMU_PMCR_X       (1 << 4) /* Export to ETM */
 #define ARMV8_PMU_PMCR_DP      (1 << 5) /* Disable CCNT if non-invasive debug*/
+/* Determines which bit of PMCCNTR_EL0 generates an overflow */
+#define ARMV8_PMU_PMCR_LC      (1 << 6)
 #define        ARMV8_PMU_PMCR_N_SHIFT  11       /* Number of counters supported */
 #define        ARMV8_PMU_PMCR_N_MASK   0x1f
-#define        ARMV8_PMU_PMCR_MASK     0x3f     /* Mask for writable bits */
+#define        ARMV8_PMU_PMCR_MASK     0x7f     /* Mask for writable bits */
 
 /*
  * PMOVSR: counters overflow flag status reg
index 10e53796926c19b23ba107e507f53e6bfbd6fac9..12f36ef8caa067906365db3b993b290676fd0d12 100644 (file)
@@ -467,6 +467,7 @@ static bool access_pmcr(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
                val &= ~ARMV8_PMU_PMCR_MASK;
                val |= p->regval & ARMV8_PMU_PMCR_MASK;
                vcpu_sys_reg(vcpu, PMCR_EL0) = val;
+               kvm_pmu_handle_pmcr(vcpu, val);
        } else {
                /* PMCR.P & PMCR.C are RAZ */
                val = vcpu_sys_reg(vcpu, PMCR_EL0)
index 348c4c9d763a651d425366cd09fd1716a683ac25..8bc92d119713d61c581231d15303a06e92b57cff 100644 (file)
@@ -45,6 +45,7 @@ void kvm_pmu_disable_counter(struct kvm_vcpu *vcpu, u64 val);
 void kvm_pmu_enable_counter(struct kvm_vcpu *vcpu, u64 val);
 void kvm_pmu_overflow_set(struct kvm_vcpu *vcpu, u64 val);
 void kvm_pmu_software_increment(struct kvm_vcpu *vcpu, u64 val);
+void kvm_pmu_handle_pmcr(struct kvm_vcpu *vcpu, u64 val);
 void kvm_pmu_set_counter_event_type(struct kvm_vcpu *vcpu, u64 data,
                                    u64 select_idx);
 #else
@@ -67,6 +68,7 @@ static inline void kvm_pmu_disable_counter(struct kvm_vcpu *vcpu, u64 val) {}
 static inline void kvm_pmu_enable_counter(struct kvm_vcpu *vcpu, u64 val) {}
 static inline void kvm_pmu_overflow_set(struct kvm_vcpu *vcpu, u64 val) {}
 static inline void kvm_pmu_software_increment(struct kvm_vcpu *vcpu, u64 val) {}
+static inline void kvm_pmu_handle_pmcr(struct kvm_vcpu *vcpu, u64 val) {}
 static inline void kvm_pmu_set_counter_event_type(struct kvm_vcpu *vcpu,
                                                  u64 data, u64 select_idx) {}
 #endif
index 9fc775ef03ec3c0db2a2752fc05adcec5067f80b..cda869c609ddb75b81f53a80cbcda5ed57a64007 100644 (file)
@@ -210,6 +210,40 @@ void kvm_pmu_software_increment(struct kvm_vcpu *vcpu, u64 val)
        }
 }
 
+/**
+ * kvm_pmu_handle_pmcr - handle PMCR register
+ * @vcpu: The vcpu pointer
+ * @val: the value guest writes to PMCR register
+ */
+void kvm_pmu_handle_pmcr(struct kvm_vcpu *vcpu, u64 val)
+{
+       struct kvm_pmu *pmu = &vcpu->arch.pmu;
+       struct kvm_pmc *pmc;
+       u64 mask;
+       int i;
+
+       mask = kvm_pmu_valid_counter_mask(vcpu);
+       if (val & ARMV8_PMU_PMCR_E) {
+               kvm_pmu_enable_counter(vcpu,
+                               vcpu_sys_reg(vcpu, PMCNTENSET_EL0) & mask);
+       } else {
+               kvm_pmu_disable_counter(vcpu, mask);
+       }
+
+       if (val & ARMV8_PMU_PMCR_C)
+               kvm_pmu_set_counter_value(vcpu, ARMV8_PMU_CYCLE_IDX, 0);
+
+       if (val & ARMV8_PMU_PMCR_P) {
+               for (i = 0; i < ARMV8_PMU_CYCLE_IDX; i++)
+                       kvm_pmu_set_counter_value(vcpu, i, 0);
+       }
+
+       if (val & ARMV8_PMU_PMCR_LC) {
+               pmc = &pmu->pmc[ARMV8_PMU_CYCLE_IDX];
+               pmc->bitmask = 0xffffffffffffffffUL;
+       }
+}
+
 static bool kvm_pmu_counter_is_enabled(struct kvm_vcpu *vcpu, u64 select_idx)
 {
        return (vcpu_sys_reg(vcpu, PMCR_EL0) & ARMV8_PMU_PMCR_E) &&