[PATCH v1 05/11] KVM: x86/pmu: Add support to reprogram PEBS event for guest counters

From: Luwei Kang
Date: Thu Mar 05 2020 - 04:59:29 EST


From: Like Xu <like.xu@xxxxxxxxxxxxxxx>

When the event precise level is non-zero, the performance counter
will be reprogramed for PEBS event and set PBES PMI bit in global_status
when the PEBS event is overflowed. Since KVM never knows the setting
of precise level in guest because it's a SW parameter, we force all PEBS
events to be precise level 1 for enough accuracy with a dedicated counter.

Originally-by: Andi Kleen <ak@xxxxxxxxxxxxxxx>
Signed-off-by: Like Xu <like.xu@xxxxxxxxxxxxxxx>
Co-developed-by: Kan Liang <kan.liang@xxxxxxxxxxxxxxx>
Signed-off-by: Kan Liang <kan.liang@xxxxxxxxxxxxxxx>
---
arch/x86/include/asm/kvm_host.h | 1 +
arch/x86/kvm/pmu.c | 69 ++++++++++++++++++++++++++++++++++++++++-
arch/x86/kvm/vmx/pmu_intel.c | 1 +
3 files changed, 70 insertions(+), 1 deletion(-)

diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h
index 98959e8..83abb49 100644
--- a/arch/x86/include/asm/kvm_host.h
+++ b/arch/x86/include/asm/kvm_host.h
@@ -478,6 +478,7 @@ struct kvm_pmu {
u64 global_ctrl_mask;
u64 global_ovf_ctrl_mask;
u64 reserved_bits;
+ u64 pebs_enable;
u8 version;
struct kvm_pmc gp_counters[INTEL_PMC_MAX_GENERIC];
struct kvm_pmc fixed_counters[INTEL_PMC_MAX_FIXED];
diff --git a/arch/x86/kvm/pmu.c b/arch/x86/kvm/pmu.c
index b4f9e97..b2bdacb 100644
--- a/arch/x86/kvm/pmu.c
+++ b/arch/x86/kvm/pmu.c
@@ -77,6 +77,11 @@ static void kvm_perf_overflow_intr(struct perf_event *perf_event,

if (!test_and_set_bit(pmc->idx, pmu->reprogram_pmi)) {
__set_bit(pmc->idx, (unsigned long *)&pmu->global_status);
+
+ /* Indicate PEBS overflow to guest. */
+ if (perf_event->attr.precise_ip)
+ __set_bit(62, (unsigned long *)&pmu->global_status);
+
kvm_make_request(KVM_REQ_PMU, pmc->vcpu);

/*
@@ -99,6 +104,7 @@ static void pmc_reprogram_counter(struct kvm_pmc *pmc, u32 type,
bool exclude_kernel, bool intr,
bool in_tx, bool in_tx_cp)
{
+ struct kvm_pmu *pmu = vcpu_to_pmu(pmc->vcpu);
struct perf_event *event;
struct perf_event_attr attr = {
.type = type,
@@ -111,6 +117,7 @@ static void pmc_reprogram_counter(struct kvm_pmc *pmc, u32 type,
.config = config,
.disabled = 1,
};
+ bool pebs = test_bit(pmc->idx, (unsigned long *)&pmu->pebs_enable);

attr.sample_period = (-pmc->counter) & pmc_bitmask(pmc);

@@ -126,8 +133,50 @@ static void pmc_reprogram_counter(struct kvm_pmc *pmc, u32 type,
attr.config |= HSW_IN_TX_CHECKPOINTED;
}

+ if (pebs) {
+ /*
+ * Host never knows the precision level set by guest.
+ * Force Host's PEBS event to precision level 1, which will
+ * not impact the accuracy of the results for guest PEBS events.
+ * Because,
+ * - For most cases, there is no difference among precision
+ * level 1 to 3 for PEBS events.
+ * - The functions as below checks the precision level in host.
+ * But the results from these functions in host are replaced
+ * by guest when sampling the guest.
+ * The accuracy for guest PEBS events will not be impacted.
+ * -- event_constraints() impacts the index of counter.
+ * The index for host event is exactly the same as guest.
+ * It's decided by guest.
+ * -- pebs_update_adaptive_cfg() impacts the value of
+ * MSR_PEBS_DATA_CFG. When guest is switched in,
+ * the MSR value will be replaced by the value from guest.
+ * -- setup_sample () impacts the output of a PEBS record.
+ * Guest handles the PEBS records.
+ */
+ attr.precise_ip = 1;
+ /*
+ * When the host's PMI handler completes, it's going to
+ * enter the guest and trigger the guest's PMI handler.
+ *
+ * At this moment, this function may be called by
+ * kvm_pmu_handle_event(). However the next sample_period
+ * hasn't been determined by guest yet and the left period,
+ * which probably be 0, is used for current sample_period.
+ *
+ * In this case, perf will mistakenly treat it as non
+ * sampling events. The PEBS event will error out.
+ *
+ * Fill it with maximum period to prevent the error out.
+ * The guest PMI handler will soon reprogram the counter.
+ */
+ if (!attr.sample_period)
+ attr.sample_period = (-1ULL) & pmc_bitmask(pmc);
+ }
+
event = perf_event_create_kernel_counter(&attr, -1, current,
- intr ? kvm_perf_overflow_intr :
+ (intr || pebs) ?
+ kvm_perf_overflow_intr :
kvm_perf_overflow, pmc);
if (IS_ERR(event)) {
pr_debug_ratelimited("kvm_pmu: event creation failed %ld for pmc->idx = %d\n",
@@ -135,6 +184,20 @@ static void pmc_reprogram_counter(struct kvm_pmc *pmc, u32 type,
return;
}

+ if (pebs) {
+ event->guest_dedicated_idx = pmc->idx;
+ /*
+ * For guest PEBS events, guest takes the responsibility to
+ * drain PEBS buffers, and load proper values to reset counters.
+ *
+ * Host will unconditionally set auto-reload flag for PEBS
+ * events with fixed period which is not necessary. Host should
+ * do nothing in drain_pebs() but inject the PMI into the guest.
+ *
+ * Unset the auto-reload flag for guest PEBS events.
+ */
+ perf_x86_pmu_unset_auto_reload(event);
+ }
pmc->perf_event = event;
pmc_to_pmu(pmc)->event_count++;
perf_event_enable(event);
@@ -158,6 +221,10 @@ static bool pmc_resume_counter(struct kvm_pmc *pmc)
if (!pmc->perf_event)
return false;

+ if (test_bit(pmc->idx, (unsigned long *)&pmc_to_pmu(pmc)->pebs_enable)
+ != (!!pmc->perf_event->attr.precise_ip))
+ return false;
+
/* recalibrate sample period and check if it's accepted by perf core */
if (perf_event_period(pmc->perf_event,
(-pmc->counter) & pmc_bitmask(pmc)))
diff --git a/arch/x86/kvm/vmx/pmu_intel.c b/arch/x86/kvm/vmx/pmu_intel.c
index fd21cdb..ebadc33 100644
--- a/arch/x86/kvm/vmx/pmu_intel.c
+++ b/arch/x86/kvm/vmx/pmu_intel.c
@@ -293,6 +293,7 @@ static void intel_pmu_refresh(struct kvm_vcpu *vcpu)
pmu->counter_bitmask[KVM_PMC_GP] = 0;
pmu->counter_bitmask[KVM_PMC_FIXED] = 0;
pmu->version = 0;
+ pmu->pebs_enable = 0;
pmu->reserved_bits = 0xffffffff00200000ull;

entry = kvm_find_cpuid_entry(vcpu, 0xa, 0);
--
1.8.3.1