[PATCH 3/3] perf/x86/intel/ds: Support monotonic clock for PEBS

From: kan . liang
Date: Mon Jan 23 2023 - 13:30:50 EST


From: Kan Liang <kan.liang@xxxxxxxxxxxxxxx>

Users try to reconcile user samples with PEBS samples and require a
common clock source. However, the current PEBS codes only convert to
sched_clock, which is not available from the user space.

Only support converting to clock monotonic. Having one common clock
source is good enough to fulfill the requirement.

Enable the large PEBS for the monotonic clock to reduce the PEBS
overhead.

There are a few rare cases that may make the conversion fails. For
example, TSC overflows. The cycle_last may be changed between samples.
The time will fallback to the inaccurate SW times. But the cases are
extremely unlikely to happen.

Signed-off-by: Kan Liang <kan.liang@xxxxxxxxxxxxxxx>
---

The patch has to be on top of the below patch
https://lore.kernel.org/all/20230123172027.125385-1-kan.liang@xxxxxxxxxxxxxxx/

arch/x86/events/intel/core.c | 2 +-
arch/x86/events/intel/ds.c | 30 ++++++++++++++++++++++++++----
2 files changed, 27 insertions(+), 5 deletions(-)

diff --git a/arch/x86/events/intel/core.c b/arch/x86/events/intel/core.c
index 14f0a746257d..ea194556cc73 100644
--- a/arch/x86/events/intel/core.c
+++ b/arch/x86/events/intel/core.c
@@ -3777,7 +3777,7 @@ static unsigned long intel_pmu_large_pebs_flags(struct perf_event *event)
{
unsigned long flags = x86_pmu.large_pebs_flags;

- if (event->attr.use_clockid)
+ if (event->attr.use_clockid && (event->attr.clockid != CLOCK_MONOTONIC))
flags &= ~PERF_SAMPLE_TIME;
if (!event->attr.exclude_kernel)
flags &= ~PERF_SAMPLE_REGS_USER;
diff --git a/arch/x86/events/intel/ds.c b/arch/x86/events/intel/ds.c
index 7980e92dec64..d7f0eaf4405c 100644
--- a/arch/x86/events/intel/ds.c
+++ b/arch/x86/events/intel/ds.c
@@ -1570,13 +1570,33 @@ static u64 get_data_src(struct perf_event *event, u64 aux)
return val;
}

+static int pebs_get_synctime(struct system_counterval_t *system,
+ void *ctx)
+{
+ *system = set_tsc_system_counterval(*(u64 *)ctx);
+ return 0;
+}
+
+static inline int pebs_clockid_time(clockid_t clk_id, u64 tsc, u64 *clk_id_time)
+{
+ /* Only support converting to clock monotonic */
+ if (clk_id != CLOCK_MONOTONIC)
+ return -EINVAL;
+
+ return get_mono_fast_from_given_time(pebs_get_synctime, &tsc, clk_id_time);
+}
+
static void setup_pebs_time(struct perf_event *event,
struct perf_sample_data *data,
u64 tsc)
{
- /* Converting to a user-defined clock is not supported yet. */
- if (event->attr.use_clockid != 0)
- return;
+ u64 time;
+
+ if (event->attr.use_clockid != 0) {
+ if (pebs_clockid_time(event->attr.clockid, tsc, &time))
+ return;
+ goto done;
+ }

/*
* Converting the TSC to perf time is only supported,
@@ -1587,8 +1607,10 @@ static void setup_pebs_time(struct perf_event *event,
*/
if (!using_native_sched_clock() || !sched_clock_stable())
return;
+ time = native_sched_clock_from_tsc(tsc) + __sched_clock_offset;

- data->time = native_sched_clock_from_tsc(tsc) + __sched_clock_offset;
+done:
+ data->time = time;
data->sample_flags |= PERF_SAMPLE_TIME;
}

--
2.35.1