[PATCH 1/2] perf_events: improve X86 constraint management for fixed counters

From: Stephane Eranian
Date: Fri Feb 04 2011 - 08:06:08 EST


This patch adds a flags field to struct event_constraint. It is used
to mark constraint which need extra work on initialization.

The patch defines the X86_EVENT_CONSTRAINT_FL_ADD_GEN flag which indicates
that the constraint mask must be dynamically adjusted for generic counters.
In other words, the constraint is extended to include generic counters once
their number is known. This applies to Intel x86 processors supporting the
architectural perfmon PMU and fixed counters. Some events may support
either a fixed counter or a generic counter. Some events only work in fixed
counters, e.g., unhalted_reference_cycles.

Signed-off-by: Stephane Eranian <eranian@xxxxxxxxxx>
---

diff --git a/arch/x86/kernel/cpu/perf_event.c b/arch/x86/kernel/cpu/perf_event.c
index 4d98789..a06261a 100644
--- a/arch/x86/kernel/cpu/perf_event.c
+++ b/arch/x86/kernel/cpu/perf_event.c
@@ -84,7 +84,9 @@ struct event_constraint {
u64 code;
u64 cmask;
int weight;
+ int flags;
};
+#define X86_EVENT_CONSTRAINT_FL_ADD_GEN 0x1 /* add generic cntrs on the fly */

struct amd_nb {
int nb_id; /* NorthBridge id */
@@ -133,21 +135,22 @@ struct cpu_hw_events {
struct amd_nb *amd_nb;
};

-#define __EVENT_CONSTRAINT(c, n, m, w) {\
+#define __EVENT_CONSTRAINT(c, n, m, w, f) {\
{ .idxmsk64 = (n) }, \
.code = (c), \
.cmask = (m), \
.weight = (w), \
+ .flags = (f), \
}

-#define EVENT_CONSTRAINT(c, n, m) \
- __EVENT_CONSTRAINT(c, n, m, HWEIGHT(n))
+#define EVENT_CONSTRAINT(c, n, m, f) \
+ __EVENT_CONSTRAINT(c, n, m, HWEIGHT(n), f)

/*
* Constraint on the Event code.
*/
#define INTEL_EVENT_CONSTRAINT(c, n) \
- EVENT_CONSTRAINT(c, n, ARCH_PERFMON_EVENTSEL_EVENT)
+ EVENT_CONSTRAINT(c, n, ARCH_PERFMON_EVENTSEL_EVENT, 0)

/*
* Constraint on the Event code + UMask + fixed-mask
@@ -161,16 +164,18 @@ struct cpu_hw_events {
* The any-thread option is supported starting with v3.
*/
#define FIXED_EVENT_CONSTRAINT(c, n) \
- EVENT_CONSTRAINT(c, (1ULL << (32+n)), X86_RAW_EVENT_MASK)
+ EVENT_CONSTRAINT(c, (1ULL << (32+n)), \
+ X86_RAW_EVENT_MASK, \
+ X86_EVENT_CONSTRAINT_FL_ADD_GEN)

/*
* Constraint on the Event code + UMask
*/
#define PEBS_EVENT_CONSTRAINT(c, n) \
- EVENT_CONSTRAINT(c, n, INTEL_ARCH_EVENT_MASK)
+ EVENT_CONSTRAINT(c, n, INTEL_ARCH_EVENT_MASK, 0)

#define EVENT_CONSTRAINT_END \
- EVENT_CONSTRAINT(0, 0, 0)
+ EVENT_CONSTRAINT(0, 0, 0, 0)

#define for_each_event_constraint(e, c) \
for ((e) = (c); (e)->weight; (e)++)
@@ -1443,11 +1448,11 @@ static int __init init_hw_perf_events(void)

unconstrained = (struct event_constraint)
__EVENT_CONSTRAINT(0, (1ULL << x86_pmu.num_counters) - 1,
- 0, x86_pmu.num_counters);
+ 0, x86_pmu.num_counters, 0);

if (x86_pmu.event_constraints) {
for_each_event_constraint(c, x86_pmu.event_constraints) {
- if (c->cmask != X86_RAW_EVENT_MASK)
+ if (!(c->flags & X86_EVENT_CONSTRAINT_FL_ADD_GEN))
continue;

c->idxmsk64 |= (1ULL << x86_pmu.num_counters) - 1;
diff --git a/arch/x86/kernel/cpu/perf_event_intel_ds.c b/arch/x86/kernel/cpu/perf_event_intel_ds.c
index b7dcd9f..68530a4 100644
--- a/arch/x86/kernel/cpu/perf_event_intel_ds.c
+++ b/arch/x86/kernel/cpu/perf_event_intel_ds.c
@@ -261,7 +261,7 @@ static void reserve_ds_buffers(void)
*/

static struct event_constraint bts_constraint =
- EVENT_CONSTRAINT(0, 1ULL << X86_PMC_IDX_FIXED_BTS, 0);
+ EVENT_CONSTRAINT(0, 1ULL << X86_PMC_IDX_FIXED_BTS, 0, 0);

static void intel_pmu_enable_bts(u64 config)
{
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/