Re: [PATCH v4 10/11] perf/x86/intel: Perform rotation on Intel CQM RMIDs

From: Peter Zijlstra
Date: Tue Jan 06 2015 - 12:37:13 EST


On Fri, Nov 14, 2014 at 09:15:11PM +0000, Matt Fleming wrote:
> @@ -417,17 +857,38 @@ static u64 intel_cqm_event_count(struct perf_event *event)
> if (!cqm_group_leader(event))
> return 0;
>
> - on_each_cpu_mask(&cqm_cpumask, __intel_cqm_event_count, &rr, 1);
> + /*
> + * Notice that we don't perform the reading of an RMID
> + * atomically, because we can't hold a spin lock across the
> + * IPIs.
> + *
> + * Speculatively perform the read, since @event might be
> + * assigned a different (possibly invalid) RMID while we're
> + * busying performing the IPI calls. It's therefore necessary to
> + * check @event's RMID afterwards, and if it has changed,
> + * discard the result of the read.
> + */
> + raw_spin_lock_irqsave(&cache_lock, flags);
> + rr.rmid = event->hw.cqm_rmid;
> + raw_spin_unlock_irqrestore(&cache_lock, flags);

You don't actually have to hold the lock here, only ACCESS_ONCE() or
whatever newfangled thing replaced that.

> +
> + if (!__rmid_valid(rr.rmid))
> + goto out;
>
> - local64_set(&event->count, atomic64_read(&rr.value));
> + on_each_cpu_mask(&cqm_cpumask, __intel_cqm_event_count, &rr, 1);
>
> + raw_spin_lock_irqsave(&cache_lock, flags);
> + if (event->hw.cqm_rmid == rr.rmid)
> + local64_set(&event->count, atomic64_read(&rr.value));
> + raw_spin_unlock_irqrestore(&cache_lock, flags);

Here you do indeed need the lock as its more than a single op :-)

> +out:
> return __perf_event_count(event);
> }
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/