Commit 88081cfb authored by Anshuman Khandual's avatar Anshuman Khandual Committed by Peter Zijlstra
Browse files

x86/perf: Assert all platform event flags are within PERF_EVENT_FLAG_ARCH



Ensure all platform specific event flags are within PERF_EVENT_FLAG_ARCH.

Signed-off-by: default avatarAnshuman Khandual <anshuman.khandual@arm.com>
Signed-off-by: default avatarPeter Zijlstra (Intel) <peterz@infradead.org>
Reviewed-by: default avatarJames Clark <james.clark@arm.com>
Link: https://lkml.kernel.org/r/20220907091924.439193-5-anshuman.khandual@arm.com
parent 91207f62
Loading
Loading
Loading
Loading
+16 −18
Original line number Diff line number Diff line
@@ -64,27 +64,25 @@ static inline bool constraint_match(struct event_constraint *c, u64 ecode)
	return ((ecode & c->cmask) - c->code) <= (u64)c->size;
}

#define PERF_ARCH(name, val)	\
	PERF_X86_EVENT_##name = val,

/*
 * struct hw_perf_event.flags flags
 */
#define PERF_X86_EVENT_PEBS_LDLAT	0x00001 /* ld+ldlat data address sampling */
#define PERF_X86_EVENT_PEBS_ST		0x00002 /* st data address sampling */
#define PERF_X86_EVENT_PEBS_ST_HSW	0x00004 /* haswell style datala, store */
#define PERF_X86_EVENT_PEBS_LD_HSW	0x00008 /* haswell style datala, load */
#define PERF_X86_EVENT_PEBS_NA_HSW	0x00010 /* haswell style datala, unknown */
#define PERF_X86_EVENT_EXCL		0x00020 /* HT exclusivity on counter */
#define PERF_X86_EVENT_DYNAMIC		0x00040 /* dynamic alloc'd constraint */

#define PERF_X86_EVENT_EXCL_ACCT	0x00100 /* accounted EXCL event */
#define PERF_X86_EVENT_AUTO_RELOAD	0x00200 /* use PEBS auto-reload */
#define PERF_X86_EVENT_LARGE_PEBS	0x00400 /* use large PEBS */
#define PERF_X86_EVENT_PEBS_VIA_PT	0x00800 /* use PT buffer for PEBS */
#define PERF_X86_EVENT_PAIR		0x01000 /* Large Increment per Cycle */
#define PERF_X86_EVENT_LBR_SELECT	0x02000 /* Save/Restore MSR_LBR_SELECT */
#define PERF_X86_EVENT_TOPDOWN		0x04000 /* Count Topdown slots/metrics events */
#define PERF_X86_EVENT_PEBS_STLAT	0x08000 /* st+stlat data address sampling */
#define PERF_X86_EVENT_AMD_BRS		0x10000 /* AMD Branch Sampling */
#define PERF_X86_EVENT_PEBS_LAT_HYBRID	0x20000 /* ld and st lat for hybrid */
enum {
#include "perf_event_flags.h"
};

#undef PERF_ARCH

#define PERF_ARCH(name, val)						\
	static_assert((PERF_X86_EVENT_##name & PERF_EVENT_FLAG_ARCH) ==	\
		      PERF_X86_EVENT_##name);

#include "perf_event_flags.h"

#undef PERF_ARCH

static inline bool is_topdown_count(struct perf_event *event)
{
+22 −0
Original line number Diff line number Diff line

/*
 * struct hw_perf_event.flags flags
 */
PERF_ARCH(PEBS_LDLAT,		0x00001) /* ld+ldlat data address sampling */
PERF_ARCH(PEBS_ST,		0x00002) /* st data address sampling */
PERF_ARCH(PEBS_ST_HSW,		0x00004) /* haswell style datala, store */
PERF_ARCH(PEBS_LD_HSW,		0x00008) /* haswell style datala, load */
PERF_ARCH(PEBS_NA_HSW,		0x00010) /* haswell style datala, unknown */
PERF_ARCH(EXCL,			0x00020) /* HT exclusivity on counter */
PERF_ARCH(DYNAMIC,		0x00040) /* dynamic alloc'd constraint */
			/*	0x00080	*/
PERF_ARCH(EXCL_ACCT,		0x00100) /* accounted EXCL event */
PERF_ARCH(AUTO_RELOAD,		0x00200) /* use PEBS auto-reload */
PERF_ARCH(LARGE_PEBS,		0x00400) /* use large PEBS */
PERF_ARCH(PEBS_VIA_PT,		0x00800) /* use PT buffer for PEBS */
PERF_ARCH(PAIR,			0x01000) /* Large Increment per Cycle */
PERF_ARCH(LBR_SELECT,		0x02000) /* Save/Restore MSR_LBR_SELECT */
PERF_ARCH(TOPDOWN,		0x04000) /* Count Topdown slots/metrics events */
PERF_ARCH(PEBS_STLAT,		0x08000) /* st+stlat data address sampling */
PERF_ARCH(AMD_BRS,		0x10000) /* AMD Branch Sampling */
PERF_ARCH(PEBS_LAT_HYBRID,	0x20000) /* ld and st lat for hybrid */