Skip to content

Commit 4e955c0

Browse files
Dapeng MiPeter Zijlstra
authored andcommitted
perf/x86/intel: Support the 4 new OMR MSRs introduced in DMR and NVL
Diamond Rapids (DMR) and Nova Lake (NVL) introduce an enhanced Off-Module Response (OMR) facility, replacing the Off-Core Response (OCR) Performance Monitoring of previous processors. Legacy microarchitectures used the OCR facility to evaluate off-core and multi-core off-module transactions. The newly named OMR facility improves OCR capabilities for scalable coverage of new memory systems in multi-core module systems. Similar to OCR, 4 additional off-module configuration MSRs (OFFMODULE_RSP_0 to OFFMODULE_RSP_3) are introduced to specify attributes of off-module transactions. When multiple identical OMR events are created, they need to occupy the same OFFMODULE_RSP_x MSR. To ensure these multiple identical OMR events can work simultaneously, the intel_alt_er() and intel_fixup_er() helpers are enhanced to rotate these OMR events across different OFFMODULE_RSP_* MSRs, similar to previous OCR events. For more details about OMR, please refer to section 16.1 "OFF-MODULE RESPONSE (OMR) FACILITY" in ISE documentation. Signed-off-by: Dapeng Mi <dapeng1.mi@linux.intel.com> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Link: https://patch.msgid.link/20260114011750.350569-2-dapeng1.mi@linux.intel.com
1 parent 4960626 commit 4e955c0

3 files changed

Lines changed: 52 additions & 17 deletions

File tree

arch/x86/events/intel/core.c

Lines changed: 42 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -3532,34 +3532,59 @@ static int intel_alt_er(struct cpu_hw_events *cpuc,
35323532
struct extra_reg *extra_regs = hybrid(cpuc->pmu, extra_regs);
35333533
int alt_idx = idx;
35343534

3535-
if (!(x86_pmu.flags & PMU_FL_HAS_RSP_1))
3536-
return idx;
3537-
3538-
if (idx == EXTRA_REG_RSP_0)
3539-
alt_idx = EXTRA_REG_RSP_1;
3535+
switch (idx) {
3536+
case EXTRA_REG_RSP_0 ... EXTRA_REG_RSP_1:
3537+
if (!(x86_pmu.flags & PMU_FL_HAS_RSP_1))
3538+
return idx;
3539+
if (++alt_idx > EXTRA_REG_RSP_1)
3540+
alt_idx = EXTRA_REG_RSP_0;
3541+
if (config & ~extra_regs[alt_idx].valid_mask)
3542+
return idx;
3543+
break;
35403544

3541-
if (idx == EXTRA_REG_RSP_1)
3542-
alt_idx = EXTRA_REG_RSP_0;
3545+
case EXTRA_REG_OMR_0 ... EXTRA_REG_OMR_3:
3546+
if (!(x86_pmu.flags & PMU_FL_HAS_OMR))
3547+
return idx;
3548+
if (++alt_idx > EXTRA_REG_OMR_3)
3549+
alt_idx = EXTRA_REG_OMR_0;
3550+
/*
3551+
* Subtracting EXTRA_REG_OMR_0 ensures to get correct
3552+
* OMR extra_reg entries which start from 0.
3553+
*/
3554+
if (config & ~extra_regs[alt_idx - EXTRA_REG_OMR_0].valid_mask)
3555+
return idx;
3556+
break;
35433557

3544-
if (config & ~extra_regs[alt_idx].valid_mask)
3545-
return idx;
3558+
default:
3559+
break;
3560+
}
35463561

35473562
return alt_idx;
35483563
}
35493564

35503565
static void intel_fixup_er(struct perf_event *event, int idx)
35513566
{
35523567
struct extra_reg *extra_regs = hybrid(event->pmu, extra_regs);
3553-
event->hw.extra_reg.idx = idx;
3568+
int er_idx;
35543569

3555-
if (idx == EXTRA_REG_RSP_0) {
3556-
event->hw.config &= ~INTEL_ARCH_EVENT_MASK;
3557-
event->hw.config |= extra_regs[EXTRA_REG_RSP_0].event;
3558-
event->hw.extra_reg.reg = MSR_OFFCORE_RSP_0;
3559-
} else if (idx == EXTRA_REG_RSP_1) {
3570+
event->hw.extra_reg.idx = idx;
3571+
switch (idx) {
3572+
case EXTRA_REG_RSP_0 ... EXTRA_REG_RSP_1:
3573+
er_idx = idx - EXTRA_REG_RSP_0;
35603574
event->hw.config &= ~INTEL_ARCH_EVENT_MASK;
3561-
event->hw.config |= extra_regs[EXTRA_REG_RSP_1].event;
3562-
event->hw.extra_reg.reg = MSR_OFFCORE_RSP_1;
3575+
event->hw.config |= extra_regs[er_idx].event;
3576+
event->hw.extra_reg.reg = MSR_OFFCORE_RSP_0 + er_idx;
3577+
break;
3578+
3579+
case EXTRA_REG_OMR_0 ... EXTRA_REG_OMR_3:
3580+
er_idx = idx - EXTRA_REG_OMR_0;
3581+
event->hw.config &= ~ARCH_PERFMON_EVENTSEL_UMASK;
3582+
event->hw.config |= 1ULL << (8 + er_idx);
3583+
event->hw.extra_reg.reg = MSR_OMR_0 + er_idx;
3584+
break;
3585+
3586+
default:
3587+
pr_warn("The extra reg idx %d is not supported.\n", idx);
35633588
}
35643589
}
35653590

arch/x86/events/perf_event.h

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -45,6 +45,10 @@ enum extra_reg_type {
4545
EXTRA_REG_FE = 4, /* fe_* */
4646
EXTRA_REG_SNOOP_0 = 5, /* snoop response 0 */
4747
EXTRA_REG_SNOOP_1 = 6, /* snoop response 1 */
48+
EXTRA_REG_OMR_0 = 7, /* OMR 0 */
49+
EXTRA_REG_OMR_1 = 8, /* OMR 1 */
50+
EXTRA_REG_OMR_2 = 9, /* OMR 2 */
51+
EXTRA_REG_OMR_3 = 10, /* OMR 3 */
4852

4953
EXTRA_REG_MAX /* number of entries needed */
5054
};
@@ -1099,6 +1103,7 @@ do { \
10991103
#define PMU_FL_RETIRE_LATENCY 0x200 /* Support Retire Latency in PEBS */
11001104
#define PMU_FL_BR_CNTR 0x400 /* Support branch counter logging */
11011105
#define PMU_FL_DYN_CONSTRAINT 0x800 /* Needs dynamic constraint */
1106+
#define PMU_FL_HAS_OMR 0x1000 /* has 4 equivalent OMR regs */
11021107

11031108
#define EVENT_VAR(_id) event_attr_##_id
11041109
#define EVENT_PTR(_id) &event_attr_##_id.attr.attr

arch/x86/include/asm/msr-index.h

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -263,6 +263,11 @@
263263
#define MSR_SNOOP_RSP_0 0x00001328
264264
#define MSR_SNOOP_RSP_1 0x00001329
265265

266+
#define MSR_OMR_0 0x000003e0
267+
#define MSR_OMR_1 0x000003e1
268+
#define MSR_OMR_2 0x000003e2
269+
#define MSR_OMR_3 0x000003e3
270+
266271
#define MSR_LBR_SELECT 0x000001c8
267272
#define MSR_LBR_TOS 0x000001c9
268273

0 commit comments

Comments
 (0)