xref: /freebsd/sys/dev/hwpmc/hwpmc_core.c (revision c7046f76)
1 /*-
2  * SPDX-License-Identifier: BSD-2-Clause-FreeBSD
3  *
4  * Copyright (c) 2008 Joseph Koshy
5  * All rights reserved.
6  *
7  * Redistribution and use in source and binary forms, with or without
8  * modification, are permitted provided that the following conditions
9  * are met:
10  * 1. Redistributions of source code must retain the above copyright
11  *    notice, this list of conditions and the following disclaimer.
12  * 2. Redistributions in binary form must reproduce the above copyright
13  *    notice, this list of conditions and the following disclaimer in the
14  *    documentation and/or other materials provided with the distribution.
15  *
16  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
17  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
18  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
19  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
20  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
21  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
22  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
23  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
24  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
25  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
26  * SUCH DAMAGE.
27  */
28 
29 /*
30  * Intel Core PMCs.
31  */
32 
33 #include <sys/cdefs.h>
34 __FBSDID("$FreeBSD$");
35 
36 #include <sys/param.h>
37 #include <sys/bus.h>
38 #include <sys/pmc.h>
39 #include <sys/pmckern.h>
40 #include <sys/smp.h>
41 #include <sys/systm.h>
42 
43 #include <machine/intr_machdep.h>
44 #include <x86/apicvar.h>
45 #include <machine/cpu.h>
46 #include <machine/cpufunc.h>
47 #include <machine/md_var.h>
48 #include <machine/specialreg.h>
49 
50 #define	CORE_CPUID_REQUEST		0xA
51 #define	CORE_CPUID_REQUEST_SIZE		0x4
52 #define	CORE_CPUID_EAX			0x0
53 #define	CORE_CPUID_EBX			0x1
54 #define	CORE_CPUID_ECX			0x2
55 #define	CORE_CPUID_EDX			0x3
56 
57 #define	IAF_PMC_CAPS			\
58 	(PMC_CAP_READ | PMC_CAP_WRITE | PMC_CAP_INTERRUPT | \
59 	 PMC_CAP_USER | PMC_CAP_SYSTEM)
60 #define	IAF_RI_TO_MSR(RI)		((RI) + (1 << 30))
61 
62 #define	IAP_PMC_CAPS (PMC_CAP_INTERRUPT | PMC_CAP_USER | PMC_CAP_SYSTEM | \
63     PMC_CAP_EDGE | PMC_CAP_THRESHOLD | PMC_CAP_READ | PMC_CAP_WRITE |	 \
64     PMC_CAP_INVERT | PMC_CAP_QUALIFIER | PMC_CAP_PRECISE)
65 
66 #define	EV_IS_NOTARCH		0
67 #define	EV_IS_ARCH_SUPP		1
68 #define	EV_IS_ARCH_NOTSUPP	-1
69 
70 /*
71  * "Architectural" events defined by Intel.  The values of these
72  * symbols correspond to positions in the bitmask returned by
73  * the CPUID.0AH instruction.
74  */
75 enum core_arch_events {
76 	CORE_AE_BRANCH_INSTRUCTION_RETIRED	= 5,
77 	CORE_AE_BRANCH_MISSES_RETIRED		= 6,
78 	CORE_AE_INSTRUCTION_RETIRED		= 1,
79 	CORE_AE_LLC_MISSES			= 4,
80 	CORE_AE_LLC_REFERENCE			= 3,
81 	CORE_AE_UNHALTED_REFERENCE_CYCLES	= 2,
82 	CORE_AE_UNHALTED_CORE_CYCLES		= 0
83 };
84 
85 static enum pmc_cputype	core_cputype;
86 static int core_version;
87 
88 struct core_cpu {
89 	volatile uint32_t	pc_iafctrl;	/* Fixed function control. */
90 	volatile uint64_t	pc_globalctrl;	/* Global control register. */
91 	struct pmc_hw		pc_corepmcs[];
92 };
93 
94 static struct core_cpu **core_pcpu;
95 
96 static uint32_t core_architectural_events;
97 static uint64_t core_pmcmask;
98 
99 static int core_iaf_ri;		/* relative index of fixed counters */
100 static int core_iaf_width;
101 static int core_iaf_npmc;
102 
103 static int core_iap_width;
104 static int core_iap_npmc;
105 static int core_iap_wroffset;
106 
107 static u_int pmc_alloc_refs;
108 static bool pmc_tsx_force_abort_set;
109 
110 static int
111 core_pcpu_noop(struct pmc_mdep *md, int cpu)
112 {
113 	(void) md;
114 	(void) cpu;
115 	return (0);
116 }
117 
118 static int
119 core_pcpu_init(struct pmc_mdep *md, int cpu)
120 {
121 	struct pmc_cpu *pc;
122 	struct core_cpu *cc;
123 	struct pmc_hw *phw;
124 	int core_ri, n, npmc;
125 
126 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
127 	    ("[iaf,%d] insane cpu number %d", __LINE__, cpu));
128 
129 	PMCDBG1(MDP,INI,1,"core-init cpu=%d", cpu);
130 
131 	core_ri = md->pmd_classdep[PMC_MDEP_CLASS_INDEX_IAP].pcd_ri;
132 	npmc = md->pmd_classdep[PMC_MDEP_CLASS_INDEX_IAP].pcd_num;
133 
134 	if (core_version >= 2)
135 		npmc += md->pmd_classdep[PMC_MDEP_CLASS_INDEX_IAF].pcd_num;
136 
137 	cc = malloc(sizeof(struct core_cpu) + npmc * sizeof(struct pmc_hw),
138 	    M_PMC, M_WAITOK | M_ZERO);
139 
140 	core_pcpu[cpu] = cc;
141 	pc = pmc_pcpu[cpu];
142 
143 	KASSERT(pc != NULL && cc != NULL,
144 	    ("[core,%d] NULL per-cpu structures cpu=%d", __LINE__, cpu));
145 
146 	for (n = 0, phw = cc->pc_corepmcs; n < npmc; n++, phw++) {
147 		phw->phw_state 	  = PMC_PHW_FLAG_IS_ENABLED |
148 		    PMC_PHW_CPU_TO_STATE(cpu) |
149 		    PMC_PHW_INDEX_TO_STATE(n + core_ri);
150 		phw->phw_pmc	  = NULL;
151 		pc->pc_hwpmcs[n + core_ri]  = phw;
152 	}
153 
154 	if (core_version >= 2 && vm_guest == VM_GUEST_NO) {
155 		/* Enable Freezing PMCs on PMI. */
156 		wrmsr(MSR_DEBUGCTLMSR, rdmsr(MSR_DEBUGCTLMSR) | 0x1000);
157 	}
158 
159 	return (0);
160 }
161 
162 static int
163 core_pcpu_fini(struct pmc_mdep *md, int cpu)
164 {
165 	int core_ri, n, npmc;
166 	struct pmc_cpu *pc;
167 	struct core_cpu *cc;
168 
169 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
170 	    ("[core,%d] insane cpu number (%d)", __LINE__, cpu));
171 
172 	PMCDBG1(MDP,INI,1,"core-pcpu-fini cpu=%d", cpu);
173 
174 	if ((cc = core_pcpu[cpu]) == NULL)
175 		return (0);
176 
177 	core_pcpu[cpu] = NULL;
178 
179 	pc = pmc_pcpu[cpu];
180 
181 	KASSERT(pc != NULL, ("[core,%d] NULL per-cpu %d state", __LINE__,
182 		cpu));
183 
184 	npmc = md->pmd_classdep[PMC_MDEP_CLASS_INDEX_IAP].pcd_num;
185 	core_ri = md->pmd_classdep[PMC_MDEP_CLASS_INDEX_IAP].pcd_ri;
186 
187 	for (n = 0; n < npmc; n++)
188 		wrmsr(IAP_EVSEL0 + n, 0);
189 
190 	if (core_version >= 2) {
191 		wrmsr(IAF_CTRL, 0);
192 		npmc += md->pmd_classdep[PMC_MDEP_CLASS_INDEX_IAF].pcd_num;
193 	}
194 
195 	for (n = 0; n < npmc; n++)
196 		pc->pc_hwpmcs[n + core_ri] = NULL;
197 
198 	free(cc, M_PMC);
199 
200 	return (0);
201 }
202 
203 /*
204  * Fixed function counters.
205  */
206 
207 static pmc_value_t
208 iaf_perfctr_value_to_reload_count(pmc_value_t v)
209 {
210 
211 	/* If the PMC has overflowed, return a reload count of zero. */
212 	if ((v & (1ULL << (core_iaf_width - 1))) == 0)
213 		return (0);
214 	v &= (1ULL << core_iaf_width) - 1;
215 	return (1ULL << core_iaf_width) - v;
216 }
217 
218 static pmc_value_t
219 iaf_reload_count_to_perfctr_value(pmc_value_t rlc)
220 {
221 	return (1ULL << core_iaf_width) - rlc;
222 }
223 
224 static int
225 iaf_allocate_pmc(int cpu, int ri, struct pmc *pm,
226     const struct pmc_op_pmcallocate *a)
227 {
228 	uint8_t ev, umask;
229 	uint32_t caps;
230 	uint64_t config, flags;
231 	const struct pmc_md_iap_op_pmcallocate *iap;
232 
233 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
234 	    ("[core,%d] illegal CPU %d", __LINE__, cpu));
235 
236 	PMCDBG2(MDP,ALL,1, "iaf-allocate ri=%d reqcaps=0x%x", ri, pm->pm_caps);
237 
238 	if (ri < 0 || ri > core_iaf_npmc)
239 		return (EINVAL);
240 
241 	if (a->pm_class != PMC_CLASS_IAF)
242 		return (EINVAL);
243 
244 	iap = &a->pm_md.pm_iap;
245 	config = iap->pm_iap_config;
246 	ev = IAP_EVSEL_GET(config);
247 	umask = IAP_UMASK_GET(config);
248 
249 	if (ev == 0x0) {
250 		if (umask != ri + 1)
251 			return (EINVAL);
252 	} else {
253 		switch (ri) {
254 		case 0:	/* INST_RETIRED.ANY */
255 			if (ev != 0xC0 || umask != 0x00)
256 				return (EINVAL);
257 			break;
258 		case 1:	/* CPU_CLK_UNHALTED.THREAD */
259 			if (ev != 0x3C || umask != 0x00)
260 				return (EINVAL);
261 			break;
262 		case 2:	/* CPU_CLK_UNHALTED.REF */
263 			if (ev != 0x3C || umask != 0x01)
264 				return (EINVAL);
265 			break;
266 		case 3:	/* TOPDOWN.SLOTS */
267 			if (ev != 0xA4 || umask != 0x01)
268 				return (EINVAL);
269 			break;
270 		default:
271 			return (EINVAL);
272 		}
273 	}
274 
275 	pmc_alloc_refs++;
276 	if ((cpu_stdext_feature3 & CPUID_STDEXT3_TSXFA) != 0 &&
277 	    !pmc_tsx_force_abort_set) {
278 		pmc_tsx_force_abort_set = true;
279 		x86_msr_op(MSR_TSX_FORCE_ABORT, MSR_OP_RENDEZVOUS_ALL |
280 		    MSR_OP_WRITE, 1, NULL);
281 	}
282 
283 	flags = 0;
284 	if (config & IAP_OS)
285 		flags |= IAF_OS;
286 	if (config & IAP_USR)
287 		flags |= IAF_USR;
288 	if (config & IAP_ANY)
289 		flags |= IAF_ANY;
290 	if (config & IAP_INT)
291 		flags |= IAF_PMI;
292 
293 	caps = a->pm_caps;
294 	if (caps & PMC_CAP_INTERRUPT)
295 		flags |= IAF_PMI;
296 	if (caps & PMC_CAP_SYSTEM)
297 		flags |= IAF_OS;
298 	if (caps & PMC_CAP_USER)
299 		flags |= IAF_USR;
300 	if ((caps & (PMC_CAP_USER | PMC_CAP_SYSTEM)) == 0)
301 		flags |= (IAF_OS | IAF_USR);
302 
303 	pm->pm_md.pm_iaf.pm_iaf_ctrl = (flags << (ri * 4));
304 
305 	PMCDBG1(MDP,ALL,2, "iaf-allocate config=0x%jx",
306 	    (uintmax_t) pm->pm_md.pm_iaf.pm_iaf_ctrl);
307 
308 	return (0);
309 }
310 
311 static int
312 iaf_config_pmc(int cpu, int ri, struct pmc *pm)
313 {
314 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
315 	    ("[core,%d] illegal CPU %d", __LINE__, cpu));
316 
317 	KASSERT(ri >= 0 && ri < core_iaf_npmc,
318 	    ("[core,%d] illegal row-index %d", __LINE__, ri));
319 
320 	PMCDBG3(MDP,CFG,1, "iaf-config cpu=%d ri=%d pm=%p", cpu, ri, pm);
321 
322 	KASSERT(core_pcpu[cpu] != NULL, ("[core,%d] null per-cpu %d", __LINE__,
323 	    cpu));
324 
325 	core_pcpu[cpu]->pc_corepmcs[ri + core_iaf_ri].phw_pmc = pm;
326 
327 	return (0);
328 }
329 
330 static int
331 iaf_describe(int cpu, int ri, struct pmc_info *pi, struct pmc **ppmc)
332 {
333 	int error;
334 	struct pmc_hw *phw;
335 	char iaf_name[PMC_NAME_MAX];
336 
337 	phw = &core_pcpu[cpu]->pc_corepmcs[ri + core_iaf_ri];
338 
339 	(void) snprintf(iaf_name, sizeof(iaf_name), "IAF-%d", ri);
340 	if ((error = copystr(iaf_name, pi->pm_name, PMC_NAME_MAX,
341 	    NULL)) != 0)
342 		return (error);
343 
344 	pi->pm_class = PMC_CLASS_IAF;
345 
346 	if (phw->phw_state & PMC_PHW_FLAG_IS_ENABLED) {
347 		pi->pm_enabled = TRUE;
348 		*ppmc          = phw->phw_pmc;
349 	} else {
350 		pi->pm_enabled = FALSE;
351 		*ppmc          = NULL;
352 	}
353 
354 	return (0);
355 }
356 
357 static int
358 iaf_get_config(int cpu, int ri, struct pmc **ppm)
359 {
360 	*ppm = core_pcpu[cpu]->pc_corepmcs[ri + core_iaf_ri].phw_pmc;
361 
362 	return (0);
363 }
364 
365 static int
366 iaf_get_msr(int ri, uint32_t *msr)
367 {
368 	KASSERT(ri >= 0 && ri < core_iaf_npmc,
369 	    ("[iaf,%d] ri %d out of range", __LINE__, ri));
370 
371 	*msr = IAF_RI_TO_MSR(ri);
372 
373 	return (0);
374 }
375 
376 static int
377 iaf_read_pmc(int cpu, int ri, pmc_value_t *v)
378 {
379 	struct pmc *pm;
380 	pmc_value_t tmp;
381 
382 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
383 	    ("[core,%d] illegal cpu value %d", __LINE__, cpu));
384 	KASSERT(ri >= 0 && ri < core_iaf_npmc,
385 	    ("[core,%d] illegal row-index %d", __LINE__, ri));
386 
387 	pm = core_pcpu[cpu]->pc_corepmcs[ri + core_iaf_ri].phw_pmc;
388 
389 	KASSERT(pm,
390 	    ("[core,%d] cpu %d ri %d(%d) pmc not configured", __LINE__, cpu,
391 		ri, ri + core_iaf_ri));
392 
393 	tmp = rdpmc(IAF_RI_TO_MSR(ri));
394 
395 	if (PMC_IS_SAMPLING_MODE(PMC_TO_MODE(pm)))
396 		*v = iaf_perfctr_value_to_reload_count(tmp);
397 	else
398 		*v = tmp & ((1ULL << core_iaf_width) - 1);
399 
400 	PMCDBG4(MDP,REA,1, "iaf-read cpu=%d ri=%d msr=0x%x -> v=%jx", cpu, ri,
401 	    IAF_RI_TO_MSR(ri), *v);
402 
403 	return (0);
404 }
405 
406 static int
407 iaf_release_pmc(int cpu, int ri, struct pmc *pmc)
408 {
409 	PMCDBG3(MDP,REL,1, "iaf-release cpu=%d ri=%d pm=%p", cpu, ri, pmc);
410 
411 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
412 	    ("[core,%d] illegal CPU value %d", __LINE__, cpu));
413 	KASSERT(ri >= 0 && ri < core_iaf_npmc,
414 	    ("[core,%d] illegal row-index %d", __LINE__, ri));
415 
416 	KASSERT(core_pcpu[cpu]->pc_corepmcs[ri + core_iaf_ri].phw_pmc == NULL,
417 	    ("[core,%d] PHW pmc non-NULL", __LINE__));
418 
419 	MPASS(pmc_alloc_refs > 0);
420 	if (pmc_alloc_refs-- == 1 && pmc_tsx_force_abort_set) {
421 		pmc_tsx_force_abort_set = false;
422 		x86_msr_op(MSR_TSX_FORCE_ABORT, MSR_OP_RENDEZVOUS_ALL |
423 		    MSR_OP_WRITE, 0, NULL);
424 	}
425 
426 	return (0);
427 }
428 
429 static int
430 iaf_start_pmc(int cpu, int ri)
431 {
432 	struct pmc *pm;
433 	struct core_cpu *cc;
434 
435 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
436 	    ("[core,%d] illegal CPU value %d", __LINE__, cpu));
437 	KASSERT(ri >= 0 && ri < core_iaf_npmc,
438 	    ("[core,%d] illegal row-index %d", __LINE__, ri));
439 
440 	PMCDBG2(MDP,STA,1,"iaf-start cpu=%d ri=%d", cpu, ri);
441 
442 	cc = core_pcpu[cpu];
443 	pm = cc->pc_corepmcs[ri + core_iaf_ri].phw_pmc;
444 
445 	cc->pc_iafctrl |= pm->pm_md.pm_iaf.pm_iaf_ctrl;
446 	wrmsr(IAF_CTRL, cc->pc_iafctrl);
447 
448 	cc->pc_globalctrl |= (1ULL << (ri + IAF_OFFSET));
449 	wrmsr(IA_GLOBAL_CTRL, cc->pc_globalctrl);
450 
451 	PMCDBG4(MDP,STA,1,"iafctrl=%x(%x) globalctrl=%jx(%jx)",
452 	    cc->pc_iafctrl, (uint32_t) rdmsr(IAF_CTRL),
453 	    cc->pc_globalctrl, rdmsr(IA_GLOBAL_CTRL));
454 
455 	return (0);
456 }
457 
458 static int
459 iaf_stop_pmc(int cpu, int ri)
460 {
461 	struct core_cpu *cc;
462 
463 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
464 	    ("[core,%d] illegal CPU value %d", __LINE__, cpu));
465 	KASSERT(ri >= 0 && ri < core_iaf_npmc,
466 	    ("[core,%d] illegal row-index %d", __LINE__, ri));
467 
468 	PMCDBG2(MDP,STA,1,"iaf-stop cpu=%d ri=%d", cpu, ri);
469 
470 	cc = core_pcpu[cpu];
471 
472 	cc->pc_iafctrl &= ~(IAF_MASK << (ri * 4));
473 	wrmsr(IAF_CTRL, cc->pc_iafctrl);
474 
475 	/* Don't need to write IA_GLOBAL_CTRL, one disable is enough. */
476 
477 	PMCDBG4(MDP,STO,1,"iafctrl=%x(%x) globalctrl=%jx(%jx)",
478 	    cc->pc_iafctrl, (uint32_t) rdmsr(IAF_CTRL),
479 	    cc->pc_globalctrl, rdmsr(IA_GLOBAL_CTRL));
480 
481 	return (0);
482 }
483 
484 static int
485 iaf_write_pmc(int cpu, int ri, pmc_value_t v)
486 {
487 	struct core_cpu *cc;
488 	struct pmc *pm;
489 
490 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
491 	    ("[core,%d] illegal cpu value %d", __LINE__, cpu));
492 	KASSERT(ri >= 0 && ri < core_iaf_npmc,
493 	    ("[core,%d] illegal row-index %d", __LINE__, ri));
494 
495 	cc = core_pcpu[cpu];
496 	pm = cc->pc_corepmcs[ri + core_iaf_ri].phw_pmc;
497 
498 	KASSERT(pm,
499 	    ("[core,%d] cpu %d ri %d pmc not configured", __LINE__, cpu, ri));
500 
501 	if (PMC_IS_SAMPLING_MODE(PMC_TO_MODE(pm)))
502 		v = iaf_reload_count_to_perfctr_value(v);
503 
504 	/* Turn off the fixed counter */
505 	wrmsr(IAF_CTRL, cc->pc_iafctrl & ~(IAF_MASK << (ri * 4)));
506 
507 	wrmsr(IAF_CTR0 + ri, v & ((1ULL << core_iaf_width) - 1));
508 
509 	/* Turn on fixed counters */
510 	wrmsr(IAF_CTRL, cc->pc_iafctrl);
511 
512 	PMCDBG6(MDP,WRI,1, "iaf-write cpu=%d ri=%d msr=0x%x v=%jx iafctrl=%jx "
513 	    "pmc=%jx", cpu, ri, IAF_RI_TO_MSR(ri), v,
514 	    (uintmax_t) rdmsr(IAF_CTRL),
515 	    (uintmax_t) rdpmc(IAF_RI_TO_MSR(ri)));
516 
517 	return (0);
518 }
519 
520 
521 static void
522 iaf_initialize(struct pmc_mdep *md, int maxcpu, int npmc, int pmcwidth)
523 {
524 	struct pmc_classdep *pcd;
525 
526 	KASSERT(md != NULL, ("[iaf,%d] md is NULL", __LINE__));
527 
528 	PMCDBG0(MDP,INI,1, "iaf-initialize");
529 
530 	pcd = &md->pmd_classdep[PMC_MDEP_CLASS_INDEX_IAF];
531 
532 	pcd->pcd_caps	= IAF_PMC_CAPS;
533 	pcd->pcd_class	= PMC_CLASS_IAF;
534 	pcd->pcd_num	= npmc;
535 	pcd->pcd_ri	= md->pmd_npmc;
536 	pcd->pcd_width	= pmcwidth;
537 
538 	pcd->pcd_allocate_pmc	= iaf_allocate_pmc;
539 	pcd->pcd_config_pmc	= iaf_config_pmc;
540 	pcd->pcd_describe	= iaf_describe;
541 	pcd->pcd_get_config	= iaf_get_config;
542 	pcd->pcd_get_msr	= iaf_get_msr;
543 	pcd->pcd_pcpu_fini	= core_pcpu_noop;
544 	pcd->pcd_pcpu_init	= core_pcpu_noop;
545 	pcd->pcd_read_pmc	= iaf_read_pmc;
546 	pcd->pcd_release_pmc	= iaf_release_pmc;
547 	pcd->pcd_start_pmc	= iaf_start_pmc;
548 	pcd->pcd_stop_pmc	= iaf_stop_pmc;
549 	pcd->pcd_write_pmc	= iaf_write_pmc;
550 
551 	md->pmd_npmc	       += npmc;
552 }
553 
554 /*
555  * Intel programmable PMCs.
556  */
557 
558 /* Sub fields of UMASK that this event supports. */
559 #define	IAP_M_CORE		(1 << 0) /* Core specificity */
560 #define	IAP_M_AGENT		(1 << 1) /* Agent specificity */
561 #define	IAP_M_PREFETCH		(1 << 2) /* Prefetch */
562 #define	IAP_M_MESI		(1 << 3) /* MESI */
563 #define	IAP_M_SNOOPRESPONSE	(1 << 4) /* Snoop response */
564 #define	IAP_M_SNOOPTYPE		(1 << 5) /* Snoop type */
565 #define	IAP_M_TRANSITION	(1 << 6) /* Transition */
566 
567 #define	IAP_F_CORE		(0x3 << 14) /* Core specificity */
568 #define	IAP_F_AGENT		(0x1 << 13) /* Agent specificity */
569 #define	IAP_F_PREFETCH		(0x3 << 12) /* Prefetch */
570 #define	IAP_F_MESI		(0xF <<  8) /* MESI */
571 #define	IAP_F_SNOOPRESPONSE	(0xB <<  8) /* Snoop response */
572 #define	IAP_F_SNOOPTYPE		(0x3 <<  8) /* Snoop type */
573 #define	IAP_F_TRANSITION	(0x1 << 12) /* Transition */
574 
575 #define	IAP_PREFETCH_RESERVED	(0x2 << 12)
576 #define	IAP_CORE_THIS		(0x1 << 14)
577 #define	IAP_CORE_ALL		(0x3 << 14)
578 #define	IAP_F_CMASK		0xFF000000
579 
580 static pmc_value_t
581 iap_perfctr_value_to_reload_count(pmc_value_t v)
582 {
583 
584 	/* If the PMC has overflowed, return a reload count of zero. */
585 	if ((v & (1ULL << (core_iap_width - 1))) == 0)
586 		return (0);
587 	v &= (1ULL << core_iap_width) - 1;
588 	return (1ULL << core_iap_width) - v;
589 }
590 
591 static pmc_value_t
592 iap_reload_count_to_perfctr_value(pmc_value_t rlc)
593 {
594 	return (1ULL << core_iap_width) - rlc;
595 }
596 
597 static int
598 iap_pmc_has_overflowed(int ri)
599 {
600 	uint64_t v;
601 
602 	/*
603 	 * We treat a Core (i.e., Intel architecture v1) PMC as has
604 	 * having overflowed if its MSB is zero.
605 	 */
606 	v = rdpmc(ri);
607 	return ((v & (1ULL << (core_iap_width - 1))) == 0);
608 }
609 
610 static int
611 iap_event_corei7_ok_on_counter(uint8_t evsel, int ri)
612 {
613 	uint32_t mask;
614 
615 	switch (evsel) {
616 	/* Events valid only on counter 0, 1. */
617 	case 0x40:
618 	case 0x41:
619 	case 0x42:
620 	case 0x43:
621 	case 0x4C:
622 	case 0x4E:
623 	case 0x51:
624 	case 0x52:
625 	case 0x53:
626 	case 0x63:
627 		mask = 0x3;
628 		break;
629 	/* Any row index is ok. */
630 	default:
631 		mask = ~0;
632 	}
633 
634 	return (mask & (1 << ri));
635 }
636 
637 static int
638 iap_event_westmere_ok_on_counter(uint8_t evsel, int ri)
639 {
640 	uint32_t mask;
641 
642 	switch (evsel) {
643 	/* Events valid only on counter 0. */
644 	case 0x60:
645 	case 0xB3:
646 		mask = 0x1;
647 		break;
648 
649 	/* Events valid only on counter 0, 1. */
650 	case 0x4C:
651 	case 0x4E:
652 	case 0x51:
653 	case 0x52:
654 	case 0x63:
655 		mask = 0x3;
656 		break;
657 	/* Any row index is ok. */
658 	default:
659 		mask = ~0;
660 	}
661 
662 	return (mask & (1 << ri));
663 }
664 
665 static int
666 iap_event_sb_sbx_ib_ibx_ok_on_counter(uint8_t evsel, int ri)
667 {
668 	uint32_t mask;
669 
670 	switch (evsel) {
671 	/* Events valid only on counter 0. */
672 	case 0xB7:
673 		mask = 0x1;
674 		break;
675 	/* Events valid only on counter 1. */
676 	case 0xC0:
677 		mask = 0x2;
678 		break;
679 	/* Events valid only on counter 2. */
680 	case 0x48:
681 	case 0xA2:
682 	case 0xA3:
683 		mask = 0x4;
684 		break;
685 	/* Events valid only on counter 3. */
686 	case 0xBB:
687 	case 0xCD:
688 		mask = 0x8;
689 		break;
690 	/* Any row index is ok. */
691 	default:
692 		mask = ~0;
693 	}
694 
695 	return (mask & (1 << ri));
696 }
697 
698 static int
699 iap_event_core_ok_on_counter(uint8_t evsel, int ri)
700 {
701 	uint32_t mask;
702 
703 	switch (evsel) {
704 		/*
705 		 * Events valid only on counter 0.
706 		 */
707 	case 0x10:
708 	case 0x14:
709 	case 0x18:
710 	case 0xB3:
711 	case 0xC1:
712 	case 0xCB:
713 		mask = (1 << 0);
714 		break;
715 
716 		/*
717 		 * Events valid only on counter 1.
718 		 */
719 	case 0x11:
720 	case 0x12:
721 	case 0x13:
722 		mask = (1 << 1);
723 		break;
724 
725 	default:
726 		mask = ~0;	/* Any row index is ok. */
727 	}
728 
729 	return (mask & (1 << ri));
730 }
731 
732 static int
733 iap_allocate_pmc(int cpu, int ri, struct pmc *pm,
734     const struct pmc_op_pmcallocate *a)
735 {
736 	uint8_t ev;
737 	const struct pmc_md_iap_op_pmcallocate *iap;
738 
739 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
740 	    ("[core,%d] illegal CPU %d", __LINE__, cpu));
741 	KASSERT(ri >= 0 && ri < core_iap_npmc,
742 	    ("[core,%d] illegal row-index value %d", __LINE__, ri));
743 
744 	if (a->pm_class != PMC_CLASS_IAP)
745 		return (EINVAL);
746 
747 	iap = &a->pm_md.pm_iap;
748 	ev = IAP_EVSEL_GET(iap->pm_iap_config);
749 
750 	switch (core_cputype) {
751 	case PMC_CPU_INTEL_CORE:
752 	case PMC_CPU_INTEL_CORE2:
753 	case PMC_CPU_INTEL_CORE2EXTREME:
754 		if (iap_event_core_ok_on_counter(ev, ri) == 0)
755 			return (EINVAL);
756 	case PMC_CPU_INTEL_COREI7:
757 	case PMC_CPU_INTEL_NEHALEM_EX:
758 		if (iap_event_corei7_ok_on_counter(ev, ri) == 0)
759 			return (EINVAL);
760 		break;
761 	case PMC_CPU_INTEL_WESTMERE:
762 	case PMC_CPU_INTEL_WESTMERE_EX:
763 		if (iap_event_westmere_ok_on_counter(ev, ri) == 0)
764 			return (EINVAL);
765 		break;
766 	case PMC_CPU_INTEL_SANDYBRIDGE:
767 	case PMC_CPU_INTEL_SANDYBRIDGE_XEON:
768 	case PMC_CPU_INTEL_IVYBRIDGE:
769 	case PMC_CPU_INTEL_IVYBRIDGE_XEON:
770 	case PMC_CPU_INTEL_HASWELL:
771 	case PMC_CPU_INTEL_HASWELL_XEON:
772 	case PMC_CPU_INTEL_BROADWELL:
773 	case PMC_CPU_INTEL_BROADWELL_XEON:
774 		if (iap_event_sb_sbx_ib_ibx_ok_on_counter(ev, ri) == 0)
775 			return (EINVAL);
776 		break;
777 	case PMC_CPU_INTEL_ATOM:
778 	case PMC_CPU_INTEL_ATOM_SILVERMONT:
779 	case PMC_CPU_INTEL_ATOM_GOLDMONT:
780 	case PMC_CPU_INTEL_ATOM_GOLDMONT_P:
781 	case PMC_CPU_INTEL_ATOM_TREMONT:
782 	case PMC_CPU_INTEL_SKYLAKE:
783 	case PMC_CPU_INTEL_SKYLAKE_XEON:
784 	case PMC_CPU_INTEL_ICELAKE:
785 	case PMC_CPU_INTEL_ICELAKE_XEON:
786 	case PMC_CPU_INTEL_ALDERLAKE:
787 	default:
788 		break;
789 	}
790 
791 	pm->pm_md.pm_iap.pm_iap_evsel = iap->pm_iap_config;
792 	return (0);
793 }
794 
795 static int
796 iap_config_pmc(int cpu, int ri, struct pmc *pm)
797 {
798 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
799 	    ("[core,%d] illegal CPU %d", __LINE__, cpu));
800 
801 	KASSERT(ri >= 0 && ri < core_iap_npmc,
802 	    ("[core,%d] illegal row-index %d", __LINE__, ri));
803 
804 	PMCDBG3(MDP,CFG,1, "iap-config cpu=%d ri=%d pm=%p", cpu, ri, pm);
805 
806 	KASSERT(core_pcpu[cpu] != NULL, ("[core,%d] null per-cpu %d", __LINE__,
807 	    cpu));
808 
809 	core_pcpu[cpu]->pc_corepmcs[ri].phw_pmc = pm;
810 
811 	return (0);
812 }
813 
814 static int
815 iap_describe(int cpu, int ri, struct pmc_info *pi, struct pmc **ppmc)
816 {
817 	int error;
818 	struct pmc_hw *phw;
819 	char iap_name[PMC_NAME_MAX];
820 
821 	phw = &core_pcpu[cpu]->pc_corepmcs[ri];
822 
823 	(void) snprintf(iap_name, sizeof(iap_name), "IAP-%d", ri);
824 	if ((error = copystr(iap_name, pi->pm_name, PMC_NAME_MAX,
825 	    NULL)) != 0)
826 		return (error);
827 
828 	pi->pm_class = PMC_CLASS_IAP;
829 
830 	if (phw->phw_state & PMC_PHW_FLAG_IS_ENABLED) {
831 		pi->pm_enabled = TRUE;
832 		*ppmc          = phw->phw_pmc;
833 	} else {
834 		pi->pm_enabled = FALSE;
835 		*ppmc          = NULL;
836 	}
837 
838 	return (0);
839 }
840 
841 static int
842 iap_get_config(int cpu, int ri, struct pmc **ppm)
843 {
844 	*ppm = core_pcpu[cpu]->pc_corepmcs[ri].phw_pmc;
845 
846 	return (0);
847 }
848 
849 static int
850 iap_get_msr(int ri, uint32_t *msr)
851 {
852 	KASSERT(ri >= 0 && ri < core_iap_npmc,
853 	    ("[iap,%d] ri %d out of range", __LINE__, ri));
854 
855 	*msr = ri;
856 
857 	return (0);
858 }
859 
860 static int
861 iap_read_pmc(int cpu, int ri, pmc_value_t *v)
862 {
863 	struct pmc *pm;
864 	pmc_value_t tmp;
865 
866 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
867 	    ("[core,%d] illegal cpu value %d", __LINE__, cpu));
868 	KASSERT(ri >= 0 && ri < core_iap_npmc,
869 	    ("[core,%d] illegal row-index %d", __LINE__, ri));
870 
871 	pm = core_pcpu[cpu]->pc_corepmcs[ri].phw_pmc;
872 
873 	KASSERT(pm,
874 	    ("[core,%d] cpu %d ri %d pmc not configured", __LINE__, cpu,
875 		ri));
876 
877 	tmp = rdpmc(ri);
878 	if (PMC_IS_SAMPLING_MODE(PMC_TO_MODE(pm)))
879 		*v = iap_perfctr_value_to_reload_count(tmp);
880 	else
881 		*v = tmp & ((1ULL << core_iap_width) - 1);
882 
883 	PMCDBG4(MDP,REA,1, "iap-read cpu=%d ri=%d msr=0x%x -> v=%jx", cpu, ri,
884 	    IAP_PMC0 + ri, *v);
885 
886 	return (0);
887 }
888 
889 static int
890 iap_release_pmc(int cpu, int ri, struct pmc *pm)
891 {
892 	(void) pm;
893 
894 	PMCDBG3(MDP,REL,1, "iap-release cpu=%d ri=%d pm=%p", cpu, ri,
895 	    pm);
896 
897 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
898 	    ("[core,%d] illegal CPU value %d", __LINE__, cpu));
899 	KASSERT(ri >= 0 && ri < core_iap_npmc,
900 	    ("[core,%d] illegal row-index %d", __LINE__, ri));
901 
902 	KASSERT(core_pcpu[cpu]->pc_corepmcs[ri].phw_pmc
903 	    == NULL, ("[core,%d] PHW pmc non-NULL", __LINE__));
904 
905 	return (0);
906 }
907 
908 static int
909 iap_start_pmc(int cpu, int ri)
910 {
911 	struct pmc *pm;
912 	uint64_t evsel;
913 	struct core_cpu *cc;
914 
915 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
916 	    ("[core,%d] illegal CPU value %d", __LINE__, cpu));
917 	KASSERT(ri >= 0 && ri < core_iap_npmc,
918 	    ("[core,%d] illegal row-index %d", __LINE__, ri));
919 
920 	cc = core_pcpu[cpu];
921 	pm = cc->pc_corepmcs[ri].phw_pmc;
922 
923 	KASSERT(pm,
924 	    ("[core,%d] starting cpu%d,ri%d with no pmc configured",
925 		__LINE__, cpu, ri));
926 
927 	PMCDBG2(MDP,STA,1, "iap-start cpu=%d ri=%d", cpu, ri);
928 
929 	evsel = pm->pm_md.pm_iap.pm_iap_evsel;
930 
931 	PMCDBG4(MDP,STA,2, "iap-start/2 cpu=%d ri=%d evselmsr=0x%x evsel=0x%x",
932 	    cpu, ri, IAP_EVSEL0 + ri, evsel);
933 
934 	/* Event specific configuration. */
935 
936 	switch (IAP_EVSEL_GET(evsel)) {
937 	case 0xB7:
938 		wrmsr(IA_OFFCORE_RSP0, pm->pm_md.pm_iap.pm_iap_rsp);
939 		break;
940 	case 0xBB:
941 		wrmsr(IA_OFFCORE_RSP1, pm->pm_md.pm_iap.pm_iap_rsp);
942 		break;
943 	default:
944 		break;
945 	}
946 
947 	wrmsr(IAP_EVSEL0 + ri, evsel | IAP_EN);
948 
949 	if (core_version >= 2) {
950 		cc->pc_globalctrl |= (1ULL << ri);
951 		wrmsr(IA_GLOBAL_CTRL, cc->pc_globalctrl);
952 	}
953 
954 	return (0);
955 }
956 
957 static int
958 iap_stop_pmc(int cpu, int ri)
959 {
960 	struct pmc *pm __diagused;
961 	struct core_cpu *cc;
962 
963 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
964 	    ("[core,%d] illegal cpu value %d", __LINE__, cpu));
965 	KASSERT(ri >= 0 && ri < core_iap_npmc,
966 	    ("[core,%d] illegal row index %d", __LINE__, ri));
967 
968 	cc = core_pcpu[cpu];
969 	pm = cc->pc_corepmcs[ri].phw_pmc;
970 
971 	KASSERT(pm,
972 	    ("[core,%d] cpu%d ri%d no configured PMC to stop", __LINE__,
973 		cpu, ri));
974 
975 	PMCDBG2(MDP,STO,1, "iap-stop cpu=%d ri=%d", cpu, ri);
976 
977 	wrmsr(IAP_EVSEL0 + ri, 0);
978 
979 	/* Don't need to write IA_GLOBAL_CTRL, one disable is enough. */
980 
981 	return (0);
982 }
983 
984 static int
985 iap_write_pmc(int cpu, int ri, pmc_value_t v)
986 {
987 	struct pmc *pm;
988 	struct core_cpu *cc;
989 
990 	KASSERT(cpu >= 0 && cpu < pmc_cpu_max(),
991 	    ("[core,%d] illegal cpu value %d", __LINE__, cpu));
992 	KASSERT(ri >= 0 && ri < core_iap_npmc,
993 	    ("[core,%d] illegal row index %d", __LINE__, ri));
994 
995 	cc = core_pcpu[cpu];
996 	pm = cc->pc_corepmcs[ri].phw_pmc;
997 
998 	KASSERT(pm,
999 	    ("[core,%d] cpu%d ri%d no configured PMC to stop", __LINE__,
1000 		cpu, ri));
1001 
1002 	if (PMC_IS_SAMPLING_MODE(PMC_TO_MODE(pm)))
1003 		v = iap_reload_count_to_perfctr_value(v);
1004 
1005 	v &= (1ULL << core_iap_width) - 1;
1006 
1007 	PMCDBG4(MDP,WRI,1, "iap-write cpu=%d ri=%d msr=0x%x v=%jx", cpu, ri,
1008 	    IAP_PMC0 + ri, v);
1009 
1010 	/*
1011 	 * Write the new value to the counter (or it's alias).  The
1012 	 * counter will be in a stopped state when the pcd_write()
1013 	 * entry point is called.
1014 	 */
1015 	wrmsr(core_iap_wroffset + IAP_PMC0 + ri, v);
1016 	return (0);
1017 }
1018 
1019 
1020 static void
1021 iap_initialize(struct pmc_mdep *md, int maxcpu, int npmc, int pmcwidth,
1022     int flags)
1023 {
1024 	struct pmc_classdep *pcd;
1025 
1026 	KASSERT(md != NULL, ("[iap,%d] md is NULL", __LINE__));
1027 
1028 	PMCDBG0(MDP,INI,1, "iap-initialize");
1029 
1030 	/* Remember the set of architectural events supported. */
1031 	core_architectural_events = ~flags;
1032 
1033 	pcd = &md->pmd_classdep[PMC_MDEP_CLASS_INDEX_IAP];
1034 
1035 	pcd->pcd_caps	= IAP_PMC_CAPS;
1036 	pcd->pcd_class	= PMC_CLASS_IAP;
1037 	pcd->pcd_num	= npmc;
1038 	pcd->pcd_ri	= md->pmd_npmc;
1039 	pcd->pcd_width	= pmcwidth;
1040 
1041 	pcd->pcd_allocate_pmc	= iap_allocate_pmc;
1042 	pcd->pcd_config_pmc	= iap_config_pmc;
1043 	pcd->pcd_describe	= iap_describe;
1044 	pcd->pcd_get_config	= iap_get_config;
1045 	pcd->pcd_get_msr	= iap_get_msr;
1046 	pcd->pcd_pcpu_fini	= core_pcpu_fini;
1047 	pcd->pcd_pcpu_init	= core_pcpu_init;
1048 	pcd->pcd_read_pmc	= iap_read_pmc;
1049 	pcd->pcd_release_pmc	= iap_release_pmc;
1050 	pcd->pcd_start_pmc	= iap_start_pmc;
1051 	pcd->pcd_stop_pmc	= iap_stop_pmc;
1052 	pcd->pcd_write_pmc	= iap_write_pmc;
1053 
1054 	md->pmd_npmc	       += npmc;
1055 }
1056 
1057 static int
1058 core_intr(struct trapframe *tf)
1059 {
1060 	pmc_value_t v;
1061 	struct pmc *pm;
1062 	struct core_cpu *cc;
1063 	int error, found_interrupt, ri;
1064 
1065 	PMCDBG3(MDP,INT, 1, "cpu=%d tf=%p um=%d", curcpu, (void *) tf,
1066 	    TRAPF_USERMODE(tf));
1067 
1068 	found_interrupt = 0;
1069 	cc = core_pcpu[curcpu];
1070 
1071 	for (ri = 0; ri < core_iap_npmc; ri++) {
1072 
1073 		if ((pm = cc->pc_corepmcs[ri].phw_pmc) == NULL ||
1074 		    !PMC_IS_SAMPLING_MODE(PMC_TO_MODE(pm)))
1075 			continue;
1076 
1077 		if (!iap_pmc_has_overflowed(ri))
1078 			continue;
1079 
1080 		found_interrupt = 1;
1081 
1082 		if (pm->pm_state != PMC_STATE_RUNNING)
1083 			continue;
1084 
1085 		error = pmc_process_interrupt(PMC_HR, pm, tf);
1086 
1087 		v = pm->pm_sc.pm_reloadcount;
1088 		v = iap_reload_count_to_perfctr_value(v);
1089 
1090 		/*
1091 		 * Stop the counter, reload it but only restart it if
1092 		 * the PMC is not stalled.
1093 		 */
1094 		wrmsr(IAP_EVSEL0 + ri, pm->pm_md.pm_iap.pm_iap_evsel);
1095 		wrmsr(core_iap_wroffset + IAP_PMC0 + ri, v);
1096 
1097 		if (__predict_false(error))
1098 			continue;
1099 
1100 		wrmsr(IAP_EVSEL0 + ri, pm->pm_md.pm_iap.pm_iap_evsel | IAP_EN);
1101 	}
1102 
1103 	if (found_interrupt)
1104 		counter_u64_add(pmc_stats.pm_intr_processed, 1);
1105 	else
1106 		counter_u64_add(pmc_stats.pm_intr_ignored, 1);
1107 
1108 	if (found_interrupt)
1109 		lapic_reenable_pmc();
1110 
1111 	return (found_interrupt);
1112 }
1113 
1114 static int
1115 core2_intr(struct trapframe *tf)
1116 {
1117 	int error, found_interrupt = 0, n, cpu;
1118 	uint64_t flag, intrstatus, intrdisable = 0;
1119 	struct pmc *pm;
1120 	struct core_cpu *cc;
1121 	pmc_value_t v;
1122 
1123 	cpu = curcpu;
1124 	PMCDBG3(MDP,INT, 1, "cpu=%d tf=0x%p um=%d", cpu, (void *) tf,
1125 	    TRAPF_USERMODE(tf));
1126 
1127 	/*
1128 	 * The IA_GLOBAL_STATUS (MSR 0x38E) register indicates which
1129 	 * PMCs have a pending PMI interrupt.  We take a 'snapshot' of
1130 	 * the current set of interrupting PMCs and process these
1131 	 * after stopping them.
1132 	 */
1133 	intrstatus = rdmsr(IA_GLOBAL_STATUS);
1134 	PMCDBG2(MDP,INT, 1, "cpu=%d intrstatus=%jx", cpu,
1135 	    (uintmax_t) intrstatus);
1136 
1137 	/*
1138 	 * Stop PMCs unless hardware already done it.
1139 	 */
1140 	if ((intrstatus & IA_GLOBAL_STATUS_FLAG_CTR_FRZ) == 0)
1141 		wrmsr(IA_GLOBAL_CTRL, 0);
1142 
1143 	cc = core_pcpu[cpu];
1144 	KASSERT(cc != NULL, ("[core,%d] null pcpu", __LINE__));
1145 
1146 	/*
1147 	 * Look for interrupts from fixed function PMCs.
1148 	 */
1149 	for (n = 0, flag = (1ULL << IAF_OFFSET); n < core_iaf_npmc;
1150 	     n++, flag <<= 1) {
1151 
1152 		if ((intrstatus & flag) == 0)
1153 			continue;
1154 
1155 		found_interrupt = 1;
1156 
1157 		pm = cc->pc_corepmcs[n + core_iaf_ri].phw_pmc;
1158 		if (pm == NULL || pm->pm_state != PMC_STATE_RUNNING ||
1159 		    !PMC_IS_SAMPLING_MODE(PMC_TO_MODE(pm)))
1160 			continue;
1161 
1162 		error = pmc_process_interrupt(PMC_HR, pm, tf);
1163 		if (__predict_false(error))
1164 			intrdisable |= flag;
1165 
1166 		v = iaf_reload_count_to_perfctr_value(pm->pm_sc.pm_reloadcount);
1167 
1168 		/* Reload sampling count. */
1169 		wrmsr(IAF_CTR0 + n, v);
1170 
1171 		PMCDBG4(MDP,INT, 1, "iaf-intr cpu=%d error=%d v=%jx(%jx)", curcpu,
1172 		    error, (uintmax_t) v, (uintmax_t) rdpmc(IAF_RI_TO_MSR(n)));
1173 	}
1174 
1175 	/*
1176 	 * Process interrupts from the programmable counters.
1177 	 */
1178 	for (n = 0, flag = 1; n < core_iap_npmc; n++, flag <<= 1) {
1179 		if ((intrstatus & flag) == 0)
1180 			continue;
1181 
1182 		found_interrupt = 1;
1183 
1184 		pm = cc->pc_corepmcs[n].phw_pmc;
1185 		if (pm == NULL || pm->pm_state != PMC_STATE_RUNNING ||
1186 		    !PMC_IS_SAMPLING_MODE(PMC_TO_MODE(pm)))
1187 			continue;
1188 
1189 		error = pmc_process_interrupt(PMC_HR, pm, tf);
1190 		if (__predict_false(error))
1191 			intrdisable |= flag;
1192 
1193 		v = iap_reload_count_to_perfctr_value(pm->pm_sc.pm_reloadcount);
1194 
1195 		PMCDBG3(MDP,INT, 1, "iap-intr cpu=%d error=%d v=%jx", cpu, error,
1196 		    (uintmax_t) v);
1197 
1198 		/* Reload sampling count. */
1199 		wrmsr(core_iap_wroffset + IAP_PMC0 + n, v);
1200 	}
1201 
1202 	if (found_interrupt)
1203 		counter_u64_add(pmc_stats.pm_intr_processed, 1);
1204 	else
1205 		counter_u64_add(pmc_stats.pm_intr_ignored, 1);
1206 
1207 	if (found_interrupt)
1208 		lapic_reenable_pmc();
1209 
1210 	/*
1211 	 * Reenable all non-stalled PMCs.
1212 	 */
1213 	if ((intrstatus & IA_GLOBAL_STATUS_FLAG_CTR_FRZ) == 0) {
1214 		wrmsr(IA_GLOBAL_OVF_CTRL, intrstatus);
1215 		cc->pc_globalctrl &= ~intrdisable;
1216 		wrmsr(IA_GLOBAL_CTRL, cc->pc_globalctrl);
1217 	} else {
1218 		if (__predict_false(intrdisable)) {
1219 			cc->pc_globalctrl &= ~intrdisable;
1220 			wrmsr(IA_GLOBAL_CTRL, cc->pc_globalctrl);
1221 		}
1222 		wrmsr(IA_GLOBAL_OVF_CTRL, intrstatus);
1223 	}
1224 
1225 	PMCDBG4(MDP, INT, 1, "cpu=%d fixedctrl=%jx globalctrl=%jx status=%jx",
1226 	    cpu, (uintmax_t) rdmsr(IAF_CTRL),
1227 	    (uintmax_t) rdmsr(IA_GLOBAL_CTRL),
1228 	    (uintmax_t) rdmsr(IA_GLOBAL_STATUS));
1229 
1230 	return (found_interrupt);
1231 }
1232 
1233 int
1234 pmc_core_initialize(struct pmc_mdep *md, int maxcpu, int version_override)
1235 {
1236 	int cpuid[CORE_CPUID_REQUEST_SIZE];
1237 	int flags, nflags;
1238 
1239 	do_cpuid(CORE_CPUID_REQUEST, cpuid);
1240 
1241 	core_cputype = md->pmd_cputype;
1242 	core_version = (version_override > 0) ? version_override :
1243 	    cpuid[CORE_CPUID_EAX] & 0xFF;
1244 
1245 	PMCDBG3(MDP,INI,1,"core-init cputype=%d ncpu=%d version=%d",
1246 	    core_cputype, maxcpu, core_version);
1247 
1248 	if (core_version < 1 || core_version > 5 ||
1249 	    (core_cputype != PMC_CPU_INTEL_CORE && core_version == 1)) {
1250 		/* Unknown PMC architecture. */
1251 		printf("hwpc_core: unknown PMC architecture: %d\n",
1252 		    core_version);
1253 		return (EPROGMISMATCH);
1254 	}
1255 
1256 	core_iap_wroffset = 0;
1257 	if (cpu_feature2 & CPUID2_PDCM) {
1258 		if (rdmsr(IA32_PERF_CAPABILITIES) & PERFCAP_FW_WRITE) {
1259 			PMCDBG0(MDP, INI, 1,
1260 			    "core-init full-width write supported");
1261 			core_iap_wroffset = IAP_A_PMC0 - IAP_PMC0;
1262 		} else
1263 			PMCDBG0(MDP, INI, 1,
1264 			    "core-init full-width write NOT supported");
1265 	} else
1266 		PMCDBG0(MDP, INI, 1, "core-init pdcm not supported");
1267 
1268 	core_pmcmask = 0;
1269 
1270 	/*
1271 	 * Initialize programmable counters.
1272 	 */
1273 	core_iap_npmc = (cpuid[CORE_CPUID_EAX] >> 8) & 0xFF;
1274 	core_iap_width = (cpuid[CORE_CPUID_EAX] >> 16) & 0xFF;
1275 
1276 	core_pmcmask |= ((1ULL << core_iap_npmc) - 1);
1277 
1278 	nflags = (cpuid[CORE_CPUID_EAX] >> 24) & 0xFF;
1279 	flags = cpuid[CORE_CPUID_EBX] & ((1 << nflags) - 1);
1280 
1281 	iap_initialize(md, maxcpu, core_iap_npmc, core_iap_width, flags);
1282 
1283 	/*
1284 	 * Initialize fixed function counters, if present.
1285 	 */
1286 	if (core_version >= 2) {
1287 		core_iaf_ri = core_iap_npmc;
1288 		core_iaf_npmc = cpuid[CORE_CPUID_EDX] & 0x1F;
1289 		core_iaf_width = (cpuid[CORE_CPUID_EDX] >> 5) & 0xFF;
1290 
1291 		iaf_initialize(md, maxcpu, core_iaf_npmc, core_iaf_width);
1292 		core_pmcmask |= ((1ULL << core_iaf_npmc) - 1) << IAF_OFFSET;
1293 	}
1294 
1295 	PMCDBG2(MDP,INI,1,"core-init pmcmask=0x%jx iafri=%d", core_pmcmask,
1296 	    core_iaf_ri);
1297 
1298 	core_pcpu = malloc(sizeof(*core_pcpu) * maxcpu, M_PMC,
1299 	    M_ZERO | M_WAITOK);
1300 
1301 	/*
1302 	 * Choose the appropriate interrupt handler.
1303 	 */
1304 	if (core_version >= 2)
1305 		md->pmd_intr = core2_intr;
1306 	else
1307 		md->pmd_intr = core_intr;
1308 
1309 	md->pmd_pcpu_fini = NULL;
1310 	md->pmd_pcpu_init = NULL;
1311 
1312 	return (0);
1313 }
1314 
1315 void
1316 pmc_core_finalize(struct pmc_mdep *md)
1317 {
1318 	PMCDBG0(MDP,INI,1, "core-finalize");
1319 
1320 	free(core_pcpu, M_PMC);
1321 	core_pcpu = NULL;
1322 }
1323