xref: /freebsd/sys/dev/xen/bus/xen_intr.c (revision f552d7ad)
1 /******************************************************************************
2  * xen_intr.c
3  *
4  * Xen event and interrupt services for x86 HVM guests.
5  *
6  * Copyright (c) 2002-2005, K A Fraser
7  * Copyright (c) 2005, Intel Corporation <xiaofeng.ling@intel.com>
8  * Copyright (c) 2012, Spectra Logic Corporation
9  * Copyright © 2021-2023, Elliott Mitchell
10  *
11  * This file may be distributed separately from the Linux kernel, or
12  * incorporated into other software packages, subject to the following license:
13  *
14  * Permission is hereby granted, free of charge, to any person obtaining a copy
15  * of this source file (the "Software"), to deal in the Software without
16  * restriction, including without limitation the rights to use, copy, modify,
17  * merge, publish, distribute, sublicense, and/or sell copies of the Software,
18  * and to permit persons to whom the Software is furnished to do so, subject to
19  * the following conditions:
20  *
21  * The above copyright notice and this permission notice shall be included in
22  * all copies or substantial portions of the Software.
23  *
24  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
25  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
26  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
27  * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
28  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
29  * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
30  * IN THE SOFTWARE.
31  */
32 
33 #include <sys/cdefs.h>
34 #include "opt_ddb.h"
35 
36 #include <sys/param.h>
37 #include <sys/systm.h>
38 #include <sys/bus.h>
39 #include <sys/kernel.h>
40 #include <sys/limits.h>
41 #include <sys/lock.h>
42 #include <sys/mutex.h>
43 #include <sys/interrupt.h>
44 #include <sys/pcpu.h>
45 #include <sys/proc.h>
46 #include <sys/smp.h>
47 #include <sys/refcount.h>
48 
49 #include <vm/vm.h>
50 #include <vm/pmap.h>
51 
52 #include <machine/smp.h>
53 #include <machine/stdarg.h>
54 
55 #include <xen/xen-os.h>
56 #include <xen/hypervisor.h>
57 #include <xen/xen_intr.h>
58 #include <xen/evtchn/evtchnvar.h>
59 
60 #include <machine/xen/arch-intr.h>
61 
62 #ifdef DDB
63 #include <ddb/ddb.h>
64 #endif
65 
66 /**
67  * Per-cpu event channel processing state.
68  */
69 struct xen_intr_pcpu_data {
70 	/**
71 	 * The last event channel bitmap section (level one bit) processed.
72 	 * This is used to ensure we scan all ports before
73 	 * servicing an already servied port again.
74 	 */
75 	u_int	last_processed_l1i;
76 
77 	/**
78 	 * The last event channel processed within the event channel
79 	 * bitmap being scanned.
80 	 */
81 	u_int	last_processed_l2i;
82 
83 	/**
84 	 * A bitmap of ports that can be serviced from this CPU.
85 	 * A set bit means interrupt handling is enabled.
86 	 */
87 	xen_ulong_t	evtchn_enabled[sizeof(xen_ulong_t) * 8];
88 };
89 
90 /*
91  * Start the scan at port 0 by initializing the last scanned
92  * location as the highest numbered event channel port.
93  */
94 DPCPU_DEFINE_STATIC(struct xen_intr_pcpu_data, xen_intr_pcpu) = {
95 	.last_processed_l1i = LONG_BIT - 1,
96 	.last_processed_l2i = LONG_BIT - 1
97 };
98 
99 DPCPU_DECLARE(struct vcpu_info *, vcpu_info);
100 
101 #define	INVALID_EVTCHN		(~(evtchn_port_t)0) /* Invalid event channel */
102 #define	is_valid_evtchn(x)	((uintmax_t)(x) < NR_EVENT_CHANNELS)
103 
104 /*
105  * Lock for interrupt core data.
106  *
107  * Modifying xen_intr_port_to_isrc[], or isrc->xi_port (implies the former)
108  * requires this lock be held.  Any time this lock is not held, the condition
109  * `!xen_intr_port_to_isrc[i] || (xen_intr_port_to_isrc[i]->ix_port == i)`
110  * MUST be true for all values of i which are valid indicies of the array.
111  *
112  * Acquire/release operations for isrc->xi_refcount require this lock be held.
113  */
114 static struct mtx	 xen_intr_isrc_lock;
115 static struct xenisrc	*xen_intr_port_to_isrc[NR_EVENT_CHANNELS];
116 
117 /*------------------------- Private Functions --------------------------------*/
118 
119 /**
120  * Retrieve a handle for a Xen interrupt source.
121  *
122  * \param isrc  A valid Xen interrupt source structure.
123  *
124  * \returns  A handle suitable for use with xen_intr_isrc_from_handle()
125  *           to retrieve the original Xen interrupt source structure.
126  */
127 
128 static inline xen_intr_handle_t
129 xen_intr_handle_from_isrc(struct xenisrc *isrc)
130 {
131 	return (isrc);
132 }
133 
134 /**
135  * Lookup a Xen interrupt source object given an interrupt binding handle.
136  *
137  * \param handle  A handle initialized by a previous call to
138  *                xen_intr_bind_isrc().
139  *
140  * \returns  A pointer to the Xen interrupt source object associated
141  *           with the given interrupt handle.  NULL if no association
142  *           currently exists.
143  */
144 static inline struct xenisrc *
145 xen_intr_isrc_from_handle(xen_intr_handle_t handle)
146 {
147 	return ((struct xenisrc *)handle);
148 }
149 
150 /**
151  * Disable signal delivery for an event channel port on the
152  * specified CPU.
153  *
154  * \param port  The event channel port to mask.
155  *
156  * This API is used to manage the port<=>CPU binding of event
157  * channel handlers.
158  *
159  * \note  This operation does not preclude reception of an event
160  *        for this event channel on another CPU.  To mask the
161  *        event channel globally, use evtchn_mask().
162  */
163 static inline void
164 evtchn_cpu_mask_port(u_int cpu, evtchn_port_t port)
165 {
166 	struct xen_intr_pcpu_data *pcpu;
167 
168 	pcpu = DPCPU_ID_PTR(cpu, xen_intr_pcpu);
169 	xen_clear_bit(port, pcpu->evtchn_enabled);
170 }
171 
172 /**
173  * Enable signal delivery for an event channel port on the
174  * specified CPU.
175  *
176  * \param port  The event channel port to unmask.
177  *
178  * This API is used to manage the port<=>CPU binding of event
179  * channel handlers.
180  *
181  * \note  This operation does not guarantee that event delivery
182  *        is enabled for this event channel port.  The port must
183  *        also be globally enabled.  See evtchn_unmask().
184  */
185 static inline void
186 evtchn_cpu_unmask_port(u_int cpu, evtchn_port_t port)
187 {
188 	struct xen_intr_pcpu_data *pcpu;
189 
190 	pcpu = DPCPU_ID_PTR(cpu, xen_intr_pcpu);
191 	xen_set_bit(port, pcpu->evtchn_enabled);
192 }
193 
194 /**
195  * Attempt to free an active Xen interrupt source object.
196  *
197  * \param isrc  The interrupt source object to release.
198  *
199  * \returns  EBUSY if the source is still in use, otherwise 0.
200  */
201 static int
202 xen_intr_release_isrc(struct xenisrc *isrc)
203 {
204 
205 	mtx_lock(&xen_intr_isrc_lock);
206 	if (is_valid_evtchn(isrc->xi_port)) {
207 		evtchn_mask_port(isrc->xi_port);
208 		evtchn_clear_port(isrc->xi_port);
209 
210 		/* Rebind port to CPU 0. */
211 		evtchn_cpu_mask_port(isrc->xi_cpu, isrc->xi_port);
212 		evtchn_cpu_unmask_port(0, isrc->xi_port);
213 
214 		if (isrc->xi_close != 0) {
215 			struct evtchn_close close = { .port = isrc->xi_port };
216 
217 			if (HYPERVISOR_event_channel_op(EVTCHNOP_close, &close))
218 				panic("EVTCHNOP_close failed");
219 		}
220 
221 		xen_intr_port_to_isrc[isrc->xi_port] = NULL;
222 	}
223 	/* not reachable from xen_intr_port_to_isrc[], unlock */
224 	mtx_unlock(&xen_intr_isrc_lock);
225 
226 	xen_arch_intr_release(isrc);
227 	return (0);
228 }
229 
230 /**
231  * Associate an interrupt handler with an already allocated local Xen
232  * event channel port.
233  *
234  * \param isrcp       The returned Xen interrupt object associated with
235  *                    the specified local port.
236  * \param local_port  The event channel to bind.
237  * \param type        The event channel type of local_port.
238  * \param intr_owner  The device making this bind request.
239  * \param filter      An interrupt filter handler.  Specify NULL
240  *                    to always dispatch to the ithread handler.
241  * \param handler     An interrupt ithread handler.  Optional (can
242  *                    specify NULL) if all necessary event actions
243  *                    are performed by filter.
244  * \param arg         Argument to present to both filter and handler.
245  * \param irqflags    Interrupt handler flags.  See sys/bus.h.
246  * \param handlep     Pointer to an opaque handle used to manage this
247  *                    registration.
248  *
249  * \returns  0 on success, otherwise an errno.
250  */
251 static int
252 xen_intr_bind_isrc(struct xenisrc **isrcp, evtchn_port_t local_port,
253     enum evtchn_type type, const char *intr_owner, driver_filter_t filter,
254     driver_intr_t handler, void *arg, enum intr_type flags,
255     xen_intr_handle_t *const port_handlep)
256 {
257 	struct xenisrc *isrc;
258 	int error;
259 
260 	*isrcp = NULL;
261 	if (port_handlep == NULL) {
262 		printf("%s: %s: Bad event handle\n", intr_owner, __func__);
263 		return (EINVAL);
264 	}
265 	*port_handlep = NULL;
266 
267 	isrc = xen_arch_intr_alloc();
268 	if (isrc == NULL)
269 		return (ENOSPC);
270 
271 	isrc->xi_cookie = NULL;
272 	isrc->xi_type = type;
273 	isrc->xi_port = local_port;
274 	isrc->xi_close = false;
275 	isrc->xi_cpu = 0;
276 	refcount_init(&isrc->xi_refcount, 1);
277 	mtx_lock(&xen_intr_isrc_lock);
278 	xen_intr_port_to_isrc[isrc->xi_port] = isrc;
279 	mtx_unlock(&xen_intr_isrc_lock);
280 
281 #ifdef SMP
282 	if (type == EVTCHN_TYPE_PORT) {
283 		/*
284 		 * By default all interrupts are assigned to vCPU#0
285 		 * unless specified otherwise, so shuffle them to balance
286 		 * the interrupt load.
287 		 */
288 		xen_intr_assign_cpu(isrc, xen_arch_intr_next_cpu(isrc));
289 	}
290 #endif
291 
292 	/*
293 	 * If a filter or handler function is provided, add it to the event.
294 	 * Otherwise the event channel is left masked and without a handler,
295 	 * the caller is in charge of setting that up.
296 	 */
297 	if (filter != NULL || handler != NULL) {
298 		error = xen_intr_add_handler(intr_owner, filter, handler, arg,
299 		    flags, xen_intr_handle_from_isrc(isrc));
300 		if (error != 0) {
301 			xen_intr_release_isrc(isrc);
302 			return (error);
303 		}
304 	}
305 
306 	*isrcp = isrc;
307 	/* Assign the opaque handler */
308 	*port_handlep = xen_intr_handle_from_isrc(isrc);
309 	return (0);
310 }
311 
312 /**
313  * Determine the event channel ports at the given section of the
314  * event port bitmap which have pending events for the given cpu.
315  *
316  * \param pcpu  The Xen interrupt pcpu data for the cpu being queried.
317  * \param sh    The Xen shared info area.
318  * \param idx   The index of the section of the event channel bitmap to
319  *              inspect.
320  *
321  * \returns  A u_long with bits set for every event channel with pending
322  *           events.
323  */
324 static inline u_long
325 xen_intr_active_ports(const struct xen_intr_pcpu_data *const pcpu,
326     const u_int idx)
327 {
328 	volatile const shared_info_t *const sh = HYPERVISOR_shared_info;
329 
330 	CTASSERT(sizeof(sh->evtchn_mask[0]) == sizeof(sh->evtchn_pending[0]));
331 	CTASSERT(sizeof(sh->evtchn_mask[0]) == sizeof(pcpu->evtchn_enabled[0]));
332 	CTASSERT(sizeof(sh->evtchn_mask) == sizeof(sh->evtchn_pending));
333 	CTASSERT(sizeof(sh->evtchn_mask) == sizeof(pcpu->evtchn_enabled));
334 	return (sh->evtchn_pending[idx]
335 	      & ~sh->evtchn_mask[idx]
336 	      & pcpu->evtchn_enabled[idx]);
337 }
338 
339 /**
340  * Interrupt handler for processing all Xen event channel events.
341  *
342  * \param trap_frame  The trap frame context for the current interrupt.
343  */
344 int
345 xen_intr_handle_upcall(void *unused __unused)
346 {
347 	struct trapframe *trap_frame = curthread->td_intr_frame;
348 	u_int l1i, l2i, port, cpu __diagused;
349 	u_long masked_l1, masked_l2;
350 	struct xenisrc *isrc;
351 	vcpu_info_t *v;
352 	struct xen_intr_pcpu_data *pc;
353 	u_long l1, l2;
354 
355 	/* We must remain on the same vCPU during this function */
356 	CRITICAL_ASSERT(curthread);
357 
358 	cpu = PCPU_GET(cpuid);
359 	pc  = DPCPU_PTR(xen_intr_pcpu);
360 	v   = DPCPU_GET(vcpu_info);
361 
362 	if (!xen_has_percpu_evtchn()) {
363 		KASSERT((cpu == 0), ("Fired PCI event callback on wrong CPU"));
364 	}
365 
366 	v->evtchn_upcall_pending = 0;
367 /* No need for a barrier on x86 -- XCHG is a barrier on x86. */
368 #if !defined(__amd64__) && !defined(__i386__)
369 	/* Clear master flag /before/ clearing selector flag. */
370 	wmb();
371 #endif
372 	l1 = atomic_readandclear_xen_ulong(&v->evtchn_pending_sel);
373 
374 	l1i = pc->last_processed_l1i;
375 	l2i = pc->last_processed_l2i;
376 
377 	while (l1 != 0) {
378 		l1i = (l1i + 1) % LONG_BIT;
379 		masked_l1 = l1 & ((~0UL) << l1i);
380 
381 		if (masked_l1 == 0) {
382 			/*
383 			 * if we masked out all events, wrap around
384 			 * to the beginning.
385 			 */
386 			l1i = LONG_BIT - 1;
387 			l2i = LONG_BIT - 1;
388 			continue;
389 		}
390 		l1i = ffsl(masked_l1) - 1;
391 
392 		do {
393 			l2 = xen_intr_active_ports(pc, l1i);
394 
395 			l2i = (l2i + 1) % LONG_BIT;
396 			masked_l2 = l2 & ((~0UL) << l2i);
397 
398 			if (masked_l2 == 0) {
399 				/* if we masked out all events, move on */
400 				l2i = LONG_BIT - 1;
401 				break;
402 			}
403 			l2i = ffsl(masked_l2) - 1;
404 
405 			/* process port */
406 			port = (l1i * LONG_BIT) + l2i;
407 			evtchn_clear_port(port);
408 
409 			isrc = xen_intr_port_to_isrc[port];
410 			if (__predict_false(isrc == NULL))
411 				continue;
412 
413 			/* Make sure we are firing on the right vCPU */
414 			KASSERT((isrc->xi_cpu == PCPU_GET(cpuid)),
415 				("Received unexpected event on vCPU#%u, event bound to vCPU#%u",
416 				PCPU_GET(cpuid), isrc->xi_cpu));
417 
418 			xen_arch_intr_execute_handlers(isrc, trap_frame);
419 
420 			/*
421 			 * If this is the final port processed,
422 			 * we'll pick up here+1 next time.
423 			 */
424 			pc->last_processed_l1i = l1i;
425 			pc->last_processed_l2i = l2i;
426 
427 		} while (l2i != LONG_BIT - 1);
428 
429 		l2 = xen_intr_active_ports(pc, l1i);
430 		if (l2 == 0) {
431 			/*
432 			 * We handled all ports, so we can clear the
433 			 * selector bit.
434 			 */
435 			l1 &= ~(1UL << l1i);
436 		}
437 	}
438 
439 	return (FILTER_HANDLED);
440 }
441 
442 static int
443 xen_intr_init(void *dummy __unused)
444 {
445 	shared_info_t *s = HYPERVISOR_shared_info;
446 	struct xen_intr_pcpu_data *pcpu;
447 	int i;
448 
449 	if (!xen_domain())
450 		return (0);
451 
452 	_Static_assert(is_valid_evtchn(0),
453 	    "is_valid_evtchn(0) fails (unused by Xen, but valid by interface");
454 	_Static_assert(is_valid_evtchn(NR_EVENT_CHANNELS - 1),
455 	    "is_valid_evtchn(max) fails (is a valid channel)");
456 	_Static_assert(!is_valid_evtchn(NR_EVENT_CHANNELS),
457 	    "is_valid_evtchn(>max) fails (NOT a valid channel)");
458 	_Static_assert(!is_valid_evtchn(~(evtchn_port_t)0),
459 	    "is_valid_evtchn(maxint) fails (overflow?)");
460 	_Static_assert(!is_valid_evtchn(INVALID_EVTCHN),
461 	    "is_valid_evtchn(INVALID_EVTCHN) fails (must be invalid!)");
462 	_Static_assert(!is_valid_evtchn(-1),
463 	    "is_valid_evtchn(-1) fails (negative are invalid)");
464 
465 	mtx_init(&xen_intr_isrc_lock, "xen-irq-lock", NULL, MTX_DEF);
466 
467 	/*
468 	 * Set the per-cpu mask of CPU#0 to enable all, since by default all
469 	 * event channels are bound to CPU#0.
470 	 */
471 	CPU_FOREACH(i) {
472 		pcpu = DPCPU_ID_PTR(i, xen_intr_pcpu);
473 		memset(pcpu->evtchn_enabled, i == 0 ? ~0 : 0,
474 		    sizeof(pcpu->evtchn_enabled));
475 	}
476 
477 	for (i = 0; i < nitems(s->evtchn_mask); i++)
478 		atomic_store_rel_xen_ulong(&s->evtchn_mask[i], ~0);
479 
480 	xen_arch_intr_init();
481 
482 	if (bootverbose)
483 		printf("Xen interrupt system initialized\n");
484 
485 	return (0);
486 }
487 SYSINIT(xen_intr_init, SI_SUB_INTR, SI_ORDER_SECOND, xen_intr_init, NULL);
488 
489 /*--------------------------- Common PIC Functions ---------------------------*/
490 
491 static void
492 xen_rebind_ipi(struct xenisrc *isrc)
493 {
494 #ifdef SMP
495 	u_int cpu = isrc->xi_cpu;
496 	u_int vcpu_id = XEN_CPUID_TO_VCPUID(cpu);
497 	int error;
498 	struct evtchn_bind_ipi bind_ipi = { .vcpu = vcpu_id };
499 
500 	error = HYPERVISOR_event_channel_op(EVTCHNOP_bind_ipi,
501 	                                    &bind_ipi);
502 	if (error != 0)
503 		panic("unable to rebind xen IPI: %d", error);
504 
505 	isrc->xi_port = bind_ipi.port;
506 #else
507 	panic("Resume IPI event channel on UP");
508 #endif
509 }
510 
511 static void
512 xen_rebind_virq(struct xenisrc *isrc)
513 {
514 	u_int cpu = isrc->xi_cpu;
515 	u_int vcpu_id = XEN_CPUID_TO_VCPUID(cpu);
516 	int error;
517 	struct evtchn_bind_virq bind_virq = { .virq = isrc->xi_virq,
518 	                                      .vcpu = vcpu_id };
519 
520 	error = HYPERVISOR_event_channel_op(EVTCHNOP_bind_virq,
521 	                                    &bind_virq);
522 	if (error != 0)
523 		panic("unable to rebind xen VIRQ#%u: %d", isrc->xi_virq, error);
524 
525 	isrc->xi_port = bind_virq.port;
526 }
527 
528 static struct xenisrc *
529 xen_intr_rebind_isrc(struct xenisrc *isrc)
530 {
531 #ifdef SMP
532 	u_int cpu = isrc->xi_cpu;
533 	int error;
534 #endif
535 	struct xenisrc *prev;
536 
537 	switch (isrc->xi_type) {
538 	case EVTCHN_TYPE_IPI:
539 		xen_rebind_ipi(isrc);
540 		break;
541 	case EVTCHN_TYPE_VIRQ:
542 		xen_rebind_virq(isrc);
543 		break;
544 	default:
545 		return (NULL);
546 	}
547 
548 	prev = xen_intr_port_to_isrc[isrc->xi_port];
549 	xen_intr_port_to_isrc[isrc->xi_port] = isrc;
550 
551 #ifdef SMP
552 	isrc->xi_cpu = 0;
553 	error = xen_intr_assign_cpu(isrc, cpu);
554 	if (error)
555 		panic("%s(): unable to rebind Xen channel %u to vCPU%u: %d",
556 		    __func__, isrc->xi_port, cpu, error);
557 #endif
558 
559 	evtchn_unmask_port(isrc->xi_port);
560 
561 	return (prev);
562 }
563 
564 /**
565  * Return this PIC to service after being suspended.
566  */
567 void
568 xen_intr_resume(void)
569 {
570 	shared_info_t *s = HYPERVISOR_shared_info;
571 	u_int isrc_idx;
572 	int i;
573 
574 	/* Reset the per-CPU masks */
575 	CPU_FOREACH(i) {
576 		struct xen_intr_pcpu_data *pcpu;
577 
578 		pcpu = DPCPU_ID_PTR(i, xen_intr_pcpu);
579 		memset(pcpu->evtchn_enabled, i == 0 ? ~0 : 0,
580 		    sizeof(pcpu->evtchn_enabled));
581 	}
582 
583 	/* Mask all event channels. */
584 	for (i = 0; i < nitems(s->evtchn_mask); i++)
585 		atomic_store_rel_xen_ulong(&s->evtchn_mask[i], ~0);
586 
587 	/* Clear existing port mappings */
588 	for (isrc_idx = 0; isrc_idx < NR_EVENT_CHANNELS; ++isrc_idx)
589 		if (xen_intr_port_to_isrc[isrc_idx] != NULL)
590 			xen_intr_port_to_isrc[isrc_idx]->xi_port =
591 			    INVALID_EVTCHN;
592 
593 	/* Remap in-use isrcs, using xen_intr_port_to_isrc as listing */
594 	for (isrc_idx = 0; isrc_idx < NR_EVENT_CHANNELS; ++isrc_idx) {
595 		struct xenisrc *cur = xen_intr_port_to_isrc[isrc_idx];
596 
597 		/* empty or entry already taken care of */
598 		if (cur == NULL || cur->xi_port == isrc_idx)
599 			continue;
600 
601 		xen_intr_port_to_isrc[isrc_idx] = NULL;
602 
603 		do {
604 			KASSERT(!is_valid_evtchn(cur->xi_port),
605 			    ("%s(): Multiple channels on single intr?",
606 			    __func__));
607 
608 			cur = xen_intr_rebind_isrc(cur);
609 		} while (cur != NULL);
610 	}
611 }
612 
613 /**
614  * Disable a Xen interrupt source.
615  *
616  * \param isrc  The interrupt source to disable.
617  */
618 void
619 xen_intr_disable_intr(struct xenisrc *isrc)
620 {
621 
622 	evtchn_mask_port(isrc->xi_port);
623 }
624 
625 /**
626  * Configure CPU affinity for interrupt source event delivery.
627  *
628  * \param isrc     The interrupt source to configure.
629  * \param to_cpu   The id of the CPU for handling future events.
630  *
631  * \returns  0 if successful, otherwise an errno.
632  */
633 int
634 xen_intr_assign_cpu(struct xenisrc *isrc, u_int to_cpu)
635 {
636 #ifdef SMP
637 	struct evtchn_bind_vcpu bind_vcpu;
638 	u_int vcpu_id = XEN_CPUID_TO_VCPUID(to_cpu);
639 	int error, masked;
640 
641 	if (!xen_has_percpu_evtchn())
642 		return (EOPNOTSUPP);
643 
644 	mtx_lock(&xen_intr_isrc_lock);
645 	if (!is_valid_evtchn(isrc->xi_port)) {
646 		mtx_unlock(&xen_intr_isrc_lock);
647 		return (EINVAL);
648 	}
649 
650 	/*
651 	 * Mask the event channel while binding it to prevent interrupt
652 	 * delivery with an inconsistent state in isrc->xi_cpu.
653 	 */
654 	masked = evtchn_test_and_set_mask(isrc->xi_port);
655 	if ((isrc->xi_type == EVTCHN_TYPE_VIRQ) ||
656 		(isrc->xi_type == EVTCHN_TYPE_IPI)) {
657 		/*
658 		 * Virtual IRQs are associated with a cpu by
659 		 * the Hypervisor at evtchn_bind_virq time, so
660 		 * all we need to do is update the per-CPU masks.
661 		 */
662 		evtchn_cpu_mask_port(isrc->xi_cpu, isrc->xi_port);
663 		isrc->xi_cpu = to_cpu;
664 		evtchn_cpu_unmask_port(isrc->xi_cpu, isrc->xi_port);
665 		goto out;
666 	}
667 
668 	bind_vcpu.port = isrc->xi_port;
669 	bind_vcpu.vcpu = vcpu_id;
670 
671 	error = HYPERVISOR_event_channel_op(EVTCHNOP_bind_vcpu, &bind_vcpu);
672 	if (isrc->xi_cpu != to_cpu) {
673 		if (error == 0) {
674 			/* Commit to new binding by removing the old one. */
675 			evtchn_cpu_mask_port(isrc->xi_cpu, isrc->xi_port);
676 			isrc->xi_cpu = to_cpu;
677 			evtchn_cpu_unmask_port(isrc->xi_cpu, isrc->xi_port);
678 		}
679 	}
680 
681 out:
682 	if (masked == 0)
683 		evtchn_unmask_port(isrc->xi_port);
684 	mtx_unlock(&xen_intr_isrc_lock);
685 	return (0);
686 #else
687 	return (EOPNOTSUPP);
688 #endif
689 }
690 
691 /*------------------- Virtual Interrupt Source PIC Functions -----------------*/
692 /*
693  * Mask a level triggered interrupt source.
694  *
695  * \param isrc  The interrupt source to mask (if necessary).
696  */
697 void
698 xen_intr_disable_source(struct xenisrc *isrc)
699 {
700 
701 	/*
702 	 * NB: checking if the event channel is already masked is
703 	 * needed because the event channel user-space device
704 	 * masks event channels on its filter as part of its
705 	 * normal operation, and those shouldn't be automatically
706 	 * unmasked by the generic interrupt code. The event channel
707 	 * device will unmask them when needed.
708 	 */
709 	isrc->xi_masked = !!evtchn_test_and_set_mask(isrc->xi_port);
710 }
711 
712 /*
713  * Unmask a level triggered interrupt source.
714  *
715  * \param isrc  The interrupt source to unmask (if necessary).
716  */
717 void
718 xen_intr_enable_source(struct xenisrc *isrc)
719 {
720 
721 	if (isrc->xi_masked == 0)
722 		evtchn_unmask_port(isrc->xi_port);
723 }
724 
725 /*
726  * Enable and unmask the interrupt source.
727  *
728  * \param isrc  The interrupt source to enable.
729  */
730 void
731 xen_intr_enable_intr(struct xenisrc *isrc)
732 {
733 
734 	evtchn_unmask_port(isrc->xi_port);
735 }
736 
737 /*--------------------------- Public Functions -------------------------------*/
738 /*------- API comments for these methods can be found in xen/xenintr.h -------*/
739 int
740 xen_intr_bind_local_port(device_t dev, evtchn_port_t local_port,
741     driver_filter_t filter, driver_intr_t handler, void *arg,
742     enum intr_type flags, xen_intr_handle_t *port_handlep)
743 {
744 	struct xenisrc *isrc;
745 	int error;
746 
747 	error = xen_intr_bind_isrc(&isrc, local_port, EVTCHN_TYPE_PORT,
748 	    device_get_nameunit(dev), filter, handler, arg, flags,
749 	    port_handlep);
750 	if (error != 0)
751 		return (error);
752 
753 	/*
754 	 * The Event Channel API didn't open this port, so it is not
755 	 * responsible for closing it automatically on unbind.
756 	 */
757 	isrc->xi_close = 0;
758 	return (0);
759 }
760 
761 int
762 xen_intr_alloc_and_bind_local_port(device_t dev, u_int remote_domain,
763     driver_filter_t filter, driver_intr_t handler, void *arg,
764     enum intr_type flags, xen_intr_handle_t *port_handlep)
765 {
766 	struct xenisrc *isrc;
767 	struct evtchn_alloc_unbound alloc_unbound;
768 	int error;
769 
770 	alloc_unbound.dom        = DOMID_SELF;
771 	alloc_unbound.remote_dom = remote_domain;
772 	error = HYPERVISOR_event_channel_op(EVTCHNOP_alloc_unbound,
773 		    &alloc_unbound);
774 	if (error != 0) {
775 		/*
776 		 * XXX Trap Hypercall error code Linuxisms in
777 		 *     the HYPERCALL layer.
778 		 */
779 		return (-error);
780 	}
781 
782 	error = xen_intr_bind_isrc(&isrc, alloc_unbound.port, EVTCHN_TYPE_PORT,
783 	    device_get_nameunit(dev), filter, handler, arg, flags,
784 	    port_handlep);
785 	if (error != 0) {
786 		evtchn_close_t close = { .port = alloc_unbound.port };
787 		if (HYPERVISOR_event_channel_op(EVTCHNOP_close, &close))
788 			panic("EVTCHNOP_close failed");
789 		return (error);
790 	}
791 
792 	isrc->xi_close = 1;
793 	return (0);
794 }
795 
796 int
797 xen_intr_bind_remote_port(device_t dev, u_int remote_domain,
798     u_int remote_port, driver_filter_t filter, driver_intr_t handler,
799     void *arg, enum intr_type flags, xen_intr_handle_t *port_handlep)
800 {
801 	struct xenisrc *isrc;
802 	struct evtchn_bind_interdomain bind_interdomain;
803 	int error;
804 
805 	bind_interdomain.remote_dom  = remote_domain;
806 	bind_interdomain.remote_port = remote_port;
807 	error = HYPERVISOR_event_channel_op(EVTCHNOP_bind_interdomain,
808 					    &bind_interdomain);
809 	if (error != 0) {
810 		/*
811 		 * XXX Trap Hypercall error code Linuxisms in
812 		 *     the HYPERCALL layer.
813 		 */
814 		return (-error);
815 	}
816 
817 	error = xen_intr_bind_isrc(&isrc, bind_interdomain.local_port,
818 	    EVTCHN_TYPE_PORT, device_get_nameunit(dev), filter, handler, arg,
819 	    flags, port_handlep);
820 	if (error) {
821 		evtchn_close_t close = { .port = bind_interdomain.local_port };
822 		if (HYPERVISOR_event_channel_op(EVTCHNOP_close, &close))
823 			panic("EVTCHNOP_close failed");
824 		return (error);
825 	}
826 
827 	/*
828 	 * The Event Channel API opened this port, so it is
829 	 * responsible for closing it automatically on unbind.
830 	 */
831 	isrc->xi_close = 1;
832 	return (0);
833 }
834 
835 int
836 xen_intr_bind_virq(device_t dev, u_int virq, u_int cpu,
837     driver_filter_t filter, driver_intr_t handler, void *arg,
838     enum intr_type flags, xen_intr_handle_t *port_handlep)
839 {
840 	u_int vcpu_id = XEN_CPUID_TO_VCPUID(cpu);
841 	struct xenisrc *isrc;
842 	struct evtchn_bind_virq bind_virq = { .virq = virq, .vcpu = vcpu_id };
843 	int error;
844 
845 	isrc = NULL;
846 	error = HYPERVISOR_event_channel_op(EVTCHNOP_bind_virq, &bind_virq);
847 	if (error != 0) {
848 		/*
849 		 * XXX Trap Hypercall error code Linuxisms in
850 		 *     the HYPERCALL layer.
851 		 */
852 		return (-error);
853 	}
854 
855 	error = xen_intr_bind_isrc(&isrc, bind_virq.port, EVTCHN_TYPE_VIRQ,
856 	    device_get_nameunit(dev), filter, handler, arg, flags,
857 	    port_handlep);
858 
859 #ifdef SMP
860 	if (error == 0)
861 		error = xen_arch_intr_event_bind(isrc, cpu);
862 #endif
863 
864 	if (error != 0) {
865 		evtchn_close_t close = { .port = bind_virq.port };
866 
867 		xen_intr_unbind(port_handlep);
868 		if (HYPERVISOR_event_channel_op(EVTCHNOP_close, &close))
869 			panic("EVTCHNOP_close failed");
870 		return (error);
871 	}
872 
873 #ifdef SMP
874 	if (isrc->xi_cpu != cpu) {
875 		/*
876 		 * Too early in the boot process for the generic interrupt
877 		 * code to perform the binding.  Update our event channel
878 		 * masks manually so events can't fire on the wrong cpu
879 		 * during AP startup.
880 		 */
881 		xen_intr_assign_cpu(isrc, cpu);
882 	}
883 #endif
884 
885 	/*
886 	 * The Event Channel API opened this port, so it is
887 	 * responsible for closing it automatically on unbind.
888 	 */
889 	isrc->xi_close = 1;
890 	isrc->xi_virq = virq;
891 
892 	return (0);
893 }
894 
895 int
896 xen_intr_alloc_and_bind_ipi(u_int cpu, driver_filter_t filter,
897     enum intr_type flags, xen_intr_handle_t *port_handlep)
898 {
899 #ifdef SMP
900 	u_int vcpu_id = XEN_CPUID_TO_VCPUID(cpu);
901 	struct xenisrc *isrc;
902 	struct evtchn_bind_ipi bind_ipi = { .vcpu = vcpu_id };
903 	/* Same size as the one used by intr_handler->ih_name. */
904 	char name[MAXCOMLEN + 1];
905 	int error;
906 
907 	isrc = NULL;
908 	error = HYPERVISOR_event_channel_op(EVTCHNOP_bind_ipi, &bind_ipi);
909 	if (error != 0) {
910 		/*
911 		 * XXX Trap Hypercall error code Linuxisms in
912 		 *     the HYPERCALL layer.
913 		 */
914 		return (-error);
915 	}
916 
917 	snprintf(name, sizeof(name), "cpu%u", cpu);
918 
919 	error = xen_intr_bind_isrc(&isrc, bind_ipi.port, EVTCHN_TYPE_IPI,
920 	    name, filter, NULL, NULL, flags, port_handlep);
921 	if (error != 0) {
922 		evtchn_close_t close = { .port = bind_ipi.port };
923 
924 		xen_intr_unbind(port_handlep);
925 		if (HYPERVISOR_event_channel_op(EVTCHNOP_close, &close))
926 			panic("EVTCHNOP_close failed");
927 		return (error);
928 	}
929 
930 	if (isrc->xi_cpu != cpu) {
931 		/*
932 		 * Too early in the boot process for the generic interrupt
933 		 * code to perform the binding.  Update our event channel
934 		 * masks manually so events can't fire on the wrong cpu
935 		 * during AP startup.
936 		 */
937 		xen_intr_assign_cpu(isrc, cpu);
938 	}
939 
940 	/*
941 	 * The Event Channel API opened this port, so it is
942 	 * responsible for closing it automatically on unbind.
943 	 */
944 	isrc->xi_close = 1;
945 	return (0);
946 #else
947 	return (EOPNOTSUPP);
948 #endif
949 }
950 
951 int
952 xen_intr_describe(xen_intr_handle_t port_handle, const char *fmt, ...)
953 {
954 	char descr[MAXCOMLEN + 1];
955 	struct xenisrc *isrc;
956 	va_list ap;
957 
958 	isrc = xen_intr_isrc_from_handle(port_handle);
959 	if (isrc == NULL)
960 		return (EINVAL);
961 
962 	va_start(ap, fmt);
963 	vsnprintf(descr, sizeof(descr), fmt, ap);
964 	va_end(ap);
965 	return (xen_arch_intr_describe(isrc, isrc->xi_cookie, descr));
966 }
967 
968 void
969 xen_intr_unbind(xen_intr_handle_t *port_handlep)
970 {
971 	struct xenisrc *isrc;
972 
973 	KASSERT(port_handlep != NULL,
974 	    ("NULL xen_intr_handle_t passed to %s", __func__));
975 
976 	isrc = xen_intr_isrc_from_handle(*port_handlep);
977 	*port_handlep = NULL;
978 	if (isrc == NULL)
979 		return;
980 
981 	mtx_lock(&xen_intr_isrc_lock);
982 	if (refcount_release(&isrc->xi_refcount) == 0) {
983 		mtx_unlock(&xen_intr_isrc_lock);
984 		return;
985 	}
986 	mtx_unlock(&xen_intr_isrc_lock);
987 
988 	if (isrc->xi_cookie != NULL)
989 		xen_arch_intr_remove_handler(isrc, isrc->xi_cookie);
990 	xen_intr_release_isrc(isrc);
991 }
992 
993 void
994 xen_intr_signal(xen_intr_handle_t handle)
995 {
996 	struct xenisrc *isrc;
997 
998 	isrc = xen_intr_isrc_from_handle(handle);
999 	if (isrc != NULL) {
1000 		KASSERT(isrc->xi_type == EVTCHN_TYPE_PORT ||
1001 			isrc->xi_type == EVTCHN_TYPE_IPI,
1002 			("evtchn_signal on something other than a local port"));
1003 		struct evtchn_send send = { .port = isrc->xi_port };
1004 		(void)HYPERVISOR_event_channel_op(EVTCHNOP_send, &send);
1005 	}
1006 }
1007 
1008 evtchn_port_t
1009 xen_intr_port(xen_intr_handle_t handle)
1010 {
1011 	struct xenisrc *isrc;
1012 
1013 	isrc = xen_intr_isrc_from_handle(handle);
1014 	if (isrc == NULL)
1015 		return (0);
1016 
1017 	return (isrc->xi_port);
1018 }
1019 
1020 int
1021 xen_intr_add_handler(const char *name, driver_filter_t filter,
1022     driver_intr_t handler, void *arg, enum intr_type flags,
1023     xen_intr_handle_t handle)
1024 {
1025 	struct xenisrc *isrc;
1026 	int error;
1027 
1028 	isrc = xen_intr_isrc_from_handle(handle);
1029 	if (isrc == NULL || isrc->xi_cookie != NULL)
1030 		return (EINVAL);
1031 
1032 	error = xen_arch_intr_add_handler(name, filter, handler, arg,
1033 	    flags | INTR_EXCL, isrc, &isrc->xi_cookie);
1034 	if (error != 0)
1035 		printf("%s: %s: add handler failed: %d\n", name, __func__,
1036 		    error);
1037 
1038 	return (error);
1039 }
1040 
1041 int
1042 xen_intr_get_evtchn_from_port(evtchn_port_t port, xen_intr_handle_t *handlep)
1043 {
1044 
1045 	if (!is_valid_evtchn(port))
1046 		return (EINVAL);
1047 
1048 	if (handlep == NULL) {
1049 		return (EINVAL);
1050 	}
1051 
1052 	mtx_lock(&xen_intr_isrc_lock);
1053 	if (xen_intr_port_to_isrc[port] == NULL) {
1054 		mtx_unlock(&xen_intr_isrc_lock);
1055 		return (EINVAL);
1056 	}
1057 	refcount_acquire(&xen_intr_port_to_isrc[port]->xi_refcount);
1058 	mtx_unlock(&xen_intr_isrc_lock);
1059 
1060 	/* Assign the opaque handler */
1061 	*handlep = xen_intr_handle_from_isrc(xen_intr_port_to_isrc[port]);
1062 
1063 	return (0);
1064 }
1065 
1066 #ifdef DDB
1067 static const char *
1068 xen_intr_print_type(enum evtchn_type type)
1069 {
1070 	static const char *evtchn_type_to_string[EVTCHN_TYPE_COUNT] = {
1071 		[EVTCHN_TYPE_UNBOUND]	= "UNBOUND",
1072 		[EVTCHN_TYPE_VIRQ]	= "VIRQ",
1073 		[EVTCHN_TYPE_IPI]	= "IPI",
1074 		[EVTCHN_TYPE_PORT]	= "PORT",
1075 	};
1076 
1077 	if (type >= EVTCHN_TYPE_COUNT)
1078 		return ("UNKNOWN");
1079 
1080 	return (evtchn_type_to_string[type]);
1081 }
1082 
1083 static void
1084 xen_intr_dump_port(struct xenisrc *isrc)
1085 {
1086 	struct xen_intr_pcpu_data *pcpu;
1087 	shared_info_t *s = HYPERVISOR_shared_info;
1088 	u_int i;
1089 
1090 	db_printf("Port %d Type: %s\n",
1091 	    isrc->xi_port, xen_intr_print_type(isrc->xi_type));
1092 	if (isrc->xi_type == EVTCHN_TYPE_VIRQ)
1093 		db_printf("\tVirq: %u\n", isrc->xi_virq);
1094 
1095 	db_printf("\tMasked: %d Pending: %d\n",
1096 	    !!xen_test_bit(isrc->xi_port, &s->evtchn_mask[0]),
1097 	    !!xen_test_bit(isrc->xi_port, &s->evtchn_pending[0]));
1098 
1099 	db_printf("\tPer-CPU Masks: ");
1100 	CPU_FOREACH(i) {
1101 		pcpu = DPCPU_ID_PTR(i, xen_intr_pcpu);
1102 		db_printf("cpu#%u: %d ", i,
1103 		    !!xen_test_bit(isrc->xi_port, pcpu->evtchn_enabled));
1104 	}
1105 	db_printf("\n");
1106 }
1107 
1108 DB_SHOW_COMMAND(xen_evtchn, db_show_xen_evtchn)
1109 {
1110 	u_int i;
1111 
1112 	if (!xen_domain()) {
1113 		db_printf("Only available on Xen guests\n");
1114 		return;
1115 	}
1116 
1117 	for (i = 0; i < NR_EVENT_CHANNELS; i++) {
1118 		struct xenisrc *isrc;
1119 
1120 		isrc = xen_intr_port_to_isrc[i];
1121 		if (isrc == NULL)
1122 			continue;
1123 
1124 		xen_intr_dump_port(isrc);
1125 	}
1126 }
1127 #endif /* DDB */
1128