xref: /freebsd/sys/dev/usb/usb_transfer.c (revision 4f52dfbb)
1 /* $FreeBSD$ */
2 /*-
3  * SPDX-License-Identifier: BSD-2-Clause-FreeBSD
4  *
5  * Copyright (c) 2008 Hans Petter Selasky. All rights reserved.
6  *
7  * Redistribution and use in source and binary forms, with or without
8  * modification, are permitted provided that the following conditions
9  * are met:
10  * 1. Redistributions of source code must retain the above copyright
11  *    notice, this list of conditions and the following disclaimer.
12  * 2. Redistributions in binary form must reproduce the above copyright
13  *    notice, this list of conditions and the following disclaimer in the
14  *    documentation and/or other materials provided with the distribution.
15  *
16  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
17  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
18  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
19  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
20  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
21  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
22  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
23  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
24  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
25  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
26  * SUCH DAMAGE.
27  */
28 
29 #ifdef USB_GLOBAL_INCLUDE_FILE
30 #include USB_GLOBAL_INCLUDE_FILE
31 #else
32 #include <sys/stdint.h>
33 #include <sys/stddef.h>
34 #include <sys/param.h>
35 #include <sys/queue.h>
36 #include <sys/types.h>
37 #include <sys/systm.h>
38 #include <sys/kernel.h>
39 #include <sys/bus.h>
40 #include <sys/module.h>
41 #include <sys/lock.h>
42 #include <sys/mutex.h>
43 #include <sys/condvar.h>
44 #include <sys/sysctl.h>
45 #include <sys/sx.h>
46 #include <sys/unistd.h>
47 #include <sys/callout.h>
48 #include <sys/malloc.h>
49 #include <sys/priv.h>
50 
51 #include <dev/usb/usb.h>
52 #include <dev/usb/usbdi.h>
53 #include <dev/usb/usbdi_util.h>
54 
55 #define	USB_DEBUG_VAR usb_debug
56 
57 #include <dev/usb/usb_core.h>
58 #include <dev/usb/usb_busdma.h>
59 #include <dev/usb/usb_process.h>
60 #include <dev/usb/usb_transfer.h>
61 #include <dev/usb/usb_device.h>
62 #include <dev/usb/usb_debug.h>
63 #include <dev/usb/usb_util.h>
64 
65 #include <dev/usb/usb_controller.h>
66 #include <dev/usb/usb_bus.h>
67 #include <dev/usb/usb_pf.h>
68 #endif			/* USB_GLOBAL_INCLUDE_FILE */
69 
70 struct usb_std_packet_size {
71 	struct {
72 		uint16_t min;		/* inclusive */
73 		uint16_t max;		/* inclusive */
74 	}	range;
75 
76 	uint16_t fixed[4];
77 };
78 
79 static usb_callback_t usb_request_callback;
80 
81 static const struct usb_config usb_control_ep_cfg[USB_CTRL_XFER_MAX] = {
82 
83 	/* This transfer is used for generic control endpoint transfers */
84 
85 	[0] = {
86 		.type = UE_CONTROL,
87 		.endpoint = 0x00,	/* Control endpoint */
88 		.direction = UE_DIR_ANY,
89 		.bufsize = USB_EP0_BUFSIZE,	/* bytes */
90 		.flags = {.proxy_buffer = 1,},
91 		.callback = &usb_request_callback,
92 		.usb_mode = USB_MODE_DUAL,	/* both modes */
93 	},
94 
95 	/* This transfer is used for generic clear stall only */
96 
97 	[1] = {
98 		.type = UE_CONTROL,
99 		.endpoint = 0x00,	/* Control pipe */
100 		.direction = UE_DIR_ANY,
101 		.bufsize = sizeof(struct usb_device_request),
102 		.callback = &usb_do_clear_stall_callback,
103 		.timeout = 1000,	/* 1 second */
104 		.interval = 50,	/* 50ms */
105 		.usb_mode = USB_MODE_HOST,
106 	},
107 };
108 
109 /* function prototypes */
110 
111 static void	usbd_update_max_frame_size(struct usb_xfer *);
112 static void	usbd_transfer_unsetup_sub(struct usb_xfer_root *, uint8_t);
113 static void	usbd_control_transfer_init(struct usb_xfer *);
114 static int	usbd_setup_ctrl_transfer(struct usb_xfer *);
115 static void	usb_callback_proc(struct usb_proc_msg *);
116 static void	usbd_callback_ss_done_defer(struct usb_xfer *);
117 static void	usbd_callback_wrapper(struct usb_xfer_queue *);
118 static void	usbd_transfer_start_cb(void *);
119 static uint8_t	usbd_callback_wrapper_sub(struct usb_xfer *);
120 static void	usbd_get_std_packet_size(struct usb_std_packet_size *ptr,
121 		    uint8_t type, enum usb_dev_speed speed);
122 
123 /*------------------------------------------------------------------------*
124  *	usb_request_callback
125  *------------------------------------------------------------------------*/
126 static void
127 usb_request_callback(struct usb_xfer *xfer, usb_error_t error)
128 {
129 	if (xfer->flags_int.usb_mode == USB_MODE_DEVICE)
130 		usb_handle_request_callback(xfer, error);
131 	else
132 		usbd_do_request_callback(xfer, error);
133 }
134 
135 /*------------------------------------------------------------------------*
136  *	usbd_update_max_frame_size
137  *
138  * This function updates the maximum frame size, hence high speed USB
139  * can transfer multiple consecutive packets.
140  *------------------------------------------------------------------------*/
141 static void
142 usbd_update_max_frame_size(struct usb_xfer *xfer)
143 {
144 	/* compute maximum frame size */
145 	/* this computation should not overflow 16-bit */
146 	/* max = 15 * 1024 */
147 
148 	xfer->max_frame_size = xfer->max_packet_size * xfer->max_packet_count;
149 }
150 
151 /*------------------------------------------------------------------------*
152  *	usbd_get_dma_delay
153  *
154  * The following function is called when we need to
155  * synchronize with DMA hardware.
156  *
157  * Returns:
158  *    0: no DMA delay required
159  * Else: milliseconds of DMA delay
160  *------------------------------------------------------------------------*/
161 usb_timeout_t
162 usbd_get_dma_delay(struct usb_device *udev)
163 {
164 	const struct usb_bus_methods *mtod;
165 	uint32_t temp;
166 
167 	mtod = udev->bus->methods;
168 	temp = 0;
169 
170 	if (mtod->get_dma_delay) {
171 		(mtod->get_dma_delay) (udev, &temp);
172 		/*
173 		 * Round up and convert to milliseconds. Note that we use
174 		 * 1024 milliseconds per second. to save a division.
175 		 */
176 		temp += 0x3FF;
177 		temp /= 0x400;
178 	}
179 	return (temp);
180 }
181 
182 /*------------------------------------------------------------------------*
183  *	usbd_transfer_setup_sub_malloc
184  *
185  * This function will allocate one or more DMA'able memory chunks
186  * according to "size", "align" and "count" arguments. "ppc" is
187  * pointed to a linear array of USB page caches afterwards.
188  *
189  * If the "align" argument is equal to "1" a non-contiguous allocation
190  * can happen. Else if the "align" argument is greater than "1", the
191  * allocation will always be contiguous in memory.
192  *
193  * Returns:
194  *    0: Success
195  * Else: Failure
196  *------------------------------------------------------------------------*/
197 #if USB_HAVE_BUSDMA
198 uint8_t
199 usbd_transfer_setup_sub_malloc(struct usb_setup_params *parm,
200     struct usb_page_cache **ppc, usb_size_t size, usb_size_t align,
201     usb_size_t count)
202 {
203 	struct usb_page_cache *pc;
204 	struct usb_page *pg;
205 	void *buf;
206 	usb_size_t n_dma_pc;
207 	usb_size_t n_dma_pg;
208 	usb_size_t n_obj;
209 	usb_size_t x;
210 	usb_size_t y;
211 	usb_size_t r;
212 	usb_size_t z;
213 
214 	USB_ASSERT(align > 0, ("Invalid alignment, 0x%08x\n",
215 	    align));
216 	USB_ASSERT(size > 0, ("Invalid size = 0\n"));
217 
218 	if (count == 0) {
219 		return (0);		/* nothing to allocate */
220 	}
221 	/*
222 	 * Make sure that the size is aligned properly.
223 	 */
224 	size = -((-size) & (-align));
225 
226 	/*
227 	 * Try multi-allocation chunks to reduce the number of DMA
228 	 * allocations, hence DMA allocations are slow.
229 	 */
230 	if (align == 1) {
231 		/* special case - non-cached multi page DMA memory */
232 		n_dma_pc = count;
233 		n_dma_pg = (2 + (size / USB_PAGE_SIZE));
234 		n_obj = 1;
235 	} else if (size >= USB_PAGE_SIZE) {
236 		n_dma_pc = count;
237 		n_dma_pg = 1;
238 		n_obj = 1;
239 	} else {
240 		/* compute number of objects per page */
241 #ifdef USB_DMA_SINGLE_ALLOC
242 		n_obj = 1;
243 #else
244 		n_obj = (USB_PAGE_SIZE / size);
245 #endif
246 		/*
247 		 * Compute number of DMA chunks, rounded up
248 		 * to nearest one:
249 		 */
250 		n_dma_pc = howmany(count, n_obj);
251 		n_dma_pg = 1;
252 	}
253 
254 	/*
255 	 * DMA memory is allocated once, but mapped twice. That's why
256 	 * there is one list for auto-free and another list for
257 	 * non-auto-free which only holds the mapping and not the
258 	 * allocation.
259 	 */
260 	if (parm->buf == NULL) {
261 		/* reserve memory (auto-free) */
262 		parm->dma_page_ptr += n_dma_pc * n_dma_pg;
263 		parm->dma_page_cache_ptr += n_dma_pc;
264 
265 		/* reserve memory (no-auto-free) */
266 		parm->dma_page_ptr += count * n_dma_pg;
267 		parm->xfer_page_cache_ptr += count;
268 		return (0);
269 	}
270 	for (x = 0; x != n_dma_pc; x++) {
271 		/* need to initialize the page cache */
272 		parm->dma_page_cache_ptr[x].tag_parent =
273 		    &parm->curr_xfer->xroot->dma_parent_tag;
274 	}
275 	for (x = 0; x != count; x++) {
276 		/* need to initialize the page cache */
277 		parm->xfer_page_cache_ptr[x].tag_parent =
278 		    &parm->curr_xfer->xroot->dma_parent_tag;
279 	}
280 
281 	if (ppc != NULL) {
282 		if (n_obj != 1)
283 			*ppc = parm->xfer_page_cache_ptr;
284 		else
285 			*ppc = parm->dma_page_cache_ptr;
286 	}
287 	r = count;			/* set remainder count */
288 	z = n_obj * size;		/* set allocation size */
289 	pc = parm->xfer_page_cache_ptr;
290 	pg = parm->dma_page_ptr;
291 
292 	if (n_obj == 1) {
293 	    /*
294 	     * Avoid mapping memory twice if only a single object
295 	     * should be allocated per page cache:
296 	     */
297 	    for (x = 0; x != n_dma_pc; x++) {
298 		if (usb_pc_alloc_mem(parm->dma_page_cache_ptr,
299 		    pg, z, align)) {
300 			return (1);	/* failure */
301 		}
302 		/* Make room for one DMA page cache and "n_dma_pg" pages */
303 		parm->dma_page_cache_ptr++;
304 		pg += n_dma_pg;
305 	    }
306 	} else {
307 	    for (x = 0; x != n_dma_pc; x++) {
308 
309 		if (r < n_obj) {
310 			/* compute last remainder */
311 			z = r * size;
312 			n_obj = r;
313 		}
314 		if (usb_pc_alloc_mem(parm->dma_page_cache_ptr,
315 		    pg, z, align)) {
316 			return (1);	/* failure */
317 		}
318 		/* Set beginning of current buffer */
319 		buf = parm->dma_page_cache_ptr->buffer;
320 		/* Make room for one DMA page cache and "n_dma_pg" pages */
321 		parm->dma_page_cache_ptr++;
322 		pg += n_dma_pg;
323 
324 		for (y = 0; (y != n_obj); y++, r--, pc++, pg += n_dma_pg) {
325 
326 			/* Load sub-chunk into DMA */
327 			if (usb_pc_dmamap_create(pc, size)) {
328 				return (1);	/* failure */
329 			}
330 			pc->buffer = USB_ADD_BYTES(buf, y * size);
331 			pc->page_start = pg;
332 
333 			USB_MTX_LOCK(pc->tag_parent->mtx);
334 			if (usb_pc_load_mem(pc, size, 1 /* synchronous */ )) {
335 				USB_MTX_UNLOCK(pc->tag_parent->mtx);
336 				return (1);	/* failure */
337 			}
338 			USB_MTX_UNLOCK(pc->tag_parent->mtx);
339 		}
340 	    }
341 	}
342 
343 	parm->xfer_page_cache_ptr = pc;
344 	parm->dma_page_ptr = pg;
345 	return (0);
346 }
347 #endif
348 
349 /*------------------------------------------------------------------------*
350  *	usbd_transfer_setup_sub - transfer setup subroutine
351  *
352  * This function must be called from the "xfer_setup" callback of the
353  * USB Host or Device controller driver when setting up an USB
354  * transfer. This function will setup correct packet sizes, buffer
355  * sizes, flags and more, that are stored in the "usb_xfer"
356  * structure.
357  *------------------------------------------------------------------------*/
358 void
359 usbd_transfer_setup_sub(struct usb_setup_params *parm)
360 {
361 	enum {
362 		REQ_SIZE = 8,
363 		MIN_PKT = 8,
364 	};
365 	struct usb_xfer *xfer = parm->curr_xfer;
366 	const struct usb_config *setup = parm->curr_setup;
367 	struct usb_endpoint_ss_comp_descriptor *ecomp;
368 	struct usb_endpoint_descriptor *edesc;
369 	struct usb_std_packet_size std_size;
370 	usb_frcount_t n_frlengths;
371 	usb_frcount_t n_frbuffers;
372 	usb_frcount_t x;
373 	uint16_t maxp_old;
374 	uint8_t type;
375 	uint8_t zmps;
376 
377 	/*
378 	 * Sanity check. The following parameters must be initialized before
379 	 * calling this function.
380 	 */
381 	if ((parm->hc_max_packet_size == 0) ||
382 	    (parm->hc_max_packet_count == 0) ||
383 	    (parm->hc_max_frame_size == 0)) {
384 		parm->err = USB_ERR_INVAL;
385 		goto done;
386 	}
387 	edesc = xfer->endpoint->edesc;
388 	ecomp = xfer->endpoint->ecomp;
389 
390 	type = (edesc->bmAttributes & UE_XFERTYPE);
391 
392 	xfer->flags = setup->flags;
393 	xfer->nframes = setup->frames;
394 	xfer->timeout = setup->timeout;
395 	xfer->callback = setup->callback;
396 	xfer->interval = setup->interval;
397 	xfer->endpointno = edesc->bEndpointAddress;
398 	xfer->max_packet_size = UGETW(edesc->wMaxPacketSize);
399 	xfer->max_packet_count = 1;
400 	/* make a shadow copy: */
401 	xfer->flags_int.usb_mode = parm->udev->flags.usb_mode;
402 
403 	parm->bufsize = setup->bufsize;
404 
405 	switch (parm->speed) {
406 	case USB_SPEED_HIGH:
407 		switch (type) {
408 		case UE_ISOCHRONOUS:
409 		case UE_INTERRUPT:
410 			xfer->max_packet_count +=
411 			    (xfer->max_packet_size >> 11) & 3;
412 
413 			/* check for invalid max packet count */
414 			if (xfer->max_packet_count > 3)
415 				xfer->max_packet_count = 3;
416 			break;
417 		default:
418 			break;
419 		}
420 		xfer->max_packet_size &= 0x7FF;
421 		break;
422 	case USB_SPEED_SUPER:
423 		xfer->max_packet_count += (xfer->max_packet_size >> 11) & 3;
424 
425 		if (ecomp != NULL)
426 			xfer->max_packet_count += ecomp->bMaxBurst;
427 
428 		if ((xfer->max_packet_count == 0) ||
429 		    (xfer->max_packet_count > 16))
430 			xfer->max_packet_count = 16;
431 
432 		switch (type) {
433 		case UE_CONTROL:
434 			xfer->max_packet_count = 1;
435 			break;
436 		case UE_ISOCHRONOUS:
437 			if (ecomp != NULL) {
438 				uint8_t mult;
439 
440 				mult = UE_GET_SS_ISO_MULT(
441 				    ecomp->bmAttributes) + 1;
442 				if (mult > 3)
443 					mult = 3;
444 
445 				xfer->max_packet_count *= mult;
446 			}
447 			break;
448 		default:
449 			break;
450 		}
451 		xfer->max_packet_size &= 0x7FF;
452 		break;
453 	default:
454 		break;
455 	}
456 	/* range check "max_packet_count" */
457 
458 	if (xfer->max_packet_count > parm->hc_max_packet_count) {
459 		xfer->max_packet_count = parm->hc_max_packet_count;
460 	}
461 
462 	/* store max packet size value before filtering */
463 
464 	maxp_old = xfer->max_packet_size;
465 
466 	/* filter "wMaxPacketSize" according to HC capabilities */
467 
468 	if ((xfer->max_packet_size > parm->hc_max_packet_size) ||
469 	    (xfer->max_packet_size == 0)) {
470 		xfer->max_packet_size = parm->hc_max_packet_size;
471 	}
472 	/* filter "wMaxPacketSize" according to standard sizes */
473 
474 	usbd_get_std_packet_size(&std_size, type, parm->speed);
475 
476 	if (std_size.range.min || std_size.range.max) {
477 
478 		if (xfer->max_packet_size < std_size.range.min) {
479 			xfer->max_packet_size = std_size.range.min;
480 		}
481 		if (xfer->max_packet_size > std_size.range.max) {
482 			xfer->max_packet_size = std_size.range.max;
483 		}
484 	} else {
485 
486 		if (xfer->max_packet_size >= std_size.fixed[3]) {
487 			xfer->max_packet_size = std_size.fixed[3];
488 		} else if (xfer->max_packet_size >= std_size.fixed[2]) {
489 			xfer->max_packet_size = std_size.fixed[2];
490 		} else if (xfer->max_packet_size >= std_size.fixed[1]) {
491 			xfer->max_packet_size = std_size.fixed[1];
492 		} else {
493 			/* only one possibility left */
494 			xfer->max_packet_size = std_size.fixed[0];
495 		}
496 	}
497 
498 	/*
499 	 * Check if the max packet size was outside its allowed range
500 	 * and clamped to a valid value:
501 	 */
502 	if (maxp_old != xfer->max_packet_size)
503 		xfer->flags_int.maxp_was_clamped = 1;
504 
505 	/* compute "max_frame_size" */
506 
507 	usbd_update_max_frame_size(xfer);
508 
509 	/* check interrupt interval and transfer pre-delay */
510 
511 	if (type == UE_ISOCHRONOUS) {
512 
513 		uint16_t frame_limit;
514 
515 		xfer->interval = 0;	/* not used, must be zero */
516 		xfer->flags_int.isochronous_xfr = 1;	/* set flag */
517 
518 		if (xfer->timeout == 0) {
519 			/*
520 			 * set a default timeout in
521 			 * case something goes wrong!
522 			 */
523 			xfer->timeout = 1000 / 4;
524 		}
525 		switch (parm->speed) {
526 		case USB_SPEED_LOW:
527 		case USB_SPEED_FULL:
528 			frame_limit = USB_MAX_FS_ISOC_FRAMES_PER_XFER;
529 			xfer->fps_shift = 0;
530 			break;
531 		default:
532 			frame_limit = USB_MAX_HS_ISOC_FRAMES_PER_XFER;
533 			xfer->fps_shift = edesc->bInterval;
534 			if (xfer->fps_shift > 0)
535 				xfer->fps_shift--;
536 			if (xfer->fps_shift > 3)
537 				xfer->fps_shift = 3;
538 			if (xfer->flags.pre_scale_frames != 0)
539 				xfer->nframes <<= (3 - xfer->fps_shift);
540 			break;
541 		}
542 
543 		if (xfer->nframes > frame_limit) {
544 			/*
545 			 * this is not going to work
546 			 * cross hardware
547 			 */
548 			parm->err = USB_ERR_INVAL;
549 			goto done;
550 		}
551 		if (xfer->nframes == 0) {
552 			/*
553 			 * this is not a valid value
554 			 */
555 			parm->err = USB_ERR_ZERO_NFRAMES;
556 			goto done;
557 		}
558 	} else {
559 
560 		/*
561 		 * If a value is specified use that else check the
562 		 * endpoint descriptor!
563 		 */
564 		if (type == UE_INTERRUPT) {
565 
566 			uint32_t temp;
567 
568 			if (xfer->interval == 0) {
569 
570 				xfer->interval = edesc->bInterval;
571 
572 				switch (parm->speed) {
573 				case USB_SPEED_LOW:
574 				case USB_SPEED_FULL:
575 					break;
576 				default:
577 					/* 125us -> 1ms */
578 					if (xfer->interval < 4)
579 						xfer->interval = 1;
580 					else if (xfer->interval > 16)
581 						xfer->interval = (1 << (16 - 4));
582 					else
583 						xfer->interval =
584 						    (1 << (xfer->interval - 4));
585 					break;
586 				}
587 			}
588 
589 			if (xfer->interval == 0) {
590 				/*
591 				 * One millisecond is the smallest
592 				 * interval we support:
593 				 */
594 				xfer->interval = 1;
595 			}
596 
597 			xfer->fps_shift = 0;
598 			temp = 1;
599 
600 			while ((temp != 0) && (temp < xfer->interval)) {
601 				xfer->fps_shift++;
602 				temp *= 2;
603 			}
604 
605 			switch (parm->speed) {
606 			case USB_SPEED_LOW:
607 			case USB_SPEED_FULL:
608 				break;
609 			default:
610 				xfer->fps_shift += 3;
611 				break;
612 			}
613 		}
614 	}
615 
616 	/*
617 	 * NOTE: we do not allow "max_packet_size" or "max_frame_size"
618 	 * to be equal to zero when setting up USB transfers, hence
619 	 * this leads to a lot of extra code in the USB kernel.
620 	 */
621 
622 	if ((xfer->max_frame_size == 0) ||
623 	    (xfer->max_packet_size == 0)) {
624 
625 		zmps = 1;
626 
627 		if ((parm->bufsize <= MIN_PKT) &&
628 		    (type != UE_CONTROL) &&
629 		    (type != UE_BULK)) {
630 
631 			/* workaround */
632 			xfer->max_packet_size = MIN_PKT;
633 			xfer->max_packet_count = 1;
634 			parm->bufsize = 0;	/* automatic setup length */
635 			usbd_update_max_frame_size(xfer);
636 
637 		} else {
638 			parm->err = USB_ERR_ZERO_MAXP;
639 			goto done;
640 		}
641 
642 	} else {
643 		zmps = 0;
644 	}
645 
646 	/*
647 	 * check if we should setup a default
648 	 * length:
649 	 */
650 
651 	if (parm->bufsize == 0) {
652 
653 		parm->bufsize = xfer->max_frame_size;
654 
655 		if (type == UE_ISOCHRONOUS) {
656 			parm->bufsize *= xfer->nframes;
657 		}
658 	}
659 	/*
660 	 * check if we are about to setup a proxy
661 	 * type of buffer:
662 	 */
663 
664 	if (xfer->flags.proxy_buffer) {
665 
666 		/* round bufsize up */
667 
668 		parm->bufsize += (xfer->max_frame_size - 1);
669 
670 		if (parm->bufsize < xfer->max_frame_size) {
671 			/* length wrapped around */
672 			parm->err = USB_ERR_INVAL;
673 			goto done;
674 		}
675 		/* subtract remainder */
676 
677 		parm->bufsize -= (parm->bufsize % xfer->max_frame_size);
678 
679 		/* add length of USB device request structure, if any */
680 
681 		if (type == UE_CONTROL) {
682 			parm->bufsize += REQ_SIZE;	/* SETUP message */
683 		}
684 	}
685 	xfer->max_data_length = parm->bufsize;
686 
687 	/* Setup "n_frlengths" and "n_frbuffers" */
688 
689 	if (type == UE_ISOCHRONOUS) {
690 		n_frlengths = xfer->nframes;
691 		n_frbuffers = 1;
692 	} else {
693 
694 		if (type == UE_CONTROL) {
695 			xfer->flags_int.control_xfr = 1;
696 			if (xfer->nframes == 0) {
697 				if (parm->bufsize <= REQ_SIZE) {
698 					/*
699 					 * there will never be any data
700 					 * stage
701 					 */
702 					xfer->nframes = 1;
703 				} else {
704 					xfer->nframes = 2;
705 				}
706 			}
707 		} else {
708 			if (xfer->nframes == 0) {
709 				xfer->nframes = 1;
710 			}
711 		}
712 
713 		n_frlengths = xfer->nframes;
714 		n_frbuffers = xfer->nframes;
715 	}
716 
717 	/*
718 	 * check if we have room for the
719 	 * USB device request structure:
720 	 */
721 
722 	if (type == UE_CONTROL) {
723 
724 		if (xfer->max_data_length < REQ_SIZE) {
725 			/* length wrapped around or too small bufsize */
726 			parm->err = USB_ERR_INVAL;
727 			goto done;
728 		}
729 		xfer->max_data_length -= REQ_SIZE;
730 	}
731 	/*
732 	 * Setup "frlengths" and shadow "frlengths" for keeping the
733 	 * initial frame lengths when a USB transfer is complete. This
734 	 * information is useful when computing isochronous offsets.
735 	 */
736 	xfer->frlengths = parm->xfer_length_ptr;
737 	parm->xfer_length_ptr += 2 * n_frlengths;
738 
739 	/* setup "frbuffers" */
740 	xfer->frbuffers = parm->xfer_page_cache_ptr;
741 	parm->xfer_page_cache_ptr += n_frbuffers;
742 
743 	/* initialize max frame count */
744 	xfer->max_frame_count = xfer->nframes;
745 
746 	/*
747 	 * check if we need to setup
748 	 * a local buffer:
749 	 */
750 
751 	if (!xfer->flags.ext_buffer) {
752 #if USB_HAVE_BUSDMA
753 		struct usb_page_search page_info;
754 		struct usb_page_cache *pc;
755 
756 		if (usbd_transfer_setup_sub_malloc(parm,
757 		    &pc, parm->bufsize, 1, 1)) {
758 			parm->err = USB_ERR_NOMEM;
759 		} else if (parm->buf != NULL) {
760 
761 			usbd_get_page(pc, 0, &page_info);
762 
763 			xfer->local_buffer = page_info.buffer;
764 
765 			usbd_xfer_set_frame_offset(xfer, 0, 0);
766 
767 			if ((type == UE_CONTROL) && (n_frbuffers > 1)) {
768 				usbd_xfer_set_frame_offset(xfer, REQ_SIZE, 1);
769 			}
770 		}
771 #else
772 		/* align data */
773 		parm->size[0] += ((-parm->size[0]) & (USB_HOST_ALIGN - 1));
774 
775 		if (parm->buf != NULL) {
776 			xfer->local_buffer =
777 			    USB_ADD_BYTES(parm->buf, parm->size[0]);
778 
779 			usbd_xfer_set_frame_offset(xfer, 0, 0);
780 
781 			if ((type == UE_CONTROL) && (n_frbuffers > 1)) {
782 				usbd_xfer_set_frame_offset(xfer, REQ_SIZE, 1);
783 			}
784 		}
785 		parm->size[0] += parm->bufsize;
786 
787 		/* align data again */
788 		parm->size[0] += ((-parm->size[0]) & (USB_HOST_ALIGN - 1));
789 #endif
790 	}
791 	/*
792 	 * Compute maximum buffer size
793 	 */
794 
795 	if (parm->bufsize_max < parm->bufsize) {
796 		parm->bufsize_max = parm->bufsize;
797 	}
798 #if USB_HAVE_BUSDMA
799 	if (xfer->flags_int.bdma_enable) {
800 		/*
801 		 * Setup "dma_page_ptr".
802 		 *
803 		 * Proof for formula below:
804 		 *
805 		 * Assume there are three USB frames having length "a", "b" and
806 		 * "c". These USB frames will at maximum need "z"
807 		 * "usb_page" structures. "z" is given by:
808 		 *
809 		 * z = ((a / USB_PAGE_SIZE) + 2) + ((b / USB_PAGE_SIZE) + 2) +
810 		 * ((c / USB_PAGE_SIZE) + 2);
811 		 *
812 		 * Constraining "a", "b" and "c" like this:
813 		 *
814 		 * (a + b + c) <= parm->bufsize
815 		 *
816 		 * We know that:
817 		 *
818 		 * z <= ((parm->bufsize / USB_PAGE_SIZE) + (3*2));
819 		 *
820 		 * Here is the general formula:
821 		 */
822 		xfer->dma_page_ptr = parm->dma_page_ptr;
823 		parm->dma_page_ptr += (2 * n_frbuffers);
824 		parm->dma_page_ptr += (parm->bufsize / USB_PAGE_SIZE);
825 	}
826 #endif
827 	if (zmps) {
828 		/* correct maximum data length */
829 		xfer->max_data_length = 0;
830 	}
831 	/* subtract USB frame remainder from "hc_max_frame_size" */
832 
833 	xfer->max_hc_frame_size =
834 	    (parm->hc_max_frame_size -
835 	    (parm->hc_max_frame_size % xfer->max_frame_size));
836 
837 	if (xfer->max_hc_frame_size == 0) {
838 		parm->err = USB_ERR_INVAL;
839 		goto done;
840 	}
841 
842 	/* initialize frame buffers */
843 
844 	if (parm->buf) {
845 		for (x = 0; x != n_frbuffers; x++) {
846 			xfer->frbuffers[x].tag_parent =
847 			    &xfer->xroot->dma_parent_tag;
848 #if USB_HAVE_BUSDMA
849 			if (xfer->flags_int.bdma_enable &&
850 			    (parm->bufsize_max > 0)) {
851 
852 				if (usb_pc_dmamap_create(
853 				    xfer->frbuffers + x,
854 				    parm->bufsize_max)) {
855 					parm->err = USB_ERR_NOMEM;
856 					goto done;
857 				}
858 			}
859 #endif
860 		}
861 	}
862 done:
863 	if (parm->err) {
864 		/*
865 		 * Set some dummy values so that we avoid division by zero:
866 		 */
867 		xfer->max_hc_frame_size = 1;
868 		xfer->max_frame_size = 1;
869 		xfer->max_packet_size = 1;
870 		xfer->max_data_length = 0;
871 		xfer->nframes = 0;
872 		xfer->max_frame_count = 0;
873 	}
874 }
875 
876 static uint8_t
877 usbd_transfer_setup_has_bulk(const struct usb_config *setup_start,
878     uint16_t n_setup)
879 {
880 	while (n_setup--) {
881 		uint8_t type = setup_start[n_setup].type;
882 		if (type == UE_BULK || type == UE_BULK_INTR ||
883 		    type == UE_TYPE_ANY)
884 			return (1);
885 	}
886 	return (0);
887 }
888 
889 /*------------------------------------------------------------------------*
890  *	usbd_transfer_setup - setup an array of USB transfers
891  *
892  * NOTE: You must always call "usbd_transfer_unsetup" after calling
893  * "usbd_transfer_setup" if success was returned.
894  *
895  * The idea is that the USB device driver should pre-allocate all its
896  * transfers by one call to this function.
897  *
898  * Return values:
899  *    0: Success
900  * Else: Failure
901  *------------------------------------------------------------------------*/
902 usb_error_t
903 usbd_transfer_setup(struct usb_device *udev,
904     const uint8_t *ifaces, struct usb_xfer **ppxfer,
905     const struct usb_config *setup_start, uint16_t n_setup,
906     void *priv_sc, struct mtx *xfer_mtx)
907 {
908 	const struct usb_config *setup_end = setup_start + n_setup;
909 	const struct usb_config *setup;
910 	struct usb_setup_params *parm;
911 	struct usb_endpoint *ep;
912 	struct usb_xfer_root *info;
913 	struct usb_xfer *xfer;
914 	void *buf = NULL;
915 	usb_error_t error = 0;
916 	uint16_t n;
917 	uint16_t refcount;
918 	uint8_t do_unlock;
919 
920 	WITNESS_WARN(WARN_GIANTOK | WARN_SLEEPOK, NULL,
921 	    "usbd_transfer_setup can sleep!");
922 
923 	/* do some checking first */
924 
925 	if (n_setup == 0) {
926 		DPRINTFN(6, "setup array has zero length!\n");
927 		return (USB_ERR_INVAL);
928 	}
929 	if (ifaces == NULL) {
930 		DPRINTFN(6, "ifaces array is NULL!\n");
931 		return (USB_ERR_INVAL);
932 	}
933 	if (xfer_mtx == NULL) {
934 		DPRINTFN(6, "using global lock\n");
935 		xfer_mtx = &Giant;
936 	}
937 
938 	/* more sanity checks */
939 
940 	for (setup = setup_start, n = 0;
941 	    setup != setup_end; setup++, n++) {
942 		if (setup->bufsize == (usb_frlength_t)-1) {
943 			error = USB_ERR_BAD_BUFSIZE;
944 			DPRINTF("invalid bufsize\n");
945 		}
946 		if (setup->callback == NULL) {
947 			error = USB_ERR_NO_CALLBACK;
948 			DPRINTF("no callback\n");
949 		}
950 		ppxfer[n] = NULL;
951 	}
952 
953 	if (error)
954 		return (error);
955 
956 	/* Protect scratch area */
957 	do_unlock = usbd_ctrl_lock(udev);
958 
959 	refcount = 0;
960 	info = NULL;
961 
962 	parm = &udev->scratch.xfer_setup[0].parm;
963 	memset(parm, 0, sizeof(*parm));
964 
965 	parm->udev = udev;
966 	parm->speed = usbd_get_speed(udev);
967 	parm->hc_max_packet_count = 1;
968 
969 	if (parm->speed >= USB_SPEED_MAX) {
970 		parm->err = USB_ERR_INVAL;
971 		goto done;
972 	}
973 	/* setup all transfers */
974 
975 	while (1) {
976 
977 		if (buf) {
978 			/*
979 			 * Initialize the "usb_xfer_root" structure,
980 			 * which is common for all our USB transfers.
981 			 */
982 			info = USB_ADD_BYTES(buf, 0);
983 
984 			info->memory_base = buf;
985 			info->memory_size = parm->size[0];
986 
987 #if USB_HAVE_BUSDMA
988 			info->dma_page_cache_start = USB_ADD_BYTES(buf, parm->size[4]);
989 			info->dma_page_cache_end = USB_ADD_BYTES(buf, parm->size[5]);
990 #endif
991 			info->xfer_page_cache_start = USB_ADD_BYTES(buf, parm->size[5]);
992 			info->xfer_page_cache_end = USB_ADD_BYTES(buf, parm->size[2]);
993 
994 			cv_init(&info->cv_drain, "WDRAIN");
995 
996 			info->xfer_mtx = xfer_mtx;
997 #if USB_HAVE_BUSDMA
998 			usb_dma_tag_setup(&info->dma_parent_tag,
999 			    parm->dma_tag_p, udev->bus->dma_parent_tag[0].tag,
1000 			    xfer_mtx, &usb_bdma_done_event, udev->bus->dma_bits,
1001 			    parm->dma_tag_max);
1002 #endif
1003 
1004 			info->bus = udev->bus;
1005 			info->udev = udev;
1006 
1007 			TAILQ_INIT(&info->done_q.head);
1008 			info->done_q.command = &usbd_callback_wrapper;
1009 #if USB_HAVE_BUSDMA
1010 			TAILQ_INIT(&info->dma_q.head);
1011 			info->dma_q.command = &usb_bdma_work_loop;
1012 #endif
1013 			info->done_m[0].hdr.pm_callback = &usb_callback_proc;
1014 			info->done_m[0].xroot = info;
1015 			info->done_m[1].hdr.pm_callback = &usb_callback_proc;
1016 			info->done_m[1].xroot = info;
1017 
1018 			/*
1019 			 * In device side mode control endpoint
1020 			 * requests need to run from a separate
1021 			 * context, else there is a chance of
1022 			 * deadlock!
1023 			 */
1024 			if (setup_start == usb_control_ep_cfg)
1025 				info->done_p =
1026 				    USB_BUS_CONTROL_XFER_PROC(udev->bus);
1027 			else if (xfer_mtx == &Giant)
1028 				info->done_p =
1029 				    USB_BUS_GIANT_PROC(udev->bus);
1030 			else if (usbd_transfer_setup_has_bulk(setup_start, n_setup))
1031 				info->done_p =
1032 				    USB_BUS_NON_GIANT_BULK_PROC(udev->bus);
1033 			else
1034 				info->done_p =
1035 				    USB_BUS_NON_GIANT_ISOC_PROC(udev->bus);
1036 		}
1037 		/* reset sizes */
1038 
1039 		parm->size[0] = 0;
1040 		parm->buf = buf;
1041 		parm->size[0] += sizeof(info[0]);
1042 
1043 		for (setup = setup_start, n = 0;
1044 		    setup != setup_end; setup++, n++) {
1045 
1046 			/* skip USB transfers without callbacks: */
1047 			if (setup->callback == NULL) {
1048 				continue;
1049 			}
1050 			/* see if there is a matching endpoint */
1051 			ep = usbd_get_endpoint(udev,
1052 			    ifaces[setup->if_index], setup);
1053 
1054 			/*
1055 			 * Check that the USB PIPE is valid and that
1056 			 * the endpoint mode is proper.
1057 			 *
1058 			 * Make sure we don't allocate a streams
1059 			 * transfer when such a combination is not
1060 			 * valid.
1061 			 */
1062 			if ((ep == NULL) || (ep->methods == NULL) ||
1063 			    ((ep->ep_mode != USB_EP_MODE_STREAMS) &&
1064 			    (ep->ep_mode != USB_EP_MODE_DEFAULT)) ||
1065 			    (setup->stream_id != 0 &&
1066 			    (setup->stream_id >= USB_MAX_EP_STREAMS ||
1067 			    (ep->ep_mode != USB_EP_MODE_STREAMS)))) {
1068 				if (setup->flags.no_pipe_ok)
1069 					continue;
1070 				if ((setup->usb_mode != USB_MODE_DUAL) &&
1071 				    (setup->usb_mode != udev->flags.usb_mode))
1072 					continue;
1073 				parm->err = USB_ERR_NO_PIPE;
1074 				goto done;
1075 			}
1076 
1077 			/* align data properly */
1078 			parm->size[0] += ((-parm->size[0]) & (USB_HOST_ALIGN - 1));
1079 
1080 			/* store current setup pointer */
1081 			parm->curr_setup = setup;
1082 
1083 			if (buf) {
1084 				/*
1085 				 * Common initialization of the
1086 				 * "usb_xfer" structure.
1087 				 */
1088 				xfer = USB_ADD_BYTES(buf, parm->size[0]);
1089 				xfer->address = udev->address;
1090 				xfer->priv_sc = priv_sc;
1091 				xfer->xroot = info;
1092 
1093 				usb_callout_init_mtx(&xfer->timeout_handle,
1094 				    &udev->bus->bus_mtx, 0);
1095 			} else {
1096 				/*
1097 				 * Setup a dummy xfer, hence we are
1098 				 * writing to the "usb_xfer"
1099 				 * structure pointed to by "xfer"
1100 				 * before we have allocated any
1101 				 * memory:
1102 				 */
1103 				xfer = &udev->scratch.xfer_setup[0].dummy;
1104 				memset(xfer, 0, sizeof(*xfer));
1105 				refcount++;
1106 			}
1107 
1108 			/* set transfer endpoint pointer */
1109 			xfer->endpoint = ep;
1110 
1111 			/* set transfer stream ID */
1112 			xfer->stream_id = setup->stream_id;
1113 
1114 			parm->size[0] += sizeof(xfer[0]);
1115 			parm->methods = xfer->endpoint->methods;
1116 			parm->curr_xfer = xfer;
1117 
1118 			/*
1119 			 * Call the Host or Device controller transfer
1120 			 * setup routine:
1121 			 */
1122 			(udev->bus->methods->xfer_setup) (parm);
1123 
1124 			/* check for error */
1125 			if (parm->err)
1126 				goto done;
1127 
1128 			if (buf) {
1129 				/*
1130 				 * Increment the endpoint refcount. This
1131 				 * basically prevents setting a new
1132 				 * configuration and alternate setting
1133 				 * when USB transfers are in use on
1134 				 * the given interface. Search the USB
1135 				 * code for "endpoint->refcount_alloc" if you
1136 				 * want more information.
1137 				 */
1138 				USB_BUS_LOCK(info->bus);
1139 				if (xfer->endpoint->refcount_alloc >= USB_EP_REF_MAX)
1140 					parm->err = USB_ERR_INVAL;
1141 
1142 				xfer->endpoint->refcount_alloc++;
1143 
1144 				if (xfer->endpoint->refcount_alloc == 0)
1145 					panic("usbd_transfer_setup(): Refcount wrapped to zero\n");
1146 				USB_BUS_UNLOCK(info->bus);
1147 
1148 				/*
1149 				 * Whenever we set ppxfer[] then we
1150 				 * also need to increment the
1151 				 * "setup_refcount":
1152 				 */
1153 				info->setup_refcount++;
1154 
1155 				/*
1156 				 * Transfer is successfully setup and
1157 				 * can be used:
1158 				 */
1159 				ppxfer[n] = xfer;
1160 			}
1161 
1162 			/* check for error */
1163 			if (parm->err)
1164 				goto done;
1165 		}
1166 
1167 		if (buf != NULL || parm->err != 0)
1168 			goto done;
1169 
1170 		/* if no transfers, nothing to do */
1171 		if (refcount == 0)
1172 			goto done;
1173 
1174 		/* align data properly */
1175 		parm->size[0] += ((-parm->size[0]) & (USB_HOST_ALIGN - 1));
1176 
1177 		/* store offset temporarily */
1178 		parm->size[1] = parm->size[0];
1179 
1180 		/*
1181 		 * The number of DMA tags required depends on
1182 		 * the number of endpoints. The current estimate
1183 		 * for maximum number of DMA tags per endpoint
1184 		 * is three:
1185 		 * 1) for loading memory
1186 		 * 2) for allocating memory
1187 		 * 3) for fixing memory [UHCI]
1188 		 */
1189 		parm->dma_tag_max += 3 * MIN(n_setup, USB_EP_MAX);
1190 
1191 		/*
1192 		 * DMA tags for QH, TD, Data and more.
1193 		 */
1194 		parm->dma_tag_max += 8;
1195 
1196 		parm->dma_tag_p += parm->dma_tag_max;
1197 
1198 		parm->size[0] += ((uint8_t *)parm->dma_tag_p) -
1199 		    ((uint8_t *)0);
1200 
1201 		/* align data properly */
1202 		parm->size[0] += ((-parm->size[0]) & (USB_HOST_ALIGN - 1));
1203 
1204 		/* store offset temporarily */
1205 		parm->size[3] = parm->size[0];
1206 
1207 		parm->size[0] += ((uint8_t *)parm->dma_page_ptr) -
1208 		    ((uint8_t *)0);
1209 
1210 		/* align data properly */
1211 		parm->size[0] += ((-parm->size[0]) & (USB_HOST_ALIGN - 1));
1212 
1213 		/* store offset temporarily */
1214 		parm->size[4] = parm->size[0];
1215 
1216 		parm->size[0] += ((uint8_t *)parm->dma_page_cache_ptr) -
1217 		    ((uint8_t *)0);
1218 
1219 		/* store end offset temporarily */
1220 		parm->size[5] = parm->size[0];
1221 
1222 		parm->size[0] += ((uint8_t *)parm->xfer_page_cache_ptr) -
1223 		    ((uint8_t *)0);
1224 
1225 		/* store end offset temporarily */
1226 
1227 		parm->size[2] = parm->size[0];
1228 
1229 		/* align data properly */
1230 		parm->size[0] += ((-parm->size[0]) & (USB_HOST_ALIGN - 1));
1231 
1232 		parm->size[6] = parm->size[0];
1233 
1234 		parm->size[0] += ((uint8_t *)parm->xfer_length_ptr) -
1235 		    ((uint8_t *)0);
1236 
1237 		/* align data properly */
1238 		parm->size[0] += ((-parm->size[0]) & (USB_HOST_ALIGN - 1));
1239 
1240 		/* allocate zeroed memory */
1241 		buf = malloc(parm->size[0], M_USB, M_WAITOK | M_ZERO);
1242 
1243 		if (buf == NULL) {
1244 			parm->err = USB_ERR_NOMEM;
1245 			DPRINTFN(0, "cannot allocate memory block for "
1246 			    "configuration (%d bytes)\n",
1247 			    parm->size[0]);
1248 			goto done;
1249 		}
1250 		parm->dma_tag_p = USB_ADD_BYTES(buf, parm->size[1]);
1251 		parm->dma_page_ptr = USB_ADD_BYTES(buf, parm->size[3]);
1252 		parm->dma_page_cache_ptr = USB_ADD_BYTES(buf, parm->size[4]);
1253 		parm->xfer_page_cache_ptr = USB_ADD_BYTES(buf, parm->size[5]);
1254 		parm->xfer_length_ptr = USB_ADD_BYTES(buf, parm->size[6]);
1255 	}
1256 
1257 done:
1258 	if (buf) {
1259 		if (info->setup_refcount == 0) {
1260 			/*
1261 			 * "usbd_transfer_unsetup_sub" will unlock
1262 			 * the bus mutex before returning !
1263 			 */
1264 			USB_BUS_LOCK(info->bus);
1265 
1266 			/* something went wrong */
1267 			usbd_transfer_unsetup_sub(info, 0);
1268 		}
1269 	}
1270 
1271 	/* check if any errors happened */
1272 	if (parm->err)
1273 		usbd_transfer_unsetup(ppxfer, n_setup);
1274 
1275 	error = parm->err;
1276 
1277 	if (do_unlock)
1278 		usbd_ctrl_unlock(udev);
1279 
1280 	return (error);
1281 }
1282 
1283 /*------------------------------------------------------------------------*
1284  *	usbd_transfer_unsetup_sub - factored out code
1285  *------------------------------------------------------------------------*/
1286 static void
1287 usbd_transfer_unsetup_sub(struct usb_xfer_root *info, uint8_t needs_delay)
1288 {
1289 #if USB_HAVE_BUSDMA
1290 	struct usb_page_cache *pc;
1291 #endif
1292 
1293 	USB_BUS_LOCK_ASSERT(info->bus, MA_OWNED);
1294 
1295 	/* wait for any outstanding DMA operations */
1296 
1297 	if (needs_delay) {
1298 		usb_timeout_t temp;
1299 		temp = usbd_get_dma_delay(info->udev);
1300 		if (temp != 0) {
1301 			usb_pause_mtx(&info->bus->bus_mtx,
1302 			    USB_MS_TO_TICKS(temp));
1303 		}
1304 	}
1305 
1306 	/* make sure that our done messages are not queued anywhere */
1307 	usb_proc_mwait(info->done_p, &info->done_m[0], &info->done_m[1]);
1308 
1309 	USB_BUS_UNLOCK(info->bus);
1310 
1311 #if USB_HAVE_BUSDMA
1312 	/* free DMA'able memory, if any */
1313 	pc = info->dma_page_cache_start;
1314 	while (pc != info->dma_page_cache_end) {
1315 		usb_pc_free_mem(pc);
1316 		pc++;
1317 	}
1318 
1319 	/* free DMA maps in all "xfer->frbuffers" */
1320 	pc = info->xfer_page_cache_start;
1321 	while (pc != info->xfer_page_cache_end) {
1322 		usb_pc_dmamap_destroy(pc);
1323 		pc++;
1324 	}
1325 
1326 	/* free all DMA tags */
1327 	usb_dma_tag_unsetup(&info->dma_parent_tag);
1328 #endif
1329 
1330 	cv_destroy(&info->cv_drain);
1331 
1332 	/*
1333 	 * free the "memory_base" last, hence the "info" structure is
1334 	 * contained within the "memory_base"!
1335 	 */
1336 	free(info->memory_base, M_USB);
1337 }
1338 
1339 /*------------------------------------------------------------------------*
1340  *	usbd_transfer_unsetup - unsetup/free an array of USB transfers
1341  *
1342  * NOTE: All USB transfers in progress will get called back passing
1343  * the error code "USB_ERR_CANCELLED" before this function
1344  * returns.
1345  *------------------------------------------------------------------------*/
1346 void
1347 usbd_transfer_unsetup(struct usb_xfer **pxfer, uint16_t n_setup)
1348 {
1349 	struct usb_xfer *xfer;
1350 	struct usb_xfer_root *info;
1351 	uint8_t needs_delay = 0;
1352 
1353 	WITNESS_WARN(WARN_GIANTOK | WARN_SLEEPOK, NULL,
1354 	    "usbd_transfer_unsetup can sleep!");
1355 
1356 	while (n_setup--) {
1357 		xfer = pxfer[n_setup];
1358 
1359 		if (xfer == NULL)
1360 			continue;
1361 
1362 		info = xfer->xroot;
1363 
1364 		USB_XFER_LOCK(xfer);
1365 		USB_BUS_LOCK(info->bus);
1366 
1367 		/*
1368 		 * HINT: when you start/stop a transfer, it might be a
1369 		 * good idea to directly use the "pxfer[]" structure:
1370 		 *
1371 		 * usbd_transfer_start(sc->pxfer[0]);
1372 		 * usbd_transfer_stop(sc->pxfer[0]);
1373 		 *
1374 		 * That way, if your code has many parts that will not
1375 		 * stop running under the same lock, in other words
1376 		 * "xfer_mtx", the usbd_transfer_start and
1377 		 * usbd_transfer_stop functions will simply return
1378 		 * when they detect a NULL pointer argument.
1379 		 *
1380 		 * To avoid any races we clear the "pxfer[]" pointer
1381 		 * while holding the private mutex of the driver:
1382 		 */
1383 		pxfer[n_setup] = NULL;
1384 
1385 		USB_BUS_UNLOCK(info->bus);
1386 		USB_XFER_UNLOCK(xfer);
1387 
1388 		usbd_transfer_drain(xfer);
1389 
1390 #if USB_HAVE_BUSDMA
1391 		if (xfer->flags_int.bdma_enable)
1392 			needs_delay = 1;
1393 #endif
1394 		/*
1395 		 * NOTE: default endpoint does not have an
1396 		 * interface, even if endpoint->iface_index == 0
1397 		 */
1398 		USB_BUS_LOCK(info->bus);
1399 		xfer->endpoint->refcount_alloc--;
1400 		USB_BUS_UNLOCK(info->bus);
1401 
1402 		usb_callout_drain(&xfer->timeout_handle);
1403 
1404 		USB_BUS_LOCK(info->bus);
1405 
1406 		USB_ASSERT(info->setup_refcount != 0, ("Invalid setup "
1407 		    "reference count\n"));
1408 
1409 		info->setup_refcount--;
1410 
1411 		if (info->setup_refcount == 0) {
1412 			usbd_transfer_unsetup_sub(info,
1413 			    needs_delay);
1414 		} else {
1415 			USB_BUS_UNLOCK(info->bus);
1416 		}
1417 	}
1418 }
1419 
1420 /*------------------------------------------------------------------------*
1421  *	usbd_control_transfer_init - factored out code
1422  *
1423  * In USB Device Mode we have to wait for the SETUP packet which
1424  * containst the "struct usb_device_request" structure, before we can
1425  * transfer any data. In USB Host Mode we already have the SETUP
1426  * packet at the moment the USB transfer is started. This leads us to
1427  * having to setup the USB transfer at two different places in
1428  * time. This function just contains factored out control transfer
1429  * initialisation code, so that we don't duplicate the code.
1430  *------------------------------------------------------------------------*/
1431 static void
1432 usbd_control_transfer_init(struct usb_xfer *xfer)
1433 {
1434 	struct usb_device_request req;
1435 
1436 	/* copy out the USB request header */
1437 
1438 	usbd_copy_out(xfer->frbuffers, 0, &req, sizeof(req));
1439 
1440 	/* setup remainder */
1441 
1442 	xfer->flags_int.control_rem = UGETW(req.wLength);
1443 
1444 	/* copy direction to endpoint variable */
1445 
1446 	xfer->endpointno &= ~(UE_DIR_IN | UE_DIR_OUT);
1447 	xfer->endpointno |=
1448 	    (req.bmRequestType & UT_READ) ? UE_DIR_IN : UE_DIR_OUT;
1449 }
1450 
1451 /*------------------------------------------------------------------------*
1452  *	usbd_control_transfer_did_data
1453  *
1454  * This function returns non-zero if a control endpoint has
1455  * transferred the first DATA packet after the SETUP packet.
1456  * Else it returns zero.
1457  *------------------------------------------------------------------------*/
1458 static uint8_t
1459 usbd_control_transfer_did_data(struct usb_xfer *xfer)
1460 {
1461 	struct usb_device_request req;
1462 
1463 	/* SETUP packet is not yet sent */
1464 	if (xfer->flags_int.control_hdr != 0)
1465 		return (0);
1466 
1467 	/* copy out the USB request header */
1468 	usbd_copy_out(xfer->frbuffers, 0, &req, sizeof(req));
1469 
1470 	/* compare remainder to the initial value */
1471 	return (xfer->flags_int.control_rem != UGETW(req.wLength));
1472 }
1473 
1474 /*------------------------------------------------------------------------*
1475  *	usbd_setup_ctrl_transfer
1476  *
1477  * This function handles initialisation of control transfers. Control
1478  * transfers are special in that regard that they can both transmit
1479  * and receive data.
1480  *
1481  * Return values:
1482  *    0: Success
1483  * Else: Failure
1484  *------------------------------------------------------------------------*/
1485 static int
1486 usbd_setup_ctrl_transfer(struct usb_xfer *xfer)
1487 {
1488 	usb_frlength_t len;
1489 
1490 	/* Check for control endpoint stall */
1491 	if (xfer->flags.stall_pipe && xfer->flags_int.control_act) {
1492 		/* the control transfer is no longer active */
1493 		xfer->flags_int.control_stall = 1;
1494 		xfer->flags_int.control_act = 0;
1495 	} else {
1496 		/* don't stall control transfer by default */
1497 		xfer->flags_int.control_stall = 0;
1498 	}
1499 
1500 	/* Check for invalid number of frames */
1501 	if (xfer->nframes > 2) {
1502 		/*
1503 		 * If you need to split a control transfer, you
1504 		 * have to do one part at a time. Only with
1505 		 * non-control transfers you can do multiple
1506 		 * parts a time.
1507 		 */
1508 		DPRINTFN(0, "Too many frames: %u\n",
1509 		    (unsigned int)xfer->nframes);
1510 		goto error;
1511 	}
1512 
1513 	/*
1514          * Check if there is a control
1515          * transfer in progress:
1516          */
1517 	if (xfer->flags_int.control_act) {
1518 
1519 		if (xfer->flags_int.control_hdr) {
1520 
1521 			/* clear send header flag */
1522 
1523 			xfer->flags_int.control_hdr = 0;
1524 
1525 			/* setup control transfer */
1526 			if (xfer->flags_int.usb_mode == USB_MODE_DEVICE) {
1527 				usbd_control_transfer_init(xfer);
1528 			}
1529 		}
1530 		/* get data length */
1531 
1532 		len = xfer->sumlen;
1533 
1534 	} else {
1535 
1536 		/* the size of the SETUP structure is hardcoded ! */
1537 
1538 		if (xfer->frlengths[0] != sizeof(struct usb_device_request)) {
1539 			DPRINTFN(0, "Wrong framelength %u != %zu\n",
1540 			    xfer->frlengths[0], sizeof(struct
1541 			    usb_device_request));
1542 			goto error;
1543 		}
1544 		/* check USB mode */
1545 		if (xfer->flags_int.usb_mode == USB_MODE_DEVICE) {
1546 
1547 			/* check number of frames */
1548 			if (xfer->nframes != 1) {
1549 				/*
1550 			         * We need to receive the setup
1551 			         * message first so that we know the
1552 			         * data direction!
1553 			         */
1554 				DPRINTF("Misconfigured transfer\n");
1555 				goto error;
1556 			}
1557 			/*
1558 			 * Set a dummy "control_rem" value.  This
1559 			 * variable will be overwritten later by a
1560 			 * call to "usbd_control_transfer_init()" !
1561 			 */
1562 			xfer->flags_int.control_rem = 0xFFFF;
1563 		} else {
1564 
1565 			/* setup "endpoint" and "control_rem" */
1566 
1567 			usbd_control_transfer_init(xfer);
1568 		}
1569 
1570 		/* set transfer-header flag */
1571 
1572 		xfer->flags_int.control_hdr = 1;
1573 
1574 		/* get data length */
1575 
1576 		len = (xfer->sumlen - sizeof(struct usb_device_request));
1577 	}
1578 
1579 	/* update did data flag */
1580 
1581 	xfer->flags_int.control_did_data =
1582 	    usbd_control_transfer_did_data(xfer);
1583 
1584 	/* check if there is a length mismatch */
1585 
1586 	if (len > xfer->flags_int.control_rem) {
1587 		DPRINTFN(0, "Length (%d) greater than "
1588 		    "remaining length (%d)\n", len,
1589 		    xfer->flags_int.control_rem);
1590 		goto error;
1591 	}
1592 	/* check if we are doing a short transfer */
1593 
1594 	if (xfer->flags.force_short_xfer) {
1595 		xfer->flags_int.control_rem = 0;
1596 	} else {
1597 		if ((len != xfer->max_data_length) &&
1598 		    (len != xfer->flags_int.control_rem) &&
1599 		    (xfer->nframes != 1)) {
1600 			DPRINTFN(0, "Short control transfer without "
1601 			    "force_short_xfer set\n");
1602 			goto error;
1603 		}
1604 		xfer->flags_int.control_rem -= len;
1605 	}
1606 
1607 	/* the status part is executed when "control_act" is 0 */
1608 
1609 	if ((xfer->flags_int.control_rem > 0) ||
1610 	    (xfer->flags.manual_status)) {
1611 		/* don't execute the STATUS stage yet */
1612 		xfer->flags_int.control_act = 1;
1613 
1614 		/* sanity check */
1615 		if ((!xfer->flags_int.control_hdr) &&
1616 		    (xfer->nframes == 1)) {
1617 			/*
1618 		         * This is not a valid operation!
1619 		         */
1620 			DPRINTFN(0, "Invalid parameter "
1621 			    "combination\n");
1622 			goto error;
1623 		}
1624 	} else {
1625 		/* time to execute the STATUS stage */
1626 		xfer->flags_int.control_act = 0;
1627 	}
1628 	return (0);			/* success */
1629 
1630 error:
1631 	return (1);			/* failure */
1632 }
1633 
1634 /*------------------------------------------------------------------------*
1635  *	usbd_transfer_submit - start USB hardware for the given transfer
1636  *
1637  * This function should only be called from the USB callback.
1638  *------------------------------------------------------------------------*/
1639 void
1640 usbd_transfer_submit(struct usb_xfer *xfer)
1641 {
1642 	struct usb_xfer_root *info;
1643 	struct usb_bus *bus;
1644 	usb_frcount_t x;
1645 
1646 	info = xfer->xroot;
1647 	bus = info->bus;
1648 
1649 	DPRINTF("xfer=%p, endpoint=%p, nframes=%d, dir=%s\n",
1650 	    xfer, xfer->endpoint, xfer->nframes, USB_GET_DATA_ISREAD(xfer) ?
1651 	    "read" : "write");
1652 
1653 #ifdef USB_DEBUG
1654 	if (USB_DEBUG_VAR > 0) {
1655 		USB_BUS_LOCK(bus);
1656 
1657 		usb_dump_endpoint(xfer->endpoint);
1658 
1659 		USB_BUS_UNLOCK(bus);
1660 	}
1661 #endif
1662 
1663 	USB_XFER_LOCK_ASSERT(xfer, MA_OWNED);
1664 	USB_BUS_LOCK_ASSERT(bus, MA_NOTOWNED);
1665 
1666 	/* Only open the USB transfer once! */
1667 	if (!xfer->flags_int.open) {
1668 		xfer->flags_int.open = 1;
1669 
1670 		DPRINTF("open\n");
1671 
1672 		USB_BUS_LOCK(bus);
1673 		(xfer->endpoint->methods->open) (xfer);
1674 		USB_BUS_UNLOCK(bus);
1675 	}
1676 	/* set "transferring" flag */
1677 	xfer->flags_int.transferring = 1;
1678 
1679 #if USB_HAVE_POWERD
1680 	/* increment power reference */
1681 	usbd_transfer_power_ref(xfer, 1);
1682 #endif
1683 	/*
1684 	 * Check if the transfer is waiting on a queue, most
1685 	 * frequently the "done_q":
1686 	 */
1687 	if (xfer->wait_queue) {
1688 		USB_BUS_LOCK(bus);
1689 		usbd_transfer_dequeue(xfer);
1690 		USB_BUS_UNLOCK(bus);
1691 	}
1692 	/* clear "did_dma_delay" flag */
1693 	xfer->flags_int.did_dma_delay = 0;
1694 
1695 	/* clear "did_close" flag */
1696 	xfer->flags_int.did_close = 0;
1697 
1698 #if USB_HAVE_BUSDMA
1699 	/* clear "bdma_setup" flag */
1700 	xfer->flags_int.bdma_setup = 0;
1701 #endif
1702 	/* by default we cannot cancel any USB transfer immediately */
1703 	xfer->flags_int.can_cancel_immed = 0;
1704 
1705 	/* clear lengths and frame counts by default */
1706 	xfer->sumlen = 0;
1707 	xfer->actlen = 0;
1708 	xfer->aframes = 0;
1709 
1710 	/* clear any previous errors */
1711 	xfer->error = 0;
1712 
1713 	/* Check if the device is still alive */
1714 	if (info->udev->state < USB_STATE_POWERED) {
1715 		USB_BUS_LOCK(bus);
1716 		/*
1717 		 * Must return cancelled error code else
1718 		 * device drivers can hang.
1719 		 */
1720 		usbd_transfer_done(xfer, USB_ERR_CANCELLED);
1721 		USB_BUS_UNLOCK(bus);
1722 		return;
1723 	}
1724 
1725 	/* sanity check */
1726 	if (xfer->nframes == 0) {
1727 		if (xfer->flags.stall_pipe) {
1728 			/*
1729 			 * Special case - want to stall without transferring
1730 			 * any data:
1731 			 */
1732 			DPRINTF("xfer=%p nframes=0: stall "
1733 			    "or clear stall!\n", xfer);
1734 			USB_BUS_LOCK(bus);
1735 			xfer->flags_int.can_cancel_immed = 1;
1736 			/* start the transfer */
1737 			usb_command_wrapper(&xfer->endpoint->
1738 			    endpoint_q[xfer->stream_id], xfer);
1739 			USB_BUS_UNLOCK(bus);
1740 			return;
1741 		}
1742 		USB_BUS_LOCK(bus);
1743 		usbd_transfer_done(xfer, USB_ERR_INVAL);
1744 		USB_BUS_UNLOCK(bus);
1745 		return;
1746 	}
1747 	/* compute some variables */
1748 
1749 	for (x = 0; x != xfer->nframes; x++) {
1750 		/* make a copy of the frlenghts[] */
1751 		xfer->frlengths[x + xfer->max_frame_count] = xfer->frlengths[x];
1752 		/* compute total transfer length */
1753 		xfer->sumlen += xfer->frlengths[x];
1754 		if (xfer->sumlen < xfer->frlengths[x]) {
1755 			/* length wrapped around */
1756 			USB_BUS_LOCK(bus);
1757 			usbd_transfer_done(xfer, USB_ERR_INVAL);
1758 			USB_BUS_UNLOCK(bus);
1759 			return;
1760 		}
1761 	}
1762 
1763 	/* clear some internal flags */
1764 
1765 	xfer->flags_int.short_xfer_ok = 0;
1766 	xfer->flags_int.short_frames_ok = 0;
1767 
1768 	/* check if this is a control transfer */
1769 
1770 	if (xfer->flags_int.control_xfr) {
1771 
1772 		if (usbd_setup_ctrl_transfer(xfer)) {
1773 			USB_BUS_LOCK(bus);
1774 			usbd_transfer_done(xfer, USB_ERR_STALLED);
1775 			USB_BUS_UNLOCK(bus);
1776 			return;
1777 		}
1778 	}
1779 	/*
1780 	 * Setup filtered version of some transfer flags,
1781 	 * in case of data read direction
1782 	 */
1783 	if (USB_GET_DATA_ISREAD(xfer)) {
1784 
1785 		if (xfer->flags.short_frames_ok) {
1786 			xfer->flags_int.short_xfer_ok = 1;
1787 			xfer->flags_int.short_frames_ok = 1;
1788 		} else if (xfer->flags.short_xfer_ok) {
1789 			xfer->flags_int.short_xfer_ok = 1;
1790 
1791 			/* check for control transfer */
1792 			if (xfer->flags_int.control_xfr) {
1793 				/*
1794 				 * 1) Control transfers do not support
1795 				 * reception of multiple short USB
1796 				 * frames in host mode and device side
1797 				 * mode, with exception of:
1798 				 *
1799 				 * 2) Due to sometimes buggy device
1800 				 * side firmware we need to do a
1801 				 * STATUS stage in case of short
1802 				 * control transfers in USB host mode.
1803 				 * The STATUS stage then becomes the
1804 				 * "alt_next" to the DATA stage.
1805 				 */
1806 				xfer->flags_int.short_frames_ok = 1;
1807 			}
1808 		}
1809 	}
1810 	/*
1811 	 * Check if BUS-DMA support is enabled and try to load virtual
1812 	 * buffers into DMA, if any:
1813 	 */
1814 #if USB_HAVE_BUSDMA
1815 	if (xfer->flags_int.bdma_enable) {
1816 		/* insert the USB transfer last in the BUS-DMA queue */
1817 		usb_command_wrapper(&xfer->xroot->dma_q, xfer);
1818 		return;
1819 	}
1820 #endif
1821 	/*
1822 	 * Enter the USB transfer into the Host Controller or
1823 	 * Device Controller schedule:
1824 	 */
1825 	usbd_pipe_enter(xfer);
1826 }
1827 
1828 /*------------------------------------------------------------------------*
1829  *	usbd_pipe_enter - factored out code
1830  *------------------------------------------------------------------------*/
1831 void
1832 usbd_pipe_enter(struct usb_xfer *xfer)
1833 {
1834 	struct usb_endpoint *ep;
1835 
1836 	USB_XFER_LOCK_ASSERT(xfer, MA_OWNED);
1837 
1838 	USB_BUS_LOCK(xfer->xroot->bus);
1839 
1840 	ep = xfer->endpoint;
1841 
1842 	DPRINTF("enter\n");
1843 
1844 	/* the transfer can now be cancelled */
1845 	xfer->flags_int.can_cancel_immed = 1;
1846 
1847 	/* enter the transfer */
1848 	(ep->methods->enter) (xfer);
1849 
1850 	/* check for transfer error */
1851 	if (xfer->error) {
1852 		/* some error has happened */
1853 		usbd_transfer_done(xfer, 0);
1854 		USB_BUS_UNLOCK(xfer->xroot->bus);
1855 		return;
1856 	}
1857 
1858 	/* start the transfer */
1859 	usb_command_wrapper(&ep->endpoint_q[xfer->stream_id], xfer);
1860 	USB_BUS_UNLOCK(xfer->xroot->bus);
1861 }
1862 
1863 /*------------------------------------------------------------------------*
1864  *	usbd_transfer_start - start an USB transfer
1865  *
1866  * NOTE: Calling this function more than one time will only
1867  *       result in a single transfer start, until the USB transfer
1868  *       completes.
1869  *------------------------------------------------------------------------*/
1870 void
1871 usbd_transfer_start(struct usb_xfer *xfer)
1872 {
1873 	if (xfer == NULL) {
1874 		/* transfer is gone */
1875 		return;
1876 	}
1877 	USB_XFER_LOCK_ASSERT(xfer, MA_OWNED);
1878 
1879 	/* mark the USB transfer started */
1880 
1881 	if (!xfer->flags_int.started) {
1882 		/* lock the BUS lock to avoid races updating flags_int */
1883 		USB_BUS_LOCK(xfer->xroot->bus);
1884 		xfer->flags_int.started = 1;
1885 		USB_BUS_UNLOCK(xfer->xroot->bus);
1886 	}
1887 	/* check if the USB transfer callback is already transferring */
1888 
1889 	if (xfer->flags_int.transferring) {
1890 		return;
1891 	}
1892 	USB_BUS_LOCK(xfer->xroot->bus);
1893 	/* call the USB transfer callback */
1894 	usbd_callback_ss_done_defer(xfer);
1895 	USB_BUS_UNLOCK(xfer->xroot->bus);
1896 }
1897 
1898 /*------------------------------------------------------------------------*
1899  *	usbd_transfer_stop - stop an USB transfer
1900  *
1901  * NOTE: Calling this function more than one time will only
1902  *       result in a single transfer stop.
1903  * NOTE: When this function returns it is not safe to free nor
1904  *       reuse any DMA buffers. See "usbd_transfer_drain()".
1905  *------------------------------------------------------------------------*/
1906 void
1907 usbd_transfer_stop(struct usb_xfer *xfer)
1908 {
1909 	struct usb_endpoint *ep;
1910 
1911 	if (xfer == NULL) {
1912 		/* transfer is gone */
1913 		return;
1914 	}
1915 	USB_XFER_LOCK_ASSERT(xfer, MA_OWNED);
1916 
1917 	/* check if the USB transfer was ever opened */
1918 
1919 	if (!xfer->flags_int.open) {
1920 		if (xfer->flags_int.started) {
1921 			/* nothing to do except clearing the "started" flag */
1922 			/* lock the BUS lock to avoid races updating flags_int */
1923 			USB_BUS_LOCK(xfer->xroot->bus);
1924 			xfer->flags_int.started = 0;
1925 			USB_BUS_UNLOCK(xfer->xroot->bus);
1926 		}
1927 		return;
1928 	}
1929 	/* try to stop the current USB transfer */
1930 
1931 	USB_BUS_LOCK(xfer->xroot->bus);
1932 	/* override any previous error */
1933 	xfer->error = USB_ERR_CANCELLED;
1934 
1935 	/*
1936 	 * Clear "open" and "started" when both private and USB lock
1937 	 * is locked so that we don't get a race updating "flags_int"
1938 	 */
1939 	xfer->flags_int.open = 0;
1940 	xfer->flags_int.started = 0;
1941 
1942 	/*
1943 	 * Check if we can cancel the USB transfer immediately.
1944 	 */
1945 	if (xfer->flags_int.transferring) {
1946 		if (xfer->flags_int.can_cancel_immed &&
1947 		    (!xfer->flags_int.did_close)) {
1948 			DPRINTF("close\n");
1949 			/*
1950 			 * The following will lead to an USB_ERR_CANCELLED
1951 			 * error code being passed to the USB callback.
1952 			 */
1953 			(xfer->endpoint->methods->close) (xfer);
1954 			/* only close once */
1955 			xfer->flags_int.did_close = 1;
1956 		} else {
1957 			/* need to wait for the next done callback */
1958 		}
1959 	} else {
1960 		DPRINTF("close\n");
1961 
1962 		/* close here and now */
1963 		(xfer->endpoint->methods->close) (xfer);
1964 
1965 		/*
1966 		 * Any additional DMA delay is done by
1967 		 * "usbd_transfer_unsetup()".
1968 		 */
1969 
1970 		/*
1971 		 * Special case. Check if we need to restart a blocked
1972 		 * endpoint.
1973 		 */
1974 		ep = xfer->endpoint;
1975 
1976 		/*
1977 		 * If the current USB transfer is completing we need
1978 		 * to start the next one:
1979 		 */
1980 		if (ep->endpoint_q[xfer->stream_id].curr == xfer) {
1981 			usb_command_wrapper(
1982 			    &ep->endpoint_q[xfer->stream_id], NULL);
1983 		}
1984 	}
1985 
1986 	USB_BUS_UNLOCK(xfer->xroot->bus);
1987 }
1988 
1989 /*------------------------------------------------------------------------*
1990  *	usbd_transfer_pending
1991  *
1992  * This function will check if an USB transfer is pending which is a
1993  * little bit complicated!
1994  * Return values:
1995  * 0: Not pending
1996  * 1: Pending: The USB transfer will receive a callback in the future.
1997  *------------------------------------------------------------------------*/
1998 uint8_t
1999 usbd_transfer_pending(struct usb_xfer *xfer)
2000 {
2001 	struct usb_xfer_root *info;
2002 	struct usb_xfer_queue *pq;
2003 
2004 	if (xfer == NULL) {
2005 		/* transfer is gone */
2006 		return (0);
2007 	}
2008 	USB_XFER_LOCK_ASSERT(xfer, MA_OWNED);
2009 
2010 	if (xfer->flags_int.transferring) {
2011 		/* trivial case */
2012 		return (1);
2013 	}
2014 	USB_BUS_LOCK(xfer->xroot->bus);
2015 	if (xfer->wait_queue) {
2016 		/* we are waiting on a queue somewhere */
2017 		USB_BUS_UNLOCK(xfer->xroot->bus);
2018 		return (1);
2019 	}
2020 	info = xfer->xroot;
2021 	pq = &info->done_q;
2022 
2023 	if (pq->curr == xfer) {
2024 		/* we are currently scheduled for callback */
2025 		USB_BUS_UNLOCK(xfer->xroot->bus);
2026 		return (1);
2027 	}
2028 	/* we are not pending */
2029 	USB_BUS_UNLOCK(xfer->xroot->bus);
2030 	return (0);
2031 }
2032 
2033 /*------------------------------------------------------------------------*
2034  *	usbd_transfer_drain
2035  *
2036  * This function will stop the USB transfer and wait for any
2037  * additional BUS-DMA and HW-DMA operations to complete. Buffers that
2038  * are loaded into DMA can safely be freed or reused after that this
2039  * function has returned.
2040  *------------------------------------------------------------------------*/
2041 void
2042 usbd_transfer_drain(struct usb_xfer *xfer)
2043 {
2044 	WITNESS_WARN(WARN_GIANTOK | WARN_SLEEPOK, NULL,
2045 	    "usbd_transfer_drain can sleep!");
2046 
2047 	if (xfer == NULL) {
2048 		/* transfer is gone */
2049 		return;
2050 	}
2051 	if (xfer->xroot->xfer_mtx != &Giant) {
2052 		USB_XFER_LOCK_ASSERT(xfer, MA_NOTOWNED);
2053 	}
2054 	USB_XFER_LOCK(xfer);
2055 
2056 	usbd_transfer_stop(xfer);
2057 
2058 	while (usbd_transfer_pending(xfer) ||
2059 	    xfer->flags_int.doing_callback) {
2060 
2061 		/*
2062 		 * It is allowed that the callback can drop its
2063 		 * transfer mutex. In that case checking only
2064 		 * "usbd_transfer_pending()" is not enough to tell if
2065 		 * the USB transfer is fully drained. We also need to
2066 		 * check the internal "doing_callback" flag.
2067 		 */
2068 		xfer->flags_int.draining = 1;
2069 
2070 		/*
2071 		 * Wait until the current outstanding USB
2072 		 * transfer is complete !
2073 		 */
2074 		cv_wait(&xfer->xroot->cv_drain, xfer->xroot->xfer_mtx);
2075 	}
2076 	USB_XFER_UNLOCK(xfer);
2077 }
2078 
2079 struct usb_page_cache *
2080 usbd_xfer_get_frame(struct usb_xfer *xfer, usb_frcount_t frindex)
2081 {
2082 	KASSERT(frindex < xfer->max_frame_count, ("frame index overflow"));
2083 
2084 	return (&xfer->frbuffers[frindex]);
2085 }
2086 
2087 void *
2088 usbd_xfer_get_frame_buffer(struct usb_xfer *xfer, usb_frcount_t frindex)
2089 {
2090 	struct usb_page_search page_info;
2091 
2092 	KASSERT(frindex < xfer->max_frame_count, ("frame index overflow"));
2093 
2094 	usbd_get_page(&xfer->frbuffers[frindex], 0, &page_info);
2095 	return (page_info.buffer);
2096 }
2097 
2098 /*------------------------------------------------------------------------*
2099  *	usbd_xfer_get_fps_shift
2100  *
2101  * The following function is only useful for isochronous transfers. It
2102  * returns how many times the frame execution rate has been shifted
2103  * down.
2104  *
2105  * Return value:
2106  * Success: 0..3
2107  * Failure: 0
2108  *------------------------------------------------------------------------*/
2109 uint8_t
2110 usbd_xfer_get_fps_shift(struct usb_xfer *xfer)
2111 {
2112 	return (xfer->fps_shift);
2113 }
2114 
2115 usb_frlength_t
2116 usbd_xfer_frame_len(struct usb_xfer *xfer, usb_frcount_t frindex)
2117 {
2118 	KASSERT(frindex < xfer->max_frame_count, ("frame index overflow"));
2119 
2120 	return (xfer->frlengths[frindex]);
2121 }
2122 
2123 /*------------------------------------------------------------------------*
2124  *	usbd_xfer_set_frame_data
2125  *
2126  * This function sets the pointer of the buffer that should
2127  * loaded directly into DMA for the given USB frame. Passing "ptr"
2128  * equal to NULL while the corresponding "frlength" is greater
2129  * than zero gives undefined results!
2130  *------------------------------------------------------------------------*/
2131 void
2132 usbd_xfer_set_frame_data(struct usb_xfer *xfer, usb_frcount_t frindex,
2133     void *ptr, usb_frlength_t len)
2134 {
2135 	KASSERT(frindex < xfer->max_frame_count, ("frame index overflow"));
2136 
2137 	/* set virtual address to load and length */
2138 	xfer->frbuffers[frindex].buffer = ptr;
2139 	usbd_xfer_set_frame_len(xfer, frindex, len);
2140 }
2141 
2142 void
2143 usbd_xfer_frame_data(struct usb_xfer *xfer, usb_frcount_t frindex,
2144     void **ptr, int *len)
2145 {
2146 	KASSERT(frindex < xfer->max_frame_count, ("frame index overflow"));
2147 
2148 	if (ptr != NULL)
2149 		*ptr = xfer->frbuffers[frindex].buffer;
2150 	if (len != NULL)
2151 		*len = xfer->frlengths[frindex];
2152 }
2153 
2154 /*------------------------------------------------------------------------*
2155  *	usbd_xfer_old_frame_length
2156  *
2157  * This function returns the framelength of the given frame at the
2158  * time the transfer was submitted. This function can be used to
2159  * compute the starting data pointer of the next isochronous frame
2160  * when an isochronous transfer has completed.
2161  *------------------------------------------------------------------------*/
2162 usb_frlength_t
2163 usbd_xfer_old_frame_length(struct usb_xfer *xfer, usb_frcount_t frindex)
2164 {
2165 	KASSERT(frindex < xfer->max_frame_count, ("frame index overflow"));
2166 
2167 	return (xfer->frlengths[frindex + xfer->max_frame_count]);
2168 }
2169 
2170 void
2171 usbd_xfer_status(struct usb_xfer *xfer, int *actlen, int *sumlen, int *aframes,
2172     int *nframes)
2173 {
2174 	if (actlen != NULL)
2175 		*actlen = xfer->actlen;
2176 	if (sumlen != NULL)
2177 		*sumlen = xfer->sumlen;
2178 	if (aframes != NULL)
2179 		*aframes = xfer->aframes;
2180 	if (nframes != NULL)
2181 		*nframes = xfer->nframes;
2182 }
2183 
2184 /*------------------------------------------------------------------------*
2185  *	usbd_xfer_set_frame_offset
2186  *
2187  * This function sets the frame data buffer offset relative to the beginning
2188  * of the USB DMA buffer allocated for this USB transfer.
2189  *------------------------------------------------------------------------*/
2190 void
2191 usbd_xfer_set_frame_offset(struct usb_xfer *xfer, usb_frlength_t offset,
2192     usb_frcount_t frindex)
2193 {
2194 	KASSERT(!xfer->flags.ext_buffer, ("Cannot offset data frame "
2195 	    "when the USB buffer is external\n"));
2196 	KASSERT(frindex < xfer->max_frame_count, ("frame index overflow"));
2197 
2198 	/* set virtual address to load */
2199 	xfer->frbuffers[frindex].buffer =
2200 	    USB_ADD_BYTES(xfer->local_buffer, offset);
2201 }
2202 
2203 void
2204 usbd_xfer_set_interval(struct usb_xfer *xfer, int i)
2205 {
2206 	xfer->interval = i;
2207 }
2208 
2209 void
2210 usbd_xfer_set_timeout(struct usb_xfer *xfer, int t)
2211 {
2212 	xfer->timeout = t;
2213 }
2214 
2215 void
2216 usbd_xfer_set_frames(struct usb_xfer *xfer, usb_frcount_t n)
2217 {
2218 	xfer->nframes = n;
2219 }
2220 
2221 usb_frcount_t
2222 usbd_xfer_max_frames(struct usb_xfer *xfer)
2223 {
2224 	return (xfer->max_frame_count);
2225 }
2226 
2227 usb_frlength_t
2228 usbd_xfer_max_len(struct usb_xfer *xfer)
2229 {
2230 	return (xfer->max_data_length);
2231 }
2232 
2233 usb_frlength_t
2234 usbd_xfer_max_framelen(struct usb_xfer *xfer)
2235 {
2236 	return (xfer->max_frame_size);
2237 }
2238 
2239 void
2240 usbd_xfer_set_frame_len(struct usb_xfer *xfer, usb_frcount_t frindex,
2241     usb_frlength_t len)
2242 {
2243 	KASSERT(frindex < xfer->max_frame_count, ("frame index overflow"));
2244 
2245 	xfer->frlengths[frindex] = len;
2246 }
2247 
2248 /*------------------------------------------------------------------------*
2249  *	usb_callback_proc - factored out code
2250  *
2251  * This function performs USB callbacks.
2252  *------------------------------------------------------------------------*/
2253 static void
2254 usb_callback_proc(struct usb_proc_msg *_pm)
2255 {
2256 	struct usb_done_msg *pm = (void *)_pm;
2257 	struct usb_xfer_root *info = pm->xroot;
2258 
2259 	/* Change locking order */
2260 	USB_BUS_UNLOCK(info->bus);
2261 
2262 	/*
2263 	 * We exploit the fact that the mutex is the same for all
2264 	 * callbacks that will be called from this thread:
2265 	 */
2266 	USB_MTX_LOCK(info->xfer_mtx);
2267 	USB_BUS_LOCK(info->bus);
2268 
2269 	/* Continue where we lost track */
2270 	usb_command_wrapper(&info->done_q,
2271 	    info->done_q.curr);
2272 
2273 	USB_MTX_UNLOCK(info->xfer_mtx);
2274 }
2275 
2276 /*------------------------------------------------------------------------*
2277  *	usbd_callback_ss_done_defer
2278  *
2279  * This function will defer the start, stop and done callback to the
2280  * correct thread.
2281  *------------------------------------------------------------------------*/
2282 static void
2283 usbd_callback_ss_done_defer(struct usb_xfer *xfer)
2284 {
2285 	struct usb_xfer_root *info = xfer->xroot;
2286 	struct usb_xfer_queue *pq = &info->done_q;
2287 
2288 	USB_BUS_LOCK_ASSERT(xfer->xroot->bus, MA_OWNED);
2289 
2290 	if (pq->curr != xfer) {
2291 		usbd_transfer_enqueue(pq, xfer);
2292 	}
2293 	if (!pq->recurse_1) {
2294 
2295 		/*
2296 	         * We have to postpone the callback due to the fact we
2297 	         * will have a Lock Order Reversal, LOR, if we try to
2298 	         * proceed !
2299 	         */
2300 		(void) usb_proc_msignal(info->done_p,
2301 		    &info->done_m[0], &info->done_m[1]);
2302 	} else {
2303 		/* clear second recurse flag */
2304 		pq->recurse_2 = 0;
2305 	}
2306 	return;
2307 
2308 }
2309 
2310 /*------------------------------------------------------------------------*
2311  *	usbd_callback_wrapper
2312  *
2313  * This is a wrapper for USB callbacks. This wrapper does some
2314  * auto-magic things like figuring out if we can call the callback
2315  * directly from the current context or if we need to wakeup the
2316  * interrupt process.
2317  *------------------------------------------------------------------------*/
2318 static void
2319 usbd_callback_wrapper(struct usb_xfer_queue *pq)
2320 {
2321 	struct usb_xfer *xfer = pq->curr;
2322 	struct usb_xfer_root *info = xfer->xroot;
2323 
2324 	USB_BUS_LOCK_ASSERT(info->bus, MA_OWNED);
2325 	if ((pq->recurse_3 != 0 || mtx_owned(info->xfer_mtx) == 0) &&
2326 	    USB_IN_POLLING_MODE_FUNC() == 0) {
2327 		/*
2328 	       	 * Cases that end up here:
2329 		 *
2330 		 * 5) HW interrupt done callback or other source.
2331 		 * 6) HW completed transfer during callback
2332 		 */
2333 		DPRINTFN(3, "case 5 and 6\n");
2334 
2335 		/*
2336 	         * We have to postpone the callback due to the fact we
2337 	         * will have a Lock Order Reversal, LOR, if we try to
2338 	         * proceed!
2339 		 *
2340 		 * Postponing the callback also ensures that other USB
2341 		 * transfer queues get a chance.
2342 	         */
2343 		(void) usb_proc_msignal(info->done_p,
2344 		    &info->done_m[0], &info->done_m[1]);
2345 		return;
2346 	}
2347 	/*
2348 	 * Cases that end up here:
2349 	 *
2350 	 * 1) We are starting a transfer
2351 	 * 2) We are prematurely calling back a transfer
2352 	 * 3) We are stopping a transfer
2353 	 * 4) We are doing an ordinary callback
2354 	 */
2355 	DPRINTFN(3, "case 1-4\n");
2356 	/* get next USB transfer in the queue */
2357 	info->done_q.curr = NULL;
2358 
2359 	/* set flag in case of drain */
2360 	xfer->flags_int.doing_callback = 1;
2361 
2362 	USB_BUS_UNLOCK(info->bus);
2363 	USB_BUS_LOCK_ASSERT(info->bus, MA_NOTOWNED);
2364 
2365 	/* set correct USB state for callback */
2366 	if (!xfer->flags_int.transferring) {
2367 		xfer->usb_state = USB_ST_SETUP;
2368 		if (!xfer->flags_int.started) {
2369 			/* we got stopped before we even got started */
2370 			USB_BUS_LOCK(info->bus);
2371 			goto done;
2372 		}
2373 	} else {
2374 
2375 		if (usbd_callback_wrapper_sub(xfer)) {
2376 			/* the callback has been deferred */
2377 			USB_BUS_LOCK(info->bus);
2378 			goto done;
2379 		}
2380 #if USB_HAVE_POWERD
2381 		/* decrement power reference */
2382 		usbd_transfer_power_ref(xfer, -1);
2383 #endif
2384 		xfer->flags_int.transferring = 0;
2385 
2386 		if (xfer->error) {
2387 			xfer->usb_state = USB_ST_ERROR;
2388 		} else {
2389 			/* set transferred state */
2390 			xfer->usb_state = USB_ST_TRANSFERRED;
2391 #if USB_HAVE_BUSDMA
2392 			/* sync DMA memory, if any */
2393 			if (xfer->flags_int.bdma_enable &&
2394 			    (!xfer->flags_int.bdma_no_post_sync)) {
2395 				usb_bdma_post_sync(xfer);
2396 			}
2397 #endif
2398 		}
2399 	}
2400 
2401 #if USB_HAVE_PF
2402 	if (xfer->usb_state != USB_ST_SETUP) {
2403 		USB_BUS_LOCK(info->bus);
2404 		usbpf_xfertap(xfer, USBPF_XFERTAP_DONE);
2405 		USB_BUS_UNLOCK(info->bus);
2406 	}
2407 #endif
2408 	/* call processing routine */
2409 	(xfer->callback) (xfer, xfer->error);
2410 
2411 	/* pickup the USB mutex again */
2412 	USB_BUS_LOCK(info->bus);
2413 
2414 	/*
2415 	 * Check if we got started after that we got cancelled, but
2416 	 * before we managed to do the callback.
2417 	 */
2418 	if ((!xfer->flags_int.open) &&
2419 	    (xfer->flags_int.started) &&
2420 	    (xfer->usb_state == USB_ST_ERROR)) {
2421 		/* clear flag in case of drain */
2422 		xfer->flags_int.doing_callback = 0;
2423 		/* try to loop, but not recursivly */
2424 		usb_command_wrapper(&info->done_q, xfer);
2425 		return;
2426 	}
2427 
2428 done:
2429 	/* clear flag in case of drain */
2430 	xfer->flags_int.doing_callback = 0;
2431 
2432 	/*
2433 	 * Check if we are draining.
2434 	 */
2435 	if (xfer->flags_int.draining &&
2436 	    (!xfer->flags_int.transferring)) {
2437 		/* "usbd_transfer_drain()" is waiting for end of transfer */
2438 		xfer->flags_int.draining = 0;
2439 		cv_broadcast(&info->cv_drain);
2440 	}
2441 
2442 	/* do the next callback, if any */
2443 	usb_command_wrapper(&info->done_q,
2444 	    info->done_q.curr);
2445 }
2446 
2447 /*------------------------------------------------------------------------*
2448  *	usb_dma_delay_done_cb
2449  *
2450  * This function is called when the DMA delay has been exectuded, and
2451  * will make sure that the callback is called to complete the USB
2452  * transfer. This code path is usually only used when there is an USB
2453  * error like USB_ERR_CANCELLED.
2454  *------------------------------------------------------------------------*/
2455 void
2456 usb_dma_delay_done_cb(struct usb_xfer *xfer)
2457 {
2458 	USB_BUS_LOCK_ASSERT(xfer->xroot->bus, MA_OWNED);
2459 
2460 	DPRINTFN(3, "Completed %p\n", xfer);
2461 
2462 	/* queue callback for execution, again */
2463 	usbd_transfer_done(xfer, 0);
2464 }
2465 
2466 /*------------------------------------------------------------------------*
2467  *	usbd_transfer_dequeue
2468  *
2469  *  - This function is used to remove an USB transfer from a USB
2470  *  transfer queue.
2471  *
2472  *  - This function can be called multiple times in a row.
2473  *------------------------------------------------------------------------*/
2474 void
2475 usbd_transfer_dequeue(struct usb_xfer *xfer)
2476 {
2477 	struct usb_xfer_queue *pq;
2478 
2479 	pq = xfer->wait_queue;
2480 	if (pq) {
2481 		TAILQ_REMOVE(&pq->head, xfer, wait_entry);
2482 		xfer->wait_queue = NULL;
2483 	}
2484 }
2485 
2486 /*------------------------------------------------------------------------*
2487  *	usbd_transfer_enqueue
2488  *
2489  *  - This function is used to insert an USB transfer into a USB *
2490  *  transfer queue.
2491  *
2492  *  - This function can be called multiple times in a row.
2493  *------------------------------------------------------------------------*/
2494 void
2495 usbd_transfer_enqueue(struct usb_xfer_queue *pq, struct usb_xfer *xfer)
2496 {
2497 	/*
2498 	 * Insert the USB transfer into the queue, if it is not
2499 	 * already on a USB transfer queue:
2500 	 */
2501 	if (xfer->wait_queue == NULL) {
2502 		xfer->wait_queue = pq;
2503 		TAILQ_INSERT_TAIL(&pq->head, xfer, wait_entry);
2504 	}
2505 }
2506 
2507 /*------------------------------------------------------------------------*
2508  *	usbd_transfer_done
2509  *
2510  *  - This function is used to remove an USB transfer from the busdma,
2511  *  pipe or interrupt queue.
2512  *
2513  *  - This function is used to queue the USB transfer on the done
2514  *  queue.
2515  *
2516  *  - This function is used to stop any USB transfer timeouts.
2517  *------------------------------------------------------------------------*/
2518 void
2519 usbd_transfer_done(struct usb_xfer *xfer, usb_error_t error)
2520 {
2521 	struct usb_xfer_root *info = xfer->xroot;
2522 
2523 	USB_BUS_LOCK_ASSERT(info->bus, MA_OWNED);
2524 
2525 	DPRINTF("err=%s\n", usbd_errstr(error));
2526 
2527 	/*
2528 	 * If we are not transferring then just return.
2529 	 * This can happen during transfer cancel.
2530 	 */
2531 	if (!xfer->flags_int.transferring) {
2532 		DPRINTF("not transferring\n");
2533 		/* end of control transfer, if any */
2534 		xfer->flags_int.control_act = 0;
2535 		return;
2536 	}
2537 	/* only set transfer error, if not already set */
2538 	if (xfer->error == USB_ERR_NORMAL_COMPLETION)
2539 		xfer->error = error;
2540 
2541 	/* stop any callouts */
2542 	usb_callout_stop(&xfer->timeout_handle);
2543 
2544 	/*
2545 	 * If we are waiting on a queue, just remove the USB transfer
2546 	 * from the queue, if any. We should have the required locks
2547 	 * locked to do the remove when this function is called.
2548 	 */
2549 	usbd_transfer_dequeue(xfer);
2550 
2551 #if USB_HAVE_BUSDMA
2552 	if (mtx_owned(info->xfer_mtx)) {
2553 		struct usb_xfer_queue *pq;
2554 
2555 		/*
2556 		 * If the private USB lock is not locked, then we assume
2557 		 * that the BUS-DMA load stage has been passed:
2558 		 */
2559 		pq = &info->dma_q;
2560 
2561 		if (pq->curr == xfer) {
2562 			/* start the next BUS-DMA load, if any */
2563 			usb_command_wrapper(pq, NULL);
2564 		}
2565 	}
2566 #endif
2567 	/* keep some statistics */
2568 	if (xfer->error) {
2569 		info->bus->stats_err.uds_requests
2570 		    [xfer->endpoint->edesc->bmAttributes & UE_XFERTYPE]++;
2571 	} else {
2572 		info->bus->stats_ok.uds_requests
2573 		    [xfer->endpoint->edesc->bmAttributes & UE_XFERTYPE]++;
2574 	}
2575 
2576 	/* call the USB transfer callback */
2577 	usbd_callback_ss_done_defer(xfer);
2578 }
2579 
2580 /*------------------------------------------------------------------------*
2581  *	usbd_transfer_start_cb
2582  *
2583  * This function is called to start the USB transfer when
2584  * "xfer->interval" is greater than zero, and and the endpoint type is
2585  * BULK or CONTROL.
2586  *------------------------------------------------------------------------*/
2587 static void
2588 usbd_transfer_start_cb(void *arg)
2589 {
2590 	struct usb_xfer *xfer = arg;
2591 	struct usb_endpoint *ep = xfer->endpoint;
2592 
2593 	USB_BUS_LOCK_ASSERT(xfer->xroot->bus, MA_OWNED);
2594 
2595 	DPRINTF("start\n");
2596 
2597 #if USB_HAVE_PF
2598 	usbpf_xfertap(xfer, USBPF_XFERTAP_SUBMIT);
2599 #endif
2600 
2601 	/* the transfer can now be cancelled */
2602 	xfer->flags_int.can_cancel_immed = 1;
2603 
2604 	/* start USB transfer, if no error */
2605 	if (xfer->error == 0)
2606 		(ep->methods->start) (xfer);
2607 
2608 	/* check for transfer error */
2609 	if (xfer->error) {
2610 		/* some error has happened */
2611 		usbd_transfer_done(xfer, 0);
2612 	}
2613 }
2614 
2615 /*------------------------------------------------------------------------*
2616  *	usbd_xfer_set_stall
2617  *
2618  * This function is used to set the stall flag outside the
2619  * callback. This function is NULL safe.
2620  *------------------------------------------------------------------------*/
2621 void
2622 usbd_xfer_set_stall(struct usb_xfer *xfer)
2623 {
2624 	if (xfer == NULL) {
2625 		/* tearing down */
2626 		return;
2627 	}
2628 	USB_XFER_LOCK_ASSERT(xfer, MA_OWNED);
2629 
2630 	/* avoid any races by locking the USB mutex */
2631 	USB_BUS_LOCK(xfer->xroot->bus);
2632 	xfer->flags.stall_pipe = 1;
2633 	USB_BUS_UNLOCK(xfer->xroot->bus);
2634 }
2635 
2636 int
2637 usbd_xfer_is_stalled(struct usb_xfer *xfer)
2638 {
2639 	return (xfer->endpoint->is_stalled);
2640 }
2641 
2642 /*------------------------------------------------------------------------*
2643  *	usbd_transfer_clear_stall
2644  *
2645  * This function is used to clear the stall flag outside the
2646  * callback. This function is NULL safe.
2647  *------------------------------------------------------------------------*/
2648 void
2649 usbd_transfer_clear_stall(struct usb_xfer *xfer)
2650 {
2651 	if (xfer == NULL) {
2652 		/* tearing down */
2653 		return;
2654 	}
2655 	USB_XFER_LOCK_ASSERT(xfer, MA_OWNED);
2656 
2657 	/* avoid any races by locking the USB mutex */
2658 	USB_BUS_LOCK(xfer->xroot->bus);
2659 
2660 	xfer->flags.stall_pipe = 0;
2661 
2662 	USB_BUS_UNLOCK(xfer->xroot->bus);
2663 }
2664 
2665 /*------------------------------------------------------------------------*
2666  *	usbd_pipe_start
2667  *
2668  * This function is used to add an USB transfer to the pipe transfer list.
2669  *------------------------------------------------------------------------*/
2670 void
2671 usbd_pipe_start(struct usb_xfer_queue *pq)
2672 {
2673 	struct usb_endpoint *ep;
2674 	struct usb_xfer *xfer;
2675 	uint8_t type;
2676 
2677 	xfer = pq->curr;
2678 	ep = xfer->endpoint;
2679 
2680 	USB_BUS_LOCK_ASSERT(xfer->xroot->bus, MA_OWNED);
2681 
2682 	/*
2683 	 * If the endpoint is already stalled we do nothing !
2684 	 */
2685 	if (ep->is_stalled) {
2686 		return;
2687 	}
2688 	/*
2689 	 * Check if we are supposed to stall the endpoint:
2690 	 */
2691 	if (xfer->flags.stall_pipe) {
2692 		struct usb_device *udev;
2693 		struct usb_xfer_root *info;
2694 
2695 		/* clear stall command */
2696 		xfer->flags.stall_pipe = 0;
2697 
2698 		/* get pointer to USB device */
2699 		info = xfer->xroot;
2700 		udev = info->udev;
2701 
2702 		/*
2703 		 * Only stall BULK and INTERRUPT endpoints.
2704 		 */
2705 		type = (ep->edesc->bmAttributes & UE_XFERTYPE);
2706 		if ((type == UE_BULK) ||
2707 		    (type == UE_INTERRUPT)) {
2708 			uint8_t did_stall;
2709 
2710 			did_stall = 1;
2711 
2712 			if (udev->flags.usb_mode == USB_MODE_DEVICE) {
2713 				(udev->bus->methods->set_stall) (
2714 				    udev, ep, &did_stall);
2715 			} else if (udev->ctrl_xfer[1]) {
2716 				info = udev->ctrl_xfer[1]->xroot;
2717 				usb_proc_msignal(
2718 				    USB_BUS_CS_PROC(info->bus),
2719 				    &udev->cs_msg[0], &udev->cs_msg[1]);
2720 			} else {
2721 				/* should not happen */
2722 				DPRINTFN(0, "No stall handler\n");
2723 			}
2724 			/*
2725 			 * Check if we should stall. Some USB hardware
2726 			 * handles set- and clear-stall in hardware.
2727 			 */
2728 			if (did_stall) {
2729 				/*
2730 				 * The transfer will be continued when
2731 				 * the clear-stall control endpoint
2732 				 * message is received.
2733 				 */
2734 				ep->is_stalled = 1;
2735 				return;
2736 			}
2737 		} else if (type == UE_ISOCHRONOUS) {
2738 
2739 			/*
2740 			 * Make sure any FIFO overflow or other FIFO
2741 			 * error conditions go away by resetting the
2742 			 * endpoint FIFO through the clear stall
2743 			 * method.
2744 			 */
2745 			if (udev->flags.usb_mode == USB_MODE_DEVICE) {
2746 				(udev->bus->methods->clear_stall) (udev, ep);
2747 			}
2748 		}
2749 	}
2750 	/* Set or clear stall complete - special case */
2751 	if (xfer->nframes == 0) {
2752 		/* we are complete */
2753 		xfer->aframes = 0;
2754 		usbd_transfer_done(xfer, 0);
2755 		return;
2756 	}
2757 	/*
2758 	 * Handled cases:
2759 	 *
2760 	 * 1) Start the first transfer queued.
2761 	 *
2762 	 * 2) Re-start the current USB transfer.
2763 	 */
2764 	/*
2765 	 * Check if there should be any
2766 	 * pre transfer start delay:
2767 	 */
2768 	if (xfer->interval > 0) {
2769 		type = (ep->edesc->bmAttributes & UE_XFERTYPE);
2770 		if ((type == UE_BULK) ||
2771 		    (type == UE_CONTROL)) {
2772 			usbd_transfer_timeout_ms(xfer,
2773 			    &usbd_transfer_start_cb,
2774 			    xfer->interval);
2775 			return;
2776 		}
2777 	}
2778 	DPRINTF("start\n");
2779 
2780 #if USB_HAVE_PF
2781 	usbpf_xfertap(xfer, USBPF_XFERTAP_SUBMIT);
2782 #endif
2783 	/* the transfer can now be cancelled */
2784 	xfer->flags_int.can_cancel_immed = 1;
2785 
2786 	/* start USB transfer, if no error */
2787 	if (xfer->error == 0)
2788 		(ep->methods->start) (xfer);
2789 
2790 	/* check for transfer error */
2791 	if (xfer->error) {
2792 		/* some error has happened */
2793 		usbd_transfer_done(xfer, 0);
2794 	}
2795 }
2796 
2797 /*------------------------------------------------------------------------*
2798  *	usbd_transfer_timeout_ms
2799  *
2800  * This function is used to setup a timeout on the given USB
2801  * transfer. If the timeout has been deferred the callback given by
2802  * "cb" will get called after "ms" milliseconds.
2803  *------------------------------------------------------------------------*/
2804 void
2805 usbd_transfer_timeout_ms(struct usb_xfer *xfer,
2806     void (*cb) (void *arg), usb_timeout_t ms)
2807 {
2808 	USB_BUS_LOCK_ASSERT(xfer->xroot->bus, MA_OWNED);
2809 
2810 	/* defer delay */
2811 	usb_callout_reset(&xfer->timeout_handle,
2812 	    USB_MS_TO_TICKS(ms) + USB_CALLOUT_ZERO_TICKS, cb, xfer);
2813 }
2814 
2815 /*------------------------------------------------------------------------*
2816  *	usbd_callback_wrapper_sub
2817  *
2818  *  - This function will update variables in an USB transfer after
2819  *  that the USB transfer is complete.
2820  *
2821  *  - This function is used to start the next USB transfer on the
2822  *  ep transfer queue, if any.
2823  *
2824  * NOTE: In some special cases the USB transfer will not be removed from
2825  * the pipe queue, but remain first. To enforce USB transfer removal call
2826  * this function passing the error code "USB_ERR_CANCELLED".
2827  *
2828  * Return values:
2829  * 0: Success.
2830  * Else: The callback has been deferred.
2831  *------------------------------------------------------------------------*/
2832 static uint8_t
2833 usbd_callback_wrapper_sub(struct usb_xfer *xfer)
2834 {
2835 	struct usb_endpoint *ep;
2836 	struct usb_bus *bus;
2837 	usb_frcount_t x;
2838 
2839 	bus = xfer->xroot->bus;
2840 
2841 	if ((!xfer->flags_int.open) &&
2842 	    (!xfer->flags_int.did_close)) {
2843 		DPRINTF("close\n");
2844 		USB_BUS_LOCK(bus);
2845 		(xfer->endpoint->methods->close) (xfer);
2846 		USB_BUS_UNLOCK(bus);
2847 		/* only close once */
2848 		xfer->flags_int.did_close = 1;
2849 		return (1);		/* wait for new callback */
2850 	}
2851 	/*
2852 	 * If we have a non-hardware induced error we
2853 	 * need to do the DMA delay!
2854 	 */
2855 	if (xfer->error != 0 && !xfer->flags_int.did_dma_delay &&
2856 	    (xfer->error == USB_ERR_CANCELLED ||
2857 	    xfer->error == USB_ERR_TIMEOUT ||
2858 	    bus->methods->start_dma_delay != NULL)) {
2859 
2860 		usb_timeout_t temp;
2861 
2862 		/* only delay once */
2863 		xfer->flags_int.did_dma_delay = 1;
2864 
2865 		/* we can not cancel this delay */
2866 		xfer->flags_int.can_cancel_immed = 0;
2867 
2868 		temp = usbd_get_dma_delay(xfer->xroot->udev);
2869 
2870 		DPRINTFN(3, "DMA delay, %u ms, "
2871 		    "on %p\n", temp, xfer);
2872 
2873 		if (temp != 0) {
2874 			USB_BUS_LOCK(bus);
2875 			/*
2876 			 * Some hardware solutions have dedicated
2877 			 * events when it is safe to free DMA'ed
2878 			 * memory. For the other hardware platforms we
2879 			 * use a static delay.
2880 			 */
2881 			if (bus->methods->start_dma_delay != NULL) {
2882 				(bus->methods->start_dma_delay) (xfer);
2883 			} else {
2884 				usbd_transfer_timeout_ms(xfer,
2885 				    (void (*)(void *))&usb_dma_delay_done_cb,
2886 				    temp);
2887 			}
2888 			USB_BUS_UNLOCK(bus);
2889 			return (1);	/* wait for new callback */
2890 		}
2891 	}
2892 	/* check actual number of frames */
2893 	if (xfer->aframes > xfer->nframes) {
2894 		if (xfer->error == 0) {
2895 			panic("%s: actual number of frames, %d, is "
2896 			    "greater than initial number of frames, %d\n",
2897 			    __FUNCTION__, xfer->aframes, xfer->nframes);
2898 		} else {
2899 			/* just set some valid value */
2900 			xfer->aframes = xfer->nframes;
2901 		}
2902 	}
2903 	/* compute actual length */
2904 	xfer->actlen = 0;
2905 
2906 	for (x = 0; x != xfer->aframes; x++) {
2907 		xfer->actlen += xfer->frlengths[x];
2908 	}
2909 
2910 	/*
2911 	 * Frames that were not transferred get zero actual length in
2912 	 * case the USB device driver does not check the actual number
2913 	 * of frames transferred, "xfer->aframes":
2914 	 */
2915 	for (; x < xfer->nframes; x++) {
2916 		usbd_xfer_set_frame_len(xfer, x, 0);
2917 	}
2918 
2919 	/* check actual length */
2920 	if (xfer->actlen > xfer->sumlen) {
2921 		if (xfer->error == 0) {
2922 			panic("%s: actual length, %d, is greater than "
2923 			    "initial length, %d\n",
2924 			    __FUNCTION__, xfer->actlen, xfer->sumlen);
2925 		} else {
2926 			/* just set some valid value */
2927 			xfer->actlen = xfer->sumlen;
2928 		}
2929 	}
2930 	DPRINTFN(1, "xfer=%p endpoint=%p sts=%d alen=%d, slen=%d, afrm=%d, nfrm=%d\n",
2931 	    xfer, xfer->endpoint, xfer->error, xfer->actlen, xfer->sumlen,
2932 	    xfer->aframes, xfer->nframes);
2933 
2934 	if (xfer->error) {
2935 		/* end of control transfer, if any */
2936 		xfer->flags_int.control_act = 0;
2937 
2938 #if USB_HAVE_TT_SUPPORT
2939 		switch (xfer->error) {
2940 		case USB_ERR_NORMAL_COMPLETION:
2941 		case USB_ERR_SHORT_XFER:
2942 		case USB_ERR_STALLED:
2943 		case USB_ERR_CANCELLED:
2944 			/* nothing to do */
2945 			break;
2946 		default:
2947 			/* try to reset the TT, if any */
2948 			USB_BUS_LOCK(bus);
2949 			uhub_tt_buffer_reset_async_locked(xfer->xroot->udev, xfer->endpoint);
2950 			USB_BUS_UNLOCK(bus);
2951 			break;
2952 		}
2953 #endif
2954 		/* check if we should block the execution queue */
2955 		if ((xfer->error != USB_ERR_CANCELLED) &&
2956 		    (xfer->flags.pipe_bof)) {
2957 			DPRINTFN(2, "xfer=%p: Block On Failure "
2958 			    "on endpoint=%p\n", xfer, xfer->endpoint);
2959 			goto done;
2960 		}
2961 	} else {
2962 		/* check for short transfers */
2963 		if (xfer->actlen < xfer->sumlen) {
2964 
2965 			/* end of control transfer, if any */
2966 			xfer->flags_int.control_act = 0;
2967 
2968 			if (!xfer->flags_int.short_xfer_ok) {
2969 				xfer->error = USB_ERR_SHORT_XFER;
2970 				if (xfer->flags.pipe_bof) {
2971 					DPRINTFN(2, "xfer=%p: Block On Failure on "
2972 					    "Short Transfer on endpoint %p.\n",
2973 					    xfer, xfer->endpoint);
2974 					goto done;
2975 				}
2976 			}
2977 		} else {
2978 			/*
2979 			 * Check if we are in the middle of a
2980 			 * control transfer:
2981 			 */
2982 			if (xfer->flags_int.control_act) {
2983 				DPRINTFN(5, "xfer=%p: Control transfer "
2984 				    "active on endpoint=%p\n", xfer, xfer->endpoint);
2985 				goto done;
2986 			}
2987 		}
2988 	}
2989 
2990 	ep = xfer->endpoint;
2991 
2992 	/*
2993 	 * If the current USB transfer is completing we need to start the
2994 	 * next one:
2995 	 */
2996 	USB_BUS_LOCK(bus);
2997 	if (ep->endpoint_q[xfer->stream_id].curr == xfer) {
2998 		usb_command_wrapper(&ep->endpoint_q[xfer->stream_id], NULL);
2999 
3000 		if (ep->endpoint_q[xfer->stream_id].curr != NULL ||
3001 		    TAILQ_FIRST(&ep->endpoint_q[xfer->stream_id].head) != NULL) {
3002 			/* there is another USB transfer waiting */
3003 		} else {
3004 			/* this is the last USB transfer */
3005 			/* clear isochronous sync flag */
3006 			xfer->endpoint->is_synced = 0;
3007 		}
3008 	}
3009 	USB_BUS_UNLOCK(bus);
3010 done:
3011 	return (0);
3012 }
3013 
3014 /*------------------------------------------------------------------------*
3015  *	usb_command_wrapper
3016  *
3017  * This function is used to execute commands non-recursivly on an USB
3018  * transfer.
3019  *------------------------------------------------------------------------*/
3020 void
3021 usb_command_wrapper(struct usb_xfer_queue *pq, struct usb_xfer *xfer)
3022 {
3023 	if (xfer) {
3024 		/*
3025 		 * If the transfer is not already processing,
3026 		 * queue it!
3027 		 */
3028 		if (pq->curr != xfer) {
3029 			usbd_transfer_enqueue(pq, xfer);
3030 			if (pq->curr != NULL) {
3031 				/* something is already processing */
3032 				DPRINTFN(6, "busy %p\n", pq->curr);
3033 				return;
3034 			}
3035 		}
3036 	} else {
3037 		/* Get next element in queue */
3038 		pq->curr = NULL;
3039 	}
3040 
3041 	if (!pq->recurse_1) {
3042 
3043 		/* clear third recurse flag */
3044 		pq->recurse_3 = 0;
3045 
3046 		do {
3047 			/* set two first recurse flags */
3048 			pq->recurse_1 = 1;
3049 			pq->recurse_2 = 1;
3050 
3051 			if (pq->curr == NULL) {
3052 				xfer = TAILQ_FIRST(&pq->head);
3053 				if (xfer) {
3054 					TAILQ_REMOVE(&pq->head, xfer,
3055 					    wait_entry);
3056 					xfer->wait_queue = NULL;
3057 					pq->curr = xfer;
3058 				} else {
3059 					break;
3060 				}
3061 			}
3062 			DPRINTFN(6, "cb %p (enter)\n", pq->curr);
3063 			(pq->command) (pq);
3064 			DPRINTFN(6, "cb %p (leave)\n", pq->curr);
3065 
3066 			/*
3067 			 * Set third recurse flag to indicate
3068 			 * recursion happened:
3069 			 */
3070 			pq->recurse_3 = 1;
3071 
3072 		} while (!pq->recurse_2);
3073 
3074 		/* clear first recurse flag */
3075 		pq->recurse_1 = 0;
3076 
3077 	} else {
3078 		/* clear second recurse flag */
3079 		pq->recurse_2 = 0;
3080 	}
3081 }
3082 
3083 /*------------------------------------------------------------------------*
3084  *	usbd_ctrl_transfer_setup
3085  *
3086  * This function is used to setup the default USB control endpoint
3087  * transfer.
3088  *------------------------------------------------------------------------*/
3089 void
3090 usbd_ctrl_transfer_setup(struct usb_device *udev)
3091 {
3092 	struct usb_xfer *xfer;
3093 	uint8_t no_resetup;
3094 	uint8_t iface_index;
3095 
3096 	/* check for root HUB */
3097 	if (udev->parent_hub == NULL)
3098 		return;
3099 repeat:
3100 
3101 	xfer = udev->ctrl_xfer[0];
3102 	if (xfer) {
3103 		USB_XFER_LOCK(xfer);
3104 		no_resetup =
3105 		    ((xfer->address == udev->address) &&
3106 		    (udev->ctrl_ep_desc.wMaxPacketSize[0] ==
3107 		    udev->ddesc.bMaxPacketSize));
3108 		if (udev->flags.usb_mode == USB_MODE_DEVICE) {
3109 			if (no_resetup) {
3110 				/*
3111 				 * NOTE: checking "xfer->address" and
3112 				 * starting the USB transfer must be
3113 				 * atomic!
3114 				 */
3115 				usbd_transfer_start(xfer);
3116 			}
3117 		}
3118 		USB_XFER_UNLOCK(xfer);
3119 	} else {
3120 		no_resetup = 0;
3121 	}
3122 
3123 	if (no_resetup) {
3124 		/*
3125 	         * All parameters are exactly the same like before.
3126 	         * Just return.
3127 	         */
3128 		return;
3129 	}
3130 	/*
3131 	 * Update wMaxPacketSize for the default control endpoint:
3132 	 */
3133 	udev->ctrl_ep_desc.wMaxPacketSize[0] =
3134 	    udev->ddesc.bMaxPacketSize;
3135 
3136 	/*
3137 	 * Unsetup any existing USB transfer:
3138 	 */
3139 	usbd_transfer_unsetup(udev->ctrl_xfer, USB_CTRL_XFER_MAX);
3140 
3141 	/*
3142 	 * Reset clear stall error counter.
3143 	 */
3144 	udev->clear_stall_errors = 0;
3145 
3146 	/*
3147 	 * Try to setup a new USB transfer for the
3148 	 * default control endpoint:
3149 	 */
3150 	iface_index = 0;
3151 	if (usbd_transfer_setup(udev, &iface_index,
3152 	    udev->ctrl_xfer, usb_control_ep_cfg, USB_CTRL_XFER_MAX, NULL,
3153 	    &udev->device_mtx)) {
3154 		DPRINTFN(0, "could not setup default "
3155 		    "USB transfer\n");
3156 	} else {
3157 		goto repeat;
3158 	}
3159 }
3160 
3161 /*------------------------------------------------------------------------*
3162  *	usbd_clear_data_toggle - factored out code
3163  *
3164  * NOTE: the intention of this function is not to reset the hardware
3165  * data toggle.
3166  *------------------------------------------------------------------------*/
3167 void
3168 usbd_clear_stall_locked(struct usb_device *udev, struct usb_endpoint *ep)
3169 {
3170 	USB_BUS_LOCK_ASSERT(udev->bus, MA_OWNED);
3171 
3172 	/* check that we have a valid case */
3173 	if (udev->flags.usb_mode == USB_MODE_HOST &&
3174 	    udev->parent_hub != NULL &&
3175 	    udev->bus->methods->clear_stall != NULL &&
3176 	    ep->methods != NULL) {
3177 		(udev->bus->methods->clear_stall) (udev, ep);
3178 	}
3179 }
3180 
3181 /*------------------------------------------------------------------------*
3182  *	usbd_clear_data_toggle - factored out code
3183  *
3184  * NOTE: the intention of this function is not to reset the hardware
3185  * data toggle on the USB device side.
3186  *------------------------------------------------------------------------*/
3187 void
3188 usbd_clear_data_toggle(struct usb_device *udev, struct usb_endpoint *ep)
3189 {
3190 	DPRINTFN(5, "udev=%p endpoint=%p\n", udev, ep);
3191 
3192 	USB_BUS_LOCK(udev->bus);
3193 	ep->toggle_next = 0;
3194 	/* some hardware needs a callback to clear the data toggle */
3195 	usbd_clear_stall_locked(udev, ep);
3196 	USB_BUS_UNLOCK(udev->bus);
3197 }
3198 
3199 /*------------------------------------------------------------------------*
3200  *	usbd_clear_stall_callback - factored out clear stall callback
3201  *
3202  * Input parameters:
3203  *  xfer1: Clear Stall Control Transfer
3204  *  xfer2: Stalled USB Transfer
3205  *
3206  * This function is NULL safe.
3207  *
3208  * Return values:
3209  *   0: In progress
3210  *   Else: Finished
3211  *
3212  * Clear stall config example:
3213  *
3214  * static const struct usb_config my_clearstall =  {
3215  *	.type = UE_CONTROL,
3216  *	.endpoint = 0,
3217  *	.direction = UE_DIR_ANY,
3218  *	.interval = 50, //50 milliseconds
3219  *	.bufsize = sizeof(struct usb_device_request),
3220  *	.timeout = 1000, //1.000 seconds
3221  *	.callback = &my_clear_stall_callback, // **
3222  *	.usb_mode = USB_MODE_HOST,
3223  * };
3224  *
3225  * ** "my_clear_stall_callback" calls "usbd_clear_stall_callback"
3226  * passing the correct parameters.
3227  *------------------------------------------------------------------------*/
3228 uint8_t
3229 usbd_clear_stall_callback(struct usb_xfer *xfer1,
3230     struct usb_xfer *xfer2)
3231 {
3232 	struct usb_device_request req;
3233 
3234 	if (xfer2 == NULL) {
3235 		/* looks like we are tearing down */
3236 		DPRINTF("NULL input parameter\n");
3237 		return (0);
3238 	}
3239 	USB_XFER_LOCK_ASSERT(xfer1, MA_OWNED);
3240 	USB_XFER_LOCK_ASSERT(xfer2, MA_OWNED);
3241 
3242 	switch (USB_GET_STATE(xfer1)) {
3243 	case USB_ST_SETUP:
3244 
3245 		/*
3246 		 * pre-clear the data toggle to DATA0 ("umass.c" and
3247 		 * "ata-usb.c" depends on this)
3248 		 */
3249 
3250 		usbd_clear_data_toggle(xfer2->xroot->udev, xfer2->endpoint);
3251 
3252 		/* setup a clear-stall packet */
3253 
3254 		req.bmRequestType = UT_WRITE_ENDPOINT;
3255 		req.bRequest = UR_CLEAR_FEATURE;
3256 		USETW(req.wValue, UF_ENDPOINT_HALT);
3257 		req.wIndex[0] = xfer2->endpoint->edesc->bEndpointAddress;
3258 		req.wIndex[1] = 0;
3259 		USETW(req.wLength, 0);
3260 
3261 		/*
3262 		 * "usbd_transfer_setup_sub()" will ensure that
3263 		 * we have sufficient room in the buffer for
3264 		 * the request structure!
3265 		 */
3266 
3267 		/* copy in the transfer */
3268 
3269 		usbd_copy_in(xfer1->frbuffers, 0, &req, sizeof(req));
3270 
3271 		/* set length */
3272 		xfer1->frlengths[0] = sizeof(req);
3273 		xfer1->nframes = 1;
3274 
3275 		usbd_transfer_submit(xfer1);
3276 		return (0);
3277 
3278 	case USB_ST_TRANSFERRED:
3279 		break;
3280 
3281 	default:			/* Error */
3282 		if (xfer1->error == USB_ERR_CANCELLED) {
3283 			return (0);
3284 		}
3285 		break;
3286 	}
3287 	return (1);			/* Clear Stall Finished */
3288 }
3289 
3290 /*------------------------------------------------------------------------*
3291  *	usbd_transfer_poll
3292  *
3293  * The following function gets called from the USB keyboard driver and
3294  * UMASS when the system has paniced.
3295  *
3296  * NOTE: It is currently not possible to resume normal operation on
3297  * the USB controller which has been polled, due to clearing of the
3298  * "up_dsleep" and "up_msleep" flags.
3299  *------------------------------------------------------------------------*/
3300 void
3301 usbd_transfer_poll(struct usb_xfer **ppxfer, uint16_t max)
3302 {
3303 	struct usb_xfer *xfer;
3304 	struct usb_xfer_root *xroot;
3305 	struct usb_device *udev;
3306 	struct usb_proc_msg *pm;
3307 	struct usb_bus *bus;
3308 	uint16_t n;
3309 	uint16_t drop_bus_spin;
3310 	uint16_t drop_bus;
3311 	uint16_t drop_xfer;
3312 
3313 	for (n = 0; n != max; n++) {
3314 		/* Extra checks to avoid panic */
3315 		xfer = ppxfer[n];
3316 		if (xfer == NULL)
3317 			continue;	/* no USB transfer */
3318 		xroot = xfer->xroot;
3319 		if (xroot == NULL)
3320 			continue;	/* no USB root */
3321 		udev = xroot->udev;
3322 		if (udev == NULL)
3323 			continue;	/* no USB device */
3324 		bus = udev->bus;
3325 		if (bus == NULL)
3326 			continue;	/* no BUS structure */
3327 		if (bus->methods == NULL)
3328 			continue;	/* no BUS methods */
3329 		if (bus->methods->xfer_poll == NULL)
3330 			continue;	/* no poll method */
3331 
3332 		drop_bus_spin = 0;
3333 		drop_bus = 0;
3334 		drop_xfer = 0;
3335 
3336 		if (USB_IN_POLLING_MODE_FUNC() == 0) {
3337 			/* make sure that the BUS spin mutex is not locked */
3338 			while (mtx_owned(&bus->bus_spin_lock)) {
3339 				mtx_unlock_spin(&bus->bus_spin_lock);
3340 				drop_bus_spin++;
3341 			}
3342 
3343 			/* make sure that the BUS mutex is not locked */
3344 			while (mtx_owned(&bus->bus_mtx)) {
3345 				mtx_unlock(&bus->bus_mtx);
3346 				drop_bus++;
3347 			}
3348 
3349 			/* make sure that the transfer mutex is not locked */
3350 			while (mtx_owned(xroot->xfer_mtx)) {
3351 				mtx_unlock(xroot->xfer_mtx);
3352 				drop_xfer++;
3353 			}
3354 		}
3355 
3356 		/* Make sure cv_signal() and cv_broadcast() is not called */
3357 		USB_BUS_CONTROL_XFER_PROC(bus)->up_msleep = 0;
3358 		USB_BUS_EXPLORE_PROC(bus)->up_msleep = 0;
3359 		USB_BUS_GIANT_PROC(bus)->up_msleep = 0;
3360 		USB_BUS_NON_GIANT_ISOC_PROC(bus)->up_msleep = 0;
3361 		USB_BUS_NON_GIANT_BULK_PROC(bus)->up_msleep = 0;
3362 
3363 		/* poll USB hardware */
3364 		(bus->methods->xfer_poll) (bus);
3365 
3366 		USB_BUS_LOCK(xroot->bus);
3367 
3368 		/* check for clear stall */
3369 		if (udev->ctrl_xfer[1] != NULL) {
3370 
3371 			/* poll clear stall start */
3372 			pm = &udev->cs_msg[0].hdr;
3373 			(pm->pm_callback) (pm);
3374 			/* poll clear stall done thread */
3375 			pm = &udev->ctrl_xfer[1]->
3376 			    xroot->done_m[0].hdr;
3377 			(pm->pm_callback) (pm);
3378 		}
3379 
3380 		/* poll done thread */
3381 		pm = &xroot->done_m[0].hdr;
3382 		(pm->pm_callback) (pm);
3383 
3384 		USB_BUS_UNLOCK(xroot->bus);
3385 
3386 		/* restore transfer mutex */
3387 		while (drop_xfer--)
3388 			mtx_lock(xroot->xfer_mtx);
3389 
3390 		/* restore BUS mutex */
3391 		while (drop_bus--)
3392 			mtx_lock(&bus->bus_mtx);
3393 
3394 		/* restore BUS spin mutex */
3395 		while (drop_bus_spin--)
3396 			mtx_lock_spin(&bus->bus_spin_lock);
3397 	}
3398 }
3399 
3400 static void
3401 usbd_get_std_packet_size(struct usb_std_packet_size *ptr,
3402     uint8_t type, enum usb_dev_speed speed)
3403 {
3404 	static const uint16_t intr_range_max[USB_SPEED_MAX] = {
3405 		[USB_SPEED_LOW] = 8,
3406 		[USB_SPEED_FULL] = 64,
3407 		[USB_SPEED_HIGH] = 1024,
3408 		[USB_SPEED_VARIABLE] = 1024,
3409 		[USB_SPEED_SUPER] = 1024,
3410 	};
3411 
3412 	static const uint16_t isoc_range_max[USB_SPEED_MAX] = {
3413 		[USB_SPEED_LOW] = 0,	/* invalid */
3414 		[USB_SPEED_FULL] = 1023,
3415 		[USB_SPEED_HIGH] = 1024,
3416 		[USB_SPEED_VARIABLE] = 3584,
3417 		[USB_SPEED_SUPER] = 1024,
3418 	};
3419 
3420 	static const uint16_t control_min[USB_SPEED_MAX] = {
3421 		[USB_SPEED_LOW] = 8,
3422 		[USB_SPEED_FULL] = 8,
3423 		[USB_SPEED_HIGH] = 64,
3424 		[USB_SPEED_VARIABLE] = 512,
3425 		[USB_SPEED_SUPER] = 512,
3426 	};
3427 
3428 	static const uint16_t bulk_min[USB_SPEED_MAX] = {
3429 		[USB_SPEED_LOW] = 8,
3430 		[USB_SPEED_FULL] = 8,
3431 		[USB_SPEED_HIGH] = 512,
3432 		[USB_SPEED_VARIABLE] = 512,
3433 		[USB_SPEED_SUPER] = 1024,
3434 	};
3435 
3436 	uint16_t temp;
3437 
3438 	memset(ptr, 0, sizeof(*ptr));
3439 
3440 	switch (type) {
3441 	case UE_INTERRUPT:
3442 		ptr->range.max = intr_range_max[speed];
3443 		break;
3444 	case UE_ISOCHRONOUS:
3445 		ptr->range.max = isoc_range_max[speed];
3446 		break;
3447 	default:
3448 		if (type == UE_BULK)
3449 			temp = bulk_min[speed];
3450 		else /* UE_CONTROL */
3451 			temp = control_min[speed];
3452 
3453 		/* default is fixed */
3454 		ptr->fixed[0] = temp;
3455 		ptr->fixed[1] = temp;
3456 		ptr->fixed[2] = temp;
3457 		ptr->fixed[3] = temp;
3458 
3459 		if (speed == USB_SPEED_FULL) {
3460 			/* multiple sizes */
3461 			ptr->fixed[1] = 16;
3462 			ptr->fixed[2] = 32;
3463 			ptr->fixed[3] = 64;
3464 		}
3465 		if ((speed == USB_SPEED_VARIABLE) &&
3466 		    (type == UE_BULK)) {
3467 			/* multiple sizes */
3468 			ptr->fixed[2] = 1024;
3469 			ptr->fixed[3] = 1536;
3470 		}
3471 		break;
3472 	}
3473 }
3474 
3475 void	*
3476 usbd_xfer_softc(struct usb_xfer *xfer)
3477 {
3478 	return (xfer->priv_sc);
3479 }
3480 
3481 void *
3482 usbd_xfer_get_priv(struct usb_xfer *xfer)
3483 {
3484 	return (xfer->priv_fifo);
3485 }
3486 
3487 void
3488 usbd_xfer_set_priv(struct usb_xfer *xfer, void *ptr)
3489 {
3490 	xfer->priv_fifo = ptr;
3491 }
3492 
3493 uint8_t
3494 usbd_xfer_state(struct usb_xfer *xfer)
3495 {
3496 	return (xfer->usb_state);
3497 }
3498 
3499 void
3500 usbd_xfer_set_flag(struct usb_xfer *xfer, int flag)
3501 {
3502 	switch (flag) {
3503 		case USB_FORCE_SHORT_XFER:
3504 			xfer->flags.force_short_xfer = 1;
3505 			break;
3506 		case USB_SHORT_XFER_OK:
3507 			xfer->flags.short_xfer_ok = 1;
3508 			break;
3509 		case USB_MULTI_SHORT_OK:
3510 			xfer->flags.short_frames_ok = 1;
3511 			break;
3512 		case USB_MANUAL_STATUS:
3513 			xfer->flags.manual_status = 1;
3514 			break;
3515 	}
3516 }
3517 
3518 void
3519 usbd_xfer_clr_flag(struct usb_xfer *xfer, int flag)
3520 {
3521 	switch (flag) {
3522 		case USB_FORCE_SHORT_XFER:
3523 			xfer->flags.force_short_xfer = 0;
3524 			break;
3525 		case USB_SHORT_XFER_OK:
3526 			xfer->flags.short_xfer_ok = 0;
3527 			break;
3528 		case USB_MULTI_SHORT_OK:
3529 			xfer->flags.short_frames_ok = 0;
3530 			break;
3531 		case USB_MANUAL_STATUS:
3532 			xfer->flags.manual_status = 0;
3533 			break;
3534 	}
3535 }
3536 
3537 /*
3538  * The following function returns in milliseconds when the isochronous
3539  * transfer was completed by the hardware. The returned value wraps
3540  * around 65536 milliseconds.
3541  */
3542 uint16_t
3543 usbd_xfer_get_timestamp(struct usb_xfer *xfer)
3544 {
3545 	return (xfer->isoc_time_complete);
3546 }
3547 
3548 /*
3549  * The following function returns non-zero if the max packet size
3550  * field was clamped to a valid value. Else it returns zero.
3551  */
3552 uint8_t
3553 usbd_xfer_maxp_was_clamped(struct usb_xfer *xfer)
3554 {
3555 	return (xfer->flags_int.maxp_was_clamped);
3556 }
3557