xref: /freebsd/sys/dev/aacraid/aacraid_cam.c (revision 069ac184)
1 /*-
2  * SPDX-License-Identifier: BSD-2-Clause
3  *
4  * Copyright (c) 2002-2010 Adaptec, Inc.
5  * Copyright (c) 2010-2012 PMC-Sierra, Inc.
6  * All rights reserved.
7  *
8  * Redistribution and use in source and binary forms, with or without
9  * modification, are permitted provided that the following conditions
10  * are met:
11  * 1. Redistributions of source code must retain the above copyright
12  *    notice, this list of conditions and the following disclaimer.
13  * 2. Redistributions in binary form must reproduce the above copyright
14  *    notice, this list of conditions and the following disclaimer in the
15  *    documentation and/or other materials provided with the distribution.
16  *
17  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
18  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
19  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
20  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
21  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
22  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
23  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
24  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
25  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
26  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
27  * SUCH DAMAGE.
28  */
29 
30 #include <sys/cdefs.h>
31 /*
32  * CAM front-end for communicating with non-DASD devices
33  */
34 
35 #include "opt_aacraid.h"
36 
37 #include <sys/param.h>
38 #include <sys/systm.h>
39 #include <sys/kernel.h>
40 #include <sys/sysctl.h>
41 #include <sys/lock.h>
42 #include <sys/malloc.h>
43 #include <sys/module.h>
44 #include <sys/mutex.h>
45 
46 #include <cam/cam.h>
47 #include <cam/cam_ccb.h>
48 #include <cam/cam_debug.h>
49 #include <cam/cam_periph.h>
50 #include <cam/cam_sim.h>
51 #include <cam/cam_xpt_sim.h>
52 #include <cam/scsi/scsi_all.h>
53 #include <cam/scsi/scsi_message.h>
54 
55 #include <sys/bus.h>
56 #include <sys/conf.h>
57 #include <sys/disk.h>
58 
59 #include <machine/md_var.h>
60 #include <machine/bus.h>
61 #include <sys/rman.h>
62 
63 #include <vm/vm.h>
64 #include <vm/pmap.h>
65 
66 #include <dev/aacraid/aacraid_reg.h>
67 #include <sys/aac_ioctl.h>
68 #include <dev/aacraid/aacraid_debug.h>
69 #include <dev/aacraid/aacraid_var.h>
70 #include <dev/aacraid/aacraid_endian.h>
71 
72 #ifndef	CAM_NEW_TRAN_CODE
73 #define	CAM_NEW_TRAN_CODE	1
74 #endif
75 
76 #ifndef SVPD_SUPPORTED_PAGE_LIST
77 struct scsi_vpd_supported_page_list
78 {
79 	u_int8_t device;
80 	u_int8_t page_code;
81 #define	SVPD_SUPPORTED_PAGE_LIST 0x00
82 	u_int8_t reserved;
83 	u_int8_t length;	/* number of VPD entries */
84 #define	SVPD_SUPPORTED_PAGES_SIZE	251
85 	u_int8_t list[SVPD_SUPPORTED_PAGES_SIZE];
86 };
87 #endif
88 
89 /************************** Version Compatibility *************************/
90 #define	aac_sim_alloc				cam_sim_alloc
91 
92 struct aac_cam {
93 	device_t		dev;
94 	struct aac_sim		*inf;
95 	struct cam_sim		*sim;
96 	struct cam_path		*path;
97 };
98 
99 static int aac_cam_probe(device_t dev);
100 static int aac_cam_attach(device_t dev);
101 static int aac_cam_detach(device_t dev);
102 static void aac_cam_action(struct cam_sim *, union ccb *);
103 static void aac_cam_poll(struct cam_sim *);
104 static void aac_cam_complete(struct aac_command *);
105 static void aac_container_complete(struct aac_command *);
106 static void aac_cam_rescan(struct aac_softc *sc, uint32_t channel,
107 	uint32_t target_id);
108 static void aac_set_scsi_error(struct aac_softc *sc, union ccb *ccb,
109 	u_int8_t status, u_int8_t key, u_int8_t asc, u_int8_t ascq);
110 static int aac_load_map_command_sg(struct aac_softc *, struct aac_command *);
111 static u_int64_t aac_eval_blockno(u_int8_t *);
112 static void aac_container_rw_command(struct cam_sim *, union ccb *, u_int8_t *);
113 static void aac_container_special_command(struct cam_sim *, union ccb *,
114 	u_int8_t *);
115 static void aac_passthrough_command(struct cam_sim *, union ccb *);
116 
117 static u_int32_t aac_cam_reset_bus(struct cam_sim *, union ccb *);
118 static u_int32_t aac_cam_abort_ccb(struct cam_sim *, union ccb *);
119 static u_int32_t aac_cam_term_io(struct cam_sim *, union ccb *);
120 
121 static device_method_t	aacraid_pass_methods[] = {
122 	DEVMETHOD(device_probe,		aac_cam_probe),
123 	DEVMETHOD(device_attach,	aac_cam_attach),
124 	DEVMETHOD(device_detach,	aac_cam_detach),
125 	{ 0, 0 }
126 };
127 
128 static driver_t	aacraid_pass_driver = {
129 	"aacraidp",
130 	aacraid_pass_methods,
131 	sizeof(struct aac_cam)
132 };
133 
134 DRIVER_MODULE(aacraidp, aacraid, aacraid_pass_driver, 0, 0);
135 MODULE_DEPEND(aacraidp, cam, 1, 1, 1);
136 
137 MALLOC_DEFINE(M_AACRAIDCAM, "aacraidcam", "AACRAID CAM info");
138 
139 static void
140 aac_set_scsi_error(struct aac_softc *sc, union ccb *ccb, u_int8_t status,
141 	u_int8_t key, u_int8_t asc, u_int8_t ascq)
142 {
143 	struct scsi_sense_data_fixed *sense =
144 		(struct scsi_sense_data_fixed *)&ccb->csio.sense_data;
145 
146 	fwprintf(sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "Error %d!", status);
147 
148 	ccb->ccb_h.status = CAM_SCSI_STATUS_ERROR;
149 	ccb->csio.scsi_status = status;
150 	if (status == SCSI_STATUS_CHECK_COND) {
151 		ccb->ccb_h.status |= CAM_AUTOSNS_VALID;
152 		bzero(&ccb->csio.sense_data, ccb->csio.sense_len);
153 		ccb->csio.sense_data.error_code =
154 			SSD_CURRENT_ERROR | SSD_ERRCODE_VALID;
155 		sense->flags = key;
156 		if (ccb->csio.sense_len >= 14) {
157 			sense->extra_len = 6;
158 			sense->add_sense_code = asc;
159 			sense->add_sense_code_qual = ascq;
160 		}
161 	}
162 }
163 
164 static void
165 aac_cam_rescan(struct aac_softc *sc, uint32_t channel, uint32_t target_id)
166 {
167 	union ccb *ccb;
168 	struct aac_sim *sim;
169 	struct aac_cam *camsc;
170 
171 	if (target_id == AAC_CAM_TARGET_WILDCARD)
172 		target_id = CAM_TARGET_WILDCARD;
173 
174 	TAILQ_FOREACH(sim, &sc->aac_sim_tqh, sim_link) {
175 		camsc = sim->aac_cam;
176 		if (camsc == NULL || camsc->inf == NULL ||
177 		    camsc->inf->BusNumber != channel)
178 			continue;
179 
180 		ccb = xpt_alloc_ccb_nowait();
181 		if (ccb == NULL) {
182 			device_printf(sc->aac_dev,
183 			    "Cannot allocate ccb for bus rescan.\n");
184 			return;
185 		}
186 
187 		if (xpt_create_path(&ccb->ccb_h.path, xpt_periph,
188 		    cam_sim_path(camsc->sim),
189 		    target_id, CAM_LUN_WILDCARD) != CAM_REQ_CMP) {
190 			xpt_free_ccb(ccb);
191 			device_printf(sc->aac_dev,
192 			    "Cannot create path for bus rescan.\n");
193 			return;
194 		}
195 		xpt_rescan(ccb);
196 		break;
197 	}
198 }
199 
200 static void
201 aac_cam_event(struct aac_softc *sc, struct aac_event *event, void *arg)
202 {
203 	union ccb *ccb;
204 	struct aac_cam *camsc;
205 
206 	switch (event->ev_type) {
207 	case AAC_EVENT_CMFREE:
208 		ccb = arg;
209 		camsc = ccb->ccb_h.sim_priv.entries[0].ptr;
210 		free(event, M_AACRAIDCAM);
211 		xpt_release_simq(camsc->sim, 1);
212 		ccb->ccb_h.status = CAM_REQUEUE_REQ;
213 		xpt_done(ccb);
214 		break;
215 	default:
216 		device_printf(sc->aac_dev, "unknown event %d in aac_cam\n",
217 		    event->ev_type);
218 		break;
219 	}
220 
221 	return;
222 }
223 
224 static int
225 aac_cam_probe(device_t dev)
226 {
227 	struct aac_cam *camsc;
228 
229 	camsc = (struct aac_cam *)device_get_softc(dev);
230 	if (!camsc->inf)
231 		return (0);
232 	fwprintf(camsc->inf->aac_sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "");
233 	return (0);
234 }
235 
236 static int
237 aac_cam_detach(device_t dev)
238 {
239 	struct aac_softc *sc;
240 	struct aac_cam *camsc;
241 
242 	camsc = (struct aac_cam *)device_get_softc(dev);
243 	if (!camsc->inf)
244 		return (0);
245 	sc = camsc->inf->aac_sc;
246 	fwprintf(sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "");
247 	camsc->inf->aac_cam = NULL;
248 
249 	mtx_lock(&sc->aac_io_lock);
250 
251 	xpt_async(AC_LOST_DEVICE, camsc->path, NULL);
252 	xpt_free_path(camsc->path);
253 	xpt_bus_deregister(cam_sim_path(camsc->sim));
254 	cam_sim_free(camsc->sim, /*free_devq*/TRUE);
255 
256 	sc->cam_rescan_cb = NULL;
257 
258 	mtx_unlock(&sc->aac_io_lock);
259 
260 	return (0);
261 }
262 
263 /*
264  * Register the driver as a CAM SIM
265  */
266 static int
267 aac_cam_attach(device_t dev)
268 {
269 	struct cam_devq *devq;
270 	struct cam_sim *sim;
271 	struct cam_path *path;
272 	struct aac_cam *camsc;
273 	struct aac_sim *inf;
274 
275 	camsc = (struct aac_cam *)device_get_softc(dev);
276 	inf = (struct aac_sim *)device_get_ivars(dev);
277 	if (!inf)
278 		return (EIO);
279 	fwprintf(inf->aac_sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "");
280 	camsc->inf = inf;
281 	camsc->inf->aac_cam = camsc;
282 
283 	devq = cam_simq_alloc(inf->TargetsPerBus);
284 	if (devq == NULL)
285 		return (EIO);
286 
287 	sim = aac_sim_alloc(aac_cam_action, aac_cam_poll, "aacraidp", camsc,
288 	    device_get_unit(dev), &inf->aac_sc->aac_io_lock, 1, 1, devq);
289 	if (sim == NULL) {
290 		cam_simq_free(devq);
291 		return (EIO);
292 	}
293 
294 	/* Since every bus has it's own sim, every bus 'appears' as bus 0 */
295 	mtx_lock(&inf->aac_sc->aac_io_lock);
296 	if (aac_xpt_bus_register(sim, dev, 0) != CAM_SUCCESS) {
297 		cam_sim_free(sim, TRUE);
298 		mtx_unlock(&inf->aac_sc->aac_io_lock);
299 		return (EIO);
300 	}
301 
302 	if (xpt_create_path(&path, NULL, cam_sim_path(sim),
303 	    CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP) {
304 		xpt_bus_deregister(cam_sim_path(sim));
305 		cam_sim_free(sim, TRUE);
306 		mtx_unlock(&inf->aac_sc->aac_io_lock);
307 		return (EIO);
308 	}
309 
310 	inf->aac_sc->cam_rescan_cb = aac_cam_rescan;
311 	mtx_unlock(&inf->aac_sc->aac_io_lock);
312 
313 	camsc->sim = sim;
314 	camsc->path = path;
315 
316 	return (0);
317 }
318 
319 static u_int64_t
320 aac_eval_blockno(u_int8_t *cmdp)
321 {
322 	u_int64_t blockno;
323 
324 	switch (cmdp[0]) {
325 	case READ_6:
326 	case WRITE_6:
327 		blockno = scsi_3btoul(((struct scsi_rw_6 *)cmdp)->addr);
328 		break;
329 	case READ_10:
330 	case WRITE_10:
331 		blockno = scsi_4btoul(((struct scsi_rw_10 *)cmdp)->addr);
332 		break;
333 	case READ_12:
334 	case WRITE_12:
335 		blockno = scsi_4btoul(((struct scsi_rw_12 *)cmdp)->addr);
336 		break;
337 	case READ_16:
338 	case WRITE_16:
339 		blockno = scsi_8btou64(((struct scsi_rw_16 *)cmdp)->addr);
340 		break;
341 	default:
342 		blockno = 0;
343 		break;
344 	}
345 	return(blockno);
346 }
347 
348 static void
349 aac_container_rw_command(struct cam_sim *sim, union ccb *ccb, u_int8_t *cmdp)
350 {
351 	struct	aac_cam *camsc;
352 	struct	aac_softc *sc;
353 	struct	aac_command *cm;
354 	struct	aac_fib *fib;
355 	u_int64_t blockno;
356 
357 	camsc = (struct aac_cam *)cam_sim_softc(sim);
358 	sc = camsc->inf->aac_sc;
359 	mtx_assert(&sc->aac_io_lock, MA_OWNED);
360 
361 	if (aacraid_alloc_command(sc, &cm)) {
362 		struct aac_event *event;
363 
364 		xpt_freeze_simq(sim, 1);
365 		ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
366 		ccb->ccb_h.sim_priv.entries[0].ptr = camsc;
367 		event = malloc(sizeof(struct aac_event), M_AACRAIDCAM,
368 		    M_NOWAIT | M_ZERO);
369 		if (event == NULL) {
370 			device_printf(sc->aac_dev,
371 			    "Warning, out of memory for event\n");
372 			return;
373 		}
374 		event->ev_callback = aac_cam_event;
375 		event->ev_arg = ccb;
376 		event->ev_type = AAC_EVENT_CMFREE;
377 		aacraid_add_event(sc, event);
378 		return;
379 	}
380 
381 	fib = cm->cm_fib;
382 	switch (ccb->ccb_h.flags & CAM_DIR_MASK) {
383 	case CAM_DIR_IN:
384 		cm->cm_flags |= AAC_CMD_DATAIN;
385 		break;
386 	case CAM_DIR_OUT:
387 		cm->cm_flags |= AAC_CMD_DATAOUT;
388 		break;
389 	case CAM_DIR_NONE:
390 		break;
391 	default:
392 		cm->cm_flags |= AAC_CMD_DATAIN | AAC_CMD_DATAOUT;
393 		break;
394 	}
395 
396 	blockno = aac_eval_blockno(cmdp);
397 
398 	cm->cm_complete = aac_container_complete;
399 	cm->cm_ccb = ccb;
400 	cm->cm_timestamp = time_uptime;
401 	cm->cm_data = (void *)ccb->csio.data_ptr;
402 	cm->cm_datalen = ccb->csio.dxfer_len;
403 
404 	fib->Header.Size = sizeof(struct aac_fib_header);
405 	fib->Header.XferState =
406 		AAC_FIBSTATE_HOSTOWNED   |
407 		AAC_FIBSTATE_INITIALISED |
408 		AAC_FIBSTATE_EMPTY	 |
409 		AAC_FIBSTATE_FROMHOST	 |
410 		AAC_FIBSTATE_REXPECTED   |
411 		AAC_FIBSTATE_NORM	 |
412 		AAC_FIBSTATE_ASYNC	 |
413 		AAC_FIBSTATE_FAST_RESPONSE;
414 
415 	if (sc->flags & AAC_FLAGS_NEW_COMM_TYPE2) {
416 		struct aac_raw_io2 *raw;
417 		/* NOTE: LE conversion handled at aacraid_map_command_sg() */
418 		raw = (struct aac_raw_io2 *)&fib->data[0];
419 		bzero(raw, sizeof(struct aac_raw_io2));
420 		fib->Header.Command = RawIo2;
421 		raw->strtBlkLow = (u_int32_t)blockno;
422 		raw->strtBlkHigh = (u_int32_t)(blockno >> 32);
423 		raw->byteCnt = cm->cm_datalen;
424 		raw->ldNum = ccb->ccb_h.target_id;
425 		fib->Header.Size += sizeof(struct aac_raw_io2);
426 		cm->cm_sgtable = (struct aac_sg_table *)raw->sge;
427 		if (cm->cm_flags & AAC_CMD_DATAIN)
428 			raw->flags = RIO2_IO_TYPE_READ | RIO2_SG_FORMAT_IEEE1212;
429 		else
430 			raw->flags = RIO2_IO_TYPE_WRITE | RIO2_SG_FORMAT_IEEE1212;
431 	} else if (sc->flags & AAC_FLAGS_RAW_IO) {
432 		struct aac_raw_io *raw;
433 		/* NOTE: LE conversion handled at aacraid_map_command_sg() */
434 		raw = (struct aac_raw_io *)&fib->data[0];
435 		bzero(raw, sizeof(struct aac_raw_io));
436 		fib->Header.Command = RawIo;
437 		raw->BlockNumber = blockno;
438 		raw->ByteCount = cm->cm_datalen;
439 		raw->ContainerId = ccb->ccb_h.target_id;
440 		fib->Header.Size += sizeof(struct aac_raw_io);
441 		cm->cm_sgtable = (struct aac_sg_table *)
442 			&raw->SgMapRaw;
443 		if (cm->cm_flags & AAC_CMD_DATAIN)
444 			raw->Flags = 1;
445 	} else if ((sc->flags & AAC_FLAGS_SG_64BIT) == 0) {
446 		fib->Header.Command = ContainerCommand;
447 		if (cm->cm_flags & AAC_CMD_DATAIN) {
448 			struct aac_blockread *br;
449 			br = (struct aac_blockread *)&fib->data[0];
450 			br->Command = VM_CtBlockRead;
451 			br->ContainerId = ccb->ccb_h.target_id;
452 			br->BlockNumber = blockno;
453 			br->ByteCount = cm->cm_datalen;
454 			aac_blockread_tole(br);
455 			fib->Header.Size += sizeof(struct aac_blockread);
456 			cm->cm_sgtable = &br->SgMap;
457 		} else {
458 			struct aac_blockwrite *bw;
459 			bw = (struct aac_blockwrite *)&fib->data[0];
460 			bw->Command = VM_CtBlockWrite;
461 			bw->ContainerId = ccb->ccb_h.target_id;
462 			bw->BlockNumber = blockno;
463 			bw->ByteCount = cm->cm_datalen;
464 			bw->Stable = CUNSTABLE;
465 			aac_blockwrite_tole(bw);
466 			fib->Header.Size += sizeof(struct aac_blockwrite);
467 			cm->cm_sgtable = &bw->SgMap;
468 		}
469 	} else {
470 		fib->Header.Command = ContainerCommand64;
471 		if (cm->cm_flags & AAC_CMD_DATAIN) {
472 			struct aac_blockread64 *br;
473 			br = (struct aac_blockread64 *)&fib->data[0];
474 			br->Command = VM_CtHostRead64;
475 			br->ContainerId = ccb->ccb_h.target_id;
476 			br->SectorCount = cm->cm_datalen/AAC_BLOCK_SIZE;
477 			br->BlockNumber = blockno;
478 			br->Pad = 0;
479 			br->Flags = 0;
480 			aac_blockread64_tole(br);
481 			fib->Header.Size += sizeof(struct aac_blockread64);
482 			cm->cm_sgtable = (struct aac_sg_table *)&br->SgMap64;
483 		} else {
484 			struct aac_blockwrite64 *bw;
485 			bw = (struct aac_blockwrite64 *)&fib->data[0];
486 			bw->Command = VM_CtHostWrite64;
487 			bw->ContainerId = ccb->ccb_h.target_id;
488 			bw->SectorCount = cm->cm_datalen/AAC_BLOCK_SIZE;
489 			bw->BlockNumber = blockno;
490 			bw->Pad = 0;
491 			bw->Flags = 0;
492 			aac_blockwrite64_tole(bw);
493 			fib->Header.Size += sizeof(struct aac_blockwrite64);
494 			cm->cm_sgtable = (struct aac_sg_table *)&bw->SgMap64;
495 		}
496 	}
497 	aac_enqueue_ready(cm);
498 	aacraid_startio(cm->cm_sc);
499 }
500 
501 static void
502 aac_container_special_command(struct cam_sim *sim, union ccb *ccb,
503 	u_int8_t *cmdp)
504 {
505 	struct	aac_cam *camsc;
506 	struct	aac_softc *sc;
507 	struct	aac_container *co;
508 
509 	camsc = (struct aac_cam *)cam_sim_softc(sim);
510 	sc = camsc->inf->aac_sc;
511 	mtx_assert(&sc->aac_io_lock, MA_OWNED);
512 
513 	TAILQ_FOREACH(co, &sc->aac_container_tqh, co_link) {
514 		fwprintf(sc, HBA_FLAGS_DBG_ERROR_B, "found container %d search for %d", co->co_mntobj.ObjectId, ccb->ccb_h.target_id);
515 		if (co->co_mntobj.ObjectId == ccb->ccb_h.target_id)
516 			break;
517 	}
518 	if (co == NULL || ccb->ccb_h.target_lun != 0) {
519 		fwprintf(sc, HBA_FLAGS_DBG_ERROR_B,
520 			"Container not present: cmd 0x%x id %d lun %d len %d",
521 			*cmdp, ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
522 			ccb->csio.dxfer_len);
523 		ccb->ccb_h.status = CAM_DEV_NOT_THERE;
524 		xpt_done(ccb);
525 		return;
526 	}
527 
528 	if (ccb->csio.dxfer_len)
529 		bzero(ccb->csio.data_ptr, ccb->csio.dxfer_len);
530 
531 	switch (*cmdp) {
532 	case INQUIRY:
533 	{
534 		struct scsi_inquiry *inq = (struct scsi_inquiry *)cmdp;
535 
536 		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
537 		"Container INQUIRY id %d lun %d len %d VPD 0x%x Page 0x%x",
538 			ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
539 			ccb->csio.dxfer_len, inq->byte2, inq->page_code);
540 		if (!(inq->byte2 & SI_EVPD)) {
541 			struct scsi_inquiry_data *p =
542 				(struct scsi_inquiry_data *)ccb->csio.data_ptr;
543 			if (inq->page_code != 0) {
544 				aac_set_scsi_error(sc, ccb,
545 					SCSI_STATUS_CHECK_COND,
546 					SSD_KEY_ILLEGAL_REQUEST, 0x24, 0x00);
547 				xpt_done(ccb);
548 				return;
549 			}
550 			p->device = T_DIRECT;
551 			p->version = SCSI_REV_SPC2;
552 			p->response_format = 2;
553 			if (ccb->csio.dxfer_len >= 36) {
554 				p->additional_length = 31;
555 				p->flags = SID_WBus16|SID_Sync|SID_CmdQue;
556 				/* OEM Vendor defines */
557 				strncpy(p->vendor, "Adaptec ", sizeof(p->vendor));
558 				strncpy(p->product, "Array           ",
559 				    sizeof(p->product));
560 				strncpy(p->revision, "V1.0",
561 				    sizeof(p->revision));
562 			}
563 		} else {
564 			if (inq->page_code == SVPD_SUPPORTED_PAGE_LIST) {
565 				struct scsi_vpd_supported_page_list *p =
566 					(struct scsi_vpd_supported_page_list *)
567 					ccb->csio.data_ptr;
568 				p->device = T_DIRECT;
569 				p->page_code = SVPD_SUPPORTED_PAGE_LIST;
570 				p->length = 2;
571 				p->list[0] = SVPD_SUPPORTED_PAGE_LIST;
572 				p->list[1] = SVPD_UNIT_SERIAL_NUMBER;
573 			} else if (inq->page_code == SVPD_UNIT_SERIAL_NUMBER) {
574 				struct scsi_vpd_unit_serial_number *p =
575 					(struct scsi_vpd_unit_serial_number *)
576 					ccb->csio.data_ptr;
577 				p->device = T_DIRECT;
578 				p->page_code = SVPD_UNIT_SERIAL_NUMBER;
579 				p->length = sprintf((char *)p->serial_num,
580 					"%08X%02X", co->co_uid,
581 					ccb->ccb_h.target_id);
582 			} else {
583 				aac_set_scsi_error(sc, ccb,
584 					SCSI_STATUS_CHECK_COND,
585 					SSD_KEY_ILLEGAL_REQUEST, 0x24, 0x00);
586 				xpt_done(ccb);
587 				return;
588 			}
589 		}
590 		ccb->ccb_h.status = CAM_REQ_CMP;
591 		break;
592 	}
593 
594 	case REPORT_LUNS:
595 		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
596 		"Container REPORT_LUNS id %d lun %d len %d",
597 		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
598 		ccb->csio.dxfer_len);
599 		ccb->ccb_h.status = CAM_REQ_CMP;
600 		break;
601 
602 	case START_STOP:
603 	{
604 		struct scsi_start_stop_unit *ss =
605 			(struct scsi_start_stop_unit *)cmdp;
606 		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
607 		"Container START_STOP id %d lun %d len %d",
608 		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
609 		ccb->csio.dxfer_len);
610 		if (sc->aac_support_opt2 & AAC_SUPPORTED_POWER_MANAGEMENT) {
611 			struct aac_command *cm;
612 			struct aac_fib *fib;
613 			struct aac_cnt_config *ccfg;
614 
615 			if (aacraid_alloc_command(sc, &cm)) {
616 				struct aac_event *event;
617 
618 				xpt_freeze_simq(sim, 1);
619 				ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
620 				ccb->ccb_h.sim_priv.entries[0].ptr = camsc;
621 				event = malloc(sizeof(struct aac_event), M_AACRAIDCAM,
622 					M_NOWAIT | M_ZERO);
623 				if (event == NULL) {
624 					device_printf(sc->aac_dev,
625 						"Warning, out of memory for event\n");
626 					return;
627 				}
628 				event->ev_callback = aac_cam_event;
629 				event->ev_arg = ccb;
630 				event->ev_type = AAC_EVENT_CMFREE;
631 				aacraid_add_event(sc, event);
632 				return;
633 			}
634 
635 			fib = cm->cm_fib;
636 			cm->cm_timestamp = time_uptime;
637 			cm->cm_datalen = 0;
638 
639 			fib->Header.Size =
640 				sizeof(struct aac_fib_header) + sizeof(struct aac_cnt_config);
641 			fib->Header.XferState =
642 				AAC_FIBSTATE_HOSTOWNED   |
643 				AAC_FIBSTATE_INITIALISED |
644 				AAC_FIBSTATE_EMPTY	 |
645 				AAC_FIBSTATE_FROMHOST	 |
646 				AAC_FIBSTATE_REXPECTED   |
647 				AAC_FIBSTATE_NORM	 |
648 				AAC_FIBSTATE_ASYNC	 |
649 				AAC_FIBSTATE_FAST_RESPONSE;
650 			fib->Header.Command = ContainerCommand;
651 
652 			/* Start unit */
653 			ccfg = (struct aac_cnt_config *)&fib->data[0];
654 			bzero(ccfg, sizeof (*ccfg) - CT_PACKET_SIZE);
655 			ccfg->Command = VM_ContainerConfig;
656 			ccfg->CTCommand.command = CT_PM_DRIVER_SUPPORT;
657 			ccfg->CTCommand.param[0] = (ss->how & SSS_START ?
658 				AAC_PM_DRIVERSUP_START_UNIT :
659 				AAC_PM_DRIVERSUP_STOP_UNIT);
660 			ccfg->CTCommand.param[1] = co->co_mntobj.ObjectId;
661 			ccfg->CTCommand.param[2] = 0;	/* 1 - immediate */
662 			aac_cnt_config_tole(ccfg);
663 
664 			if (aacraid_wait_command(cm) != 0 ||
665 				le32toh(*(u_int32_t *)&fib->data[0]) != 0) {
666 				printf("Power Management: Error start/stop container %d\n",
667 				co->co_mntobj.ObjectId);
668 			}
669 			aacraid_release_command(cm);
670 		}
671 		ccb->ccb_h.status = CAM_REQ_CMP;
672 		break;
673 	}
674 
675 	case TEST_UNIT_READY:
676 		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
677 		"Container TEST_UNIT_READY id %d lun %d len %d",
678 		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
679 		ccb->csio.dxfer_len);
680 		ccb->ccb_h.status = CAM_REQ_CMP;
681 		break;
682 
683 	case REQUEST_SENSE:
684 		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
685 		"Container REQUEST_SENSE id %d lun %d len %d",
686 		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
687 		ccb->csio.dxfer_len);
688 		ccb->ccb_h.status = CAM_REQ_CMP;
689 		break;
690 
691 	case READ_CAPACITY:
692 	{
693 		struct scsi_read_capacity_data *p =
694 			(struct scsi_read_capacity_data *)ccb->csio.data_ptr;
695 		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
696 		"Container READ_CAPACITY id %d lun %d len %d",
697 		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
698 		ccb->csio.dxfer_len);
699 		scsi_ulto4b(co->co_mntobj.ObjExtension.BlockDevice.BlockSize, p->length);
700 		/* check if greater than 2TB */
701 		if (co->co_mntobj.CapacityHigh) {
702 			if (sc->flags & AAC_FLAGS_LBA_64BIT)
703 				scsi_ulto4b(0xffffffff, p->addr);
704 		} else {
705 			scsi_ulto4b(co->co_mntobj.Capacity-1, p->addr);
706 		}
707 		ccb->ccb_h.status = CAM_REQ_CMP;
708 		break;
709 	}
710 
711 	case SERVICE_ACTION_IN:
712 	{
713 		struct scsi_read_capacity_data_long *p =
714 			(struct scsi_read_capacity_data_long *)
715 			ccb->csio.data_ptr;
716 		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
717 		"Container SERVICE_ACTION_IN id %d lun %d len %d",
718 		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
719 		ccb->csio.dxfer_len);
720 		if (((struct scsi_read_capacity_16 *)cmdp)->service_action !=
721 			SRC16_SERVICE_ACTION) {
722 			aac_set_scsi_error(sc, ccb, SCSI_STATUS_CHECK_COND,
723 				SSD_KEY_ILLEGAL_REQUEST, 0x24, 0x00);
724 			xpt_done(ccb);
725 			return;
726 		}
727 		scsi_ulto4b(co->co_mntobj.ObjExtension.BlockDevice.BlockSize, p->length);
728 		scsi_ulto4b(co->co_mntobj.CapacityHigh, p->addr);
729 		scsi_ulto4b(co->co_mntobj.Capacity-1, &p->addr[4]);
730 
731 		if (ccb->csio.dxfer_len >= 14) {
732 			u_int32_t mapping = co->co_mntobj.ObjExtension.BlockDevice.bdLgclPhysMap;
733 			p->prot_lbppbe = 0;
734 			while (mapping > 1) {
735 				mapping >>= 1;
736 				p->prot_lbppbe++;
737 			}
738 			p->prot_lbppbe &= 0x0f;
739 		}
740 
741 		ccb->ccb_h.status = CAM_REQ_CMP;
742 		break;
743 	}
744 
745 	case MODE_SENSE_6:
746 	{
747 		struct scsi_mode_sense_6 *msp =(struct scsi_mode_sense_6 *)cmdp;
748 		struct ms6_data {
749 			struct scsi_mode_hdr_6 hd;
750 			struct scsi_mode_block_descr bd;
751 			char pages;
752 		} *p = (struct ms6_data *)ccb->csio.data_ptr;
753 		char *pagep;
754 		int return_all_pages = FALSE;
755 
756 		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
757 		"Container MODE_SENSE id %d lun %d len %d page %d",
758 		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
759 		ccb->csio.dxfer_len, msp->page);
760 		p->hd.datalen = sizeof(struct scsi_mode_hdr_6) - 1;
761 		if (co->co_mntobj.ContentState & AAC_FSCS_READONLY)
762 			p->hd.dev_specific = 0x80;	/* WP */
763 		p->hd.dev_specific |= 0x10;	/* DPOFUA */
764 		if (msp->byte2 & SMS_DBD) {
765 			p->hd.block_descr_len = 0;
766 		} else {
767 			p->hd.block_descr_len =
768 				sizeof(struct scsi_mode_block_descr);
769 			p->hd.datalen += p->hd.block_descr_len;
770 			scsi_ulto3b(co->co_mntobj.ObjExtension.BlockDevice.BlockSize, p->bd.block_len);
771 			if (co->co_mntobj.Capacity > 0xffffff ||
772 				co->co_mntobj.CapacityHigh) {
773 				p->bd.num_blocks[0] = 0xff;
774 				p->bd.num_blocks[1] = 0xff;
775 				p->bd.num_blocks[2] = 0xff;
776 			} else {
777 				p->bd.num_blocks[0] = (u_int8_t)
778 					(co->co_mntobj.Capacity >> 16);
779 				p->bd.num_blocks[1] = (u_int8_t)
780 					(co->co_mntobj.Capacity >> 8);
781 				p->bd.num_blocks[2] = (u_int8_t)
782 					(co->co_mntobj.Capacity);
783 			}
784 		}
785 		pagep = &p->pages;
786 		switch (msp->page & SMS_PAGE_CODE) {
787 		case SMS_ALL_PAGES_PAGE:
788 			return_all_pages = TRUE;
789 		case SMS_CONTROL_MODE_PAGE:
790 		{
791 			struct scsi_control_page *cp =
792 				(struct scsi_control_page *)pagep;
793 
794 			if (ccb->csio.dxfer_len <= p->hd.datalen + 8) {
795 				aac_set_scsi_error(sc, ccb,
796 					SCSI_STATUS_CHECK_COND,
797 					SSD_KEY_ILLEGAL_REQUEST, 0x24, 0x00);
798 				xpt_done(ccb);
799 				return;
800 			}
801 			cp->page_code = SMS_CONTROL_MODE_PAGE;
802 			cp->page_length = 6;
803 			p->hd.datalen += 8;
804 			pagep += 8;
805 			if (!return_all_pages)
806 				break;
807 		}
808 		case SMS_VENDOR_SPECIFIC_PAGE:
809 			break;
810 		default:
811 			aac_set_scsi_error(sc, ccb, SCSI_STATUS_CHECK_COND,
812 				SSD_KEY_ILLEGAL_REQUEST, 0x24, 0x00);
813 			xpt_done(ccb);
814 			return;
815 		}
816 		ccb->ccb_h.status = CAM_REQ_CMP;
817 		break;
818 	}
819 
820 	case SYNCHRONIZE_CACHE:
821 		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
822 		"Container SYNCHRONIZE_CACHE id %d lun %d len %d",
823 		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
824 		ccb->csio.dxfer_len);
825 		ccb->ccb_h.status = CAM_REQ_CMP;
826 		break;
827 
828 	default:
829 		fwprintf(sc, HBA_FLAGS_DBG_ERROR_B,
830 		"Container unsupp. cmd 0x%x id %d lun %d len %d",
831 		*cmdp, ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
832 		ccb->csio.dxfer_len);
833 		ccb->ccb_h.status = CAM_REQ_CMP; /*CAM_REQ_INVALID*/
834 		break;
835 	}
836 	xpt_done(ccb);
837 }
838 
839 static void
840 aac_passthrough_command(struct cam_sim *sim, union ccb *ccb)
841 {
842 	struct	aac_cam *camsc;
843 	struct	aac_softc *sc;
844 	struct	aac_command *cm;
845 	struct	aac_fib *fib;
846 	struct	aac_srb *srb;
847 
848 	camsc = (struct aac_cam *)cam_sim_softc(sim);
849 	sc = camsc->inf->aac_sc;
850 	mtx_assert(&sc->aac_io_lock, MA_OWNED);
851 
852 	if (aacraid_alloc_command(sc, &cm)) {
853 		struct aac_event *event;
854 
855 		xpt_freeze_simq(sim, 1);
856 		ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
857 		ccb->ccb_h.sim_priv.entries[0].ptr = camsc;
858 		event = malloc(sizeof(struct aac_event), M_AACRAIDCAM,
859 		    M_NOWAIT | M_ZERO);
860 		if (event == NULL) {
861 			device_printf(sc->aac_dev,
862 			    "Warning, out of memory for event\n");
863 			return;
864 		}
865 		event->ev_callback = aac_cam_event;
866 		event->ev_arg = ccb;
867 		event->ev_type = AAC_EVENT_CMFREE;
868 		aacraid_add_event(sc, event);
869 		return;
870 	}
871 
872 	fib = cm->cm_fib;
873 	switch (ccb->ccb_h.flags & CAM_DIR_MASK) {
874 	case CAM_DIR_IN:
875 		cm->cm_flags |= AAC_CMD_DATAIN;
876 		break;
877 	case CAM_DIR_OUT:
878 		cm->cm_flags |= AAC_CMD_DATAOUT;
879 		break;
880 	case CAM_DIR_NONE:
881 		break;
882 	default:
883 		cm->cm_flags |= AAC_CMD_DATAIN | AAC_CMD_DATAOUT;
884 		break;
885 	}
886 
887 	srb = (struct aac_srb *)&fib->data[0];
888 	srb->function = AAC_SRB_FUNC_EXECUTE_SCSI;
889 	if (cm->cm_flags & (AAC_CMD_DATAIN|AAC_CMD_DATAOUT))
890 		srb->flags = AAC_SRB_FLAGS_UNSPECIFIED_DIRECTION;
891 	if (cm->cm_flags & AAC_CMD_DATAIN)
892 		srb->flags = AAC_SRB_FLAGS_DATA_IN;
893 	else if (cm->cm_flags & AAC_CMD_DATAOUT)
894 		srb->flags = AAC_SRB_FLAGS_DATA_OUT;
895 	else
896 		srb->flags = AAC_SRB_FLAGS_NO_DATA_XFER;
897 
898 	/*
899 	 * Copy the CDB into the SRB.  It's only 6-16 bytes,
900 	 * so a copy is not too expensive.
901 	 */
902 	srb->cdb_len = ccb->csio.cdb_len;
903 	if (ccb->ccb_h.flags & CAM_CDB_POINTER)
904 		bcopy(ccb->csio.cdb_io.cdb_ptr, (u_int8_t *)&srb->cdb[0],
905 			srb->cdb_len);
906 	else
907 		bcopy(ccb->csio.cdb_io.cdb_bytes, (u_int8_t *)&srb->cdb[0],
908 			srb->cdb_len);
909 
910 	/* Set command */
911 	fib->Header.Command = (sc->flags & AAC_FLAGS_SG_64BIT) ?
912 		ScsiPortCommandU64 : ScsiPortCommand;
913 	fib->Header.Size = sizeof(struct aac_fib_header) +
914 			sizeof(struct aac_srb);
915 
916 	/* Map the s/g list */
917 	cm->cm_sgtable = &srb->sg_map;
918 	if ((ccb->ccb_h.flags & CAM_DIR_MASK) != CAM_DIR_NONE) {
919 		/*
920 		 * Arrange things so that the S/G
921 		 * map will get set up automagically
922 		 */
923 		cm->cm_data = (void *)ccb->csio.data_ptr;
924 		cm->cm_datalen = ccb->csio.dxfer_len;
925 		srb->data_len = ccb->csio.dxfer_len;
926 	} else {
927 		cm->cm_data = NULL;
928 		cm->cm_datalen = 0;
929 		srb->data_len = 0;
930 	}
931 
932 	srb->bus = camsc->inf->BusNumber - 1; /* Bus no. rel. to the card */
933 	srb->target = ccb->ccb_h.target_id;
934 	srb->lun = ccb->ccb_h.target_lun;
935 	srb->timeout = ccb->ccb_h.timeout;	/* XXX */
936 	srb->retry_limit = 0;
937 	aac_srb_tole(srb);
938 
939 	cm->cm_complete = aac_cam_complete;
940 	cm->cm_ccb = ccb;
941 	cm->cm_timestamp = time_uptime;
942 
943 	fib->Header.XferState =
944 			AAC_FIBSTATE_HOSTOWNED	|
945 			AAC_FIBSTATE_INITIALISED	|
946 			AAC_FIBSTATE_FROMHOST	|
947 			AAC_FIBSTATE_REXPECTED	|
948 			AAC_FIBSTATE_NORM	|
949 			AAC_FIBSTATE_ASYNC	 |
950 			AAC_FIBSTATE_FAST_RESPONSE;
951 
952 	aac_enqueue_ready(cm);
953 	aacraid_startio(cm->cm_sc);
954 }
955 
956 static void
957 aac_cam_action(struct cam_sim *sim, union ccb *ccb)
958 {
959 	struct	aac_cam *camsc;
960 	struct	aac_softc *sc;
961 
962 	camsc = (struct aac_cam *)cam_sim_softc(sim);
963 	sc = camsc->inf->aac_sc;
964 	fwprintf(sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "");
965 	mtx_assert(&sc->aac_io_lock, MA_OWNED);
966 
967 	/* Synchronous ops, and ops that don't require communication with the
968 	 * controller */
969 	switch(ccb->ccb_h.func_code) {
970 	case XPT_SCSI_IO:
971 		/* This is handled down below */
972 		break;
973 	case XPT_CALC_GEOMETRY:
974 	{
975 		struct ccb_calc_geometry *ccg;
976 		u_int32_t size_mb;
977 		u_int32_t secs_per_cylinder;
978 
979 		ccg = &ccb->ccg;
980 		size_mb = ccg->volume_size /
981 		    ((1024L * 1024L) / ccg->block_size);
982 		if (size_mb >= (2 * 1024)) {		/* 2GB */
983 			ccg->heads = 255;
984 			ccg->secs_per_track = 63;
985 		} else if (size_mb >= (1 * 1024)) {	/* 1GB */
986 			ccg->heads = 128;
987 			ccg->secs_per_track = 32;
988 		} else {
989 			ccg->heads = 64;
990 			ccg->secs_per_track = 32;
991 		}
992 		secs_per_cylinder = ccg->heads * ccg->secs_per_track;
993 		ccg->cylinders = ccg->volume_size / secs_per_cylinder;
994 
995 		ccb->ccb_h.status = CAM_REQ_CMP;
996 		xpt_done(ccb);
997 		return;
998 	}
999 	case XPT_PATH_INQ:
1000 	{
1001 		struct ccb_pathinq *cpi = &ccb->cpi;
1002 
1003 		cpi->version_num = 1;
1004 		cpi->target_sprt = 0;
1005 		cpi->hba_eng_cnt = 0;
1006 		cpi->max_target = camsc->inf->TargetsPerBus - 1;
1007 		cpi->max_lun = 7;	/* Per the controller spec */
1008 		cpi->initiator_id = camsc->inf->InitiatorBusId;
1009 		cpi->bus_id = camsc->inf->BusNumber;
1010 		cpi->maxio = AAC_MAXIO_SIZE(sc);
1011 
1012 		/*
1013 		 * Resetting via the passthrough or parallel bus scan
1014 		 * causes problems.
1015 		 */
1016 		cpi->hba_misc = PIM_NOBUSRESET;
1017 		cpi->hba_inquiry = PI_TAG_ABLE;
1018 		cpi->base_transfer_speed = 300000;
1019 #ifdef CAM_NEW_TRAN_CODE
1020 		cpi->hba_misc |= PIM_SEQSCAN;
1021 		cpi->protocol = PROTO_SCSI;
1022 		cpi->transport = XPORT_SAS;
1023 		cpi->transport_version = 0;
1024 		cpi->protocol_version = SCSI_REV_SPC2;
1025 #endif
1026 		strlcpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN);
1027 		strlcpy(cpi->hba_vid, "PMC-Sierra", HBA_IDLEN);
1028 		strlcpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN);
1029 		cpi->unit_number = cam_sim_unit(sim);
1030 		ccb->ccb_h.status = CAM_REQ_CMP;
1031 		xpt_done(ccb);
1032 		return;
1033 	}
1034 	case XPT_GET_TRAN_SETTINGS:
1035 	{
1036 #ifdef CAM_NEW_TRAN_CODE
1037 		struct ccb_trans_settings_scsi *scsi =
1038 			&ccb->cts.proto_specific.scsi;
1039 		struct ccb_trans_settings_spi *spi =
1040 			&ccb->cts.xport_specific.spi;
1041 		ccb->cts.protocol = PROTO_SCSI;
1042 		ccb->cts.protocol_version = SCSI_REV_SPC2;
1043 		ccb->cts.transport = XPORT_SAS;
1044 		ccb->cts.transport_version = 0;
1045 		scsi->valid = CTS_SCSI_VALID_TQ;
1046 		scsi->flags = CTS_SCSI_FLAGS_TAG_ENB;
1047 		spi->valid |= CTS_SPI_VALID_DISC;
1048 		spi->flags |= CTS_SPI_FLAGS_DISC_ENB;
1049 #else
1050 		ccb->cts.flags = ~(CCB_TRANS_DISC_ENB | CCB_TRANS_TAG_ENB);
1051 		ccb->cts.valid = CCB_TRANS_DISC_VALID | CCB_TRANS_TQ_VALID;
1052 #endif
1053 		ccb->ccb_h.status = CAM_REQ_CMP;
1054 		xpt_done(ccb);
1055 		return;
1056 	}
1057 	case XPT_SET_TRAN_SETTINGS:
1058 		ccb->ccb_h.status = CAM_FUNC_NOTAVAIL;
1059 		xpt_done(ccb);
1060 		return;
1061 	case XPT_RESET_BUS:
1062 		if (!(sc->flags & AAC_FLAGS_CAM_NORESET) &&
1063 			camsc->inf->BusType != CONTAINER_BUS) {
1064 			ccb->ccb_h.status = aac_cam_reset_bus(sim, ccb);
1065 		} else {
1066 			ccb->ccb_h.status = CAM_REQ_CMP;
1067 		}
1068 		xpt_done(ccb);
1069 		return;
1070 	case XPT_RESET_DEV:
1071 		ccb->ccb_h.status = CAM_REQ_CMP;
1072 		xpt_done(ccb);
1073 		return;
1074 	case XPT_ABORT:
1075 		ccb->ccb_h.status = aac_cam_abort_ccb(sim, ccb);
1076 		xpt_done(ccb);
1077 		return;
1078 	case XPT_TERM_IO:
1079 		ccb->ccb_h.status = aac_cam_term_io(sim, ccb);
1080 		xpt_done(ccb);
1081 		return;
1082 	default:
1083 		device_printf(sc->aac_dev, "Unsupported command 0x%x\n",
1084 		    ccb->ccb_h.func_code);
1085 		ccb->ccb_h.status = CAM_PROVIDE_FAIL;
1086 		xpt_done(ccb);
1087 		return;
1088 	}
1089 
1090 	/* Async ops that require communcation with the controller */
1091 	if (camsc->inf->BusType == CONTAINER_BUS) {
1092 		u_int8_t *cmdp;
1093 
1094 		if (ccb->ccb_h.flags & CAM_CDB_POINTER)
1095 			cmdp = ccb->csio.cdb_io.cdb_ptr;
1096 		else
1097 			cmdp = &ccb->csio.cdb_io.cdb_bytes[0];
1098 
1099 		if (*cmdp==READ_6 || *cmdp==WRITE_6 || *cmdp==READ_10 ||
1100 			*cmdp==WRITE_10 || *cmdp==READ_12 || *cmdp==WRITE_12 ||
1101 			*cmdp==READ_16 || *cmdp==WRITE_16)
1102 			aac_container_rw_command(sim, ccb, cmdp);
1103 		else
1104 			aac_container_special_command(sim, ccb, cmdp);
1105 	} else {
1106 		aac_passthrough_command(sim, ccb);
1107 	}
1108 }
1109 
1110 static void
1111 aac_cam_poll(struct cam_sim *sim)
1112 {
1113 	/*
1114 	 * Pinging the interrupt routine isn't very safe, nor is it
1115 	 * really necessary.  Do nothing.
1116 	 */
1117 }
1118 
1119 static void
1120 aac_container_complete(struct aac_command *cm)
1121 {
1122 	union	ccb *ccb;
1123 	u_int32_t status;
1124 
1125 	fwprintf(cm->cm_sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "");
1126 	ccb = cm->cm_ccb;
1127 	status = le32toh(((u_int32_t *)cm->cm_fib->data)[0]);
1128 
1129 	if (cm->cm_flags & AAC_CMD_RESET) {
1130 		ccb->ccb_h.status = CAM_SCSI_BUS_RESET;
1131 	} else if (status == ST_OK) {
1132 		ccb->ccb_h.status = CAM_REQ_CMP;
1133 	} else if (status == ST_NOT_READY) {
1134 		ccb->ccb_h.status = CAM_BUSY;
1135 	} else {
1136 		ccb->ccb_h.status = CAM_REQ_CMP_ERR;
1137 	}
1138 
1139 	aacraid_release_command(cm);
1140 	xpt_done(ccb);
1141 }
1142 
1143 static void
1144 aac_cam_complete(struct aac_command *cm)
1145 {
1146 	union	ccb *ccb;
1147 	struct 	aac_srb_response *srbr;
1148 	struct	aac_softc *sc;
1149 
1150 	sc = cm->cm_sc;
1151 	fwprintf(sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "");
1152 	ccb = cm->cm_ccb;
1153 	srbr = (struct aac_srb_response *)&cm->cm_fib->data[0];
1154 	aac_srb_response_toh(srbr);
1155 
1156 	if (cm->cm_flags & AAC_CMD_FASTRESP) {
1157 		/* fast response */
1158 		srbr->srb_status = CAM_REQ_CMP;
1159 		srbr->scsi_status = SCSI_STATUS_OK;
1160 		srbr->sense_len = 0;
1161 	}
1162 
1163 	if (cm->cm_flags & AAC_CMD_RESET) {
1164 		ccb->ccb_h.status = CAM_SCSI_BUS_RESET;
1165 	} else if (srbr->fib_status != 0) {
1166 		device_printf(sc->aac_dev, "Passthru FIB failed!\n");
1167 		ccb->ccb_h.status = CAM_REQ_ABORTED;
1168 	} else {
1169 		/*
1170 		 * The SRB error codes just happen to match the CAM error
1171 		 * codes.  How convenient!
1172 		 */
1173 		ccb->ccb_h.status = srbr->srb_status;
1174 
1175 		/* Take care of SCSI_IO ops. */
1176 		if (ccb->ccb_h.func_code == XPT_SCSI_IO) {
1177 			u_int8_t command, device;
1178 
1179 			ccb->csio.scsi_status = srbr->scsi_status;
1180 
1181 			/* Take care of autosense */
1182 			if (srbr->sense_len) {
1183 				int sense_len, scsi_sense_len;
1184 
1185 				scsi_sense_len = sizeof(struct scsi_sense_data);
1186 				bzero(&ccb->csio.sense_data, scsi_sense_len);
1187 				sense_len = (srbr->sense_len >
1188 				    scsi_sense_len) ? scsi_sense_len :
1189 				    srbr->sense_len;
1190 				bcopy(&srbr->sense[0], &ccb->csio.sense_data,
1191 				    sense_len);
1192 				ccb->csio.sense_len = sense_len;
1193 				ccb->ccb_h.status |= CAM_AUTOSNS_VALID;
1194 				// scsi_sense_print(&ccb->csio);
1195 			}
1196 
1197 			/* If this is an inquiry command, fake things out */
1198 			if (ccb->ccb_h.flags & CAM_CDB_POINTER)
1199 				command = ccb->csio.cdb_io.cdb_ptr[0];
1200 			else
1201 				command = ccb->csio.cdb_io.cdb_bytes[0];
1202 
1203 			if (command == INQUIRY) {
1204 				/* Ignore Data Overrun errors on INQUIRY */
1205 				if ((ccb->ccb_h.status & CAM_STATUS_MASK) ==
1206 				    CAM_DATA_RUN_ERR)
1207 					ccb->ccb_h.status = (ccb->ccb_h.status &
1208 					    ~CAM_STATUS_MASK) | CAM_REQ_CMP;
1209 
1210 				if (ccb->ccb_h.status == CAM_REQ_CMP) {
1211 				  device = ccb->csio.data_ptr[0] & 0x1f;
1212 				  /*
1213 				   * We want DASD and PROC devices to only be
1214 				   * visible through the pass device.
1215 				   */
1216 				  if ((device == T_DIRECT &&
1217 				    !(sc->aac_feature_bits & AAC_SUPPL_SUPPORTED_JBOD)) ||
1218 				    (device == T_PROCESSOR))
1219 				    ccb->csio.data_ptr[0] =
1220 				  	((device & 0xe0) | T_NODEVICE);
1221 
1222 				  /* handle phys. components of a log. drive */
1223 				  if (ccb->csio.data_ptr[0] & 0x20) {
1224 					if (sc->hint_flags & 8) {
1225 					  /* expose phys. device (daXX) */
1226 					  ccb->csio.data_ptr[0] &= 0xdf;
1227 					} else {
1228 					  /* phys. device only visible through pass device (passXX) */
1229 					  ccb->csio.data_ptr[0] |= 0x10;
1230 					}
1231 				  }
1232 				} else if (ccb->ccb_h.status == CAM_SEL_TIMEOUT &&
1233 				  ccb->ccb_h.target_lun != 0) {
1234 				  /* fix for INQUIRYs on Lun>0 */
1235 				  ccb->ccb_h.status = CAM_DEV_NOT_THERE;
1236 				}
1237 			}
1238 		}
1239 	}
1240 
1241 	aacraid_release_command(cm);
1242 	xpt_done(ccb);
1243 }
1244 
1245 static u_int32_t
1246 aac_cam_reset_bus(struct cam_sim *sim, union ccb *ccb)
1247 {
1248 	struct aac_command *cm;
1249 	struct aac_fib *fib;
1250 	struct aac_softc *sc;
1251 	struct aac_cam *camsc;
1252 	struct aac_vmioctl *vmi;
1253 	struct aac_resetbus *rbc;
1254 	u_int32_t rval;
1255 
1256 	camsc = (struct aac_cam *)cam_sim_softc(sim);
1257 	sc = camsc->inf->aac_sc;
1258 
1259 	if (sc == NULL) {
1260 		printf("aac: Null sc?\n");
1261 		return (CAM_REQ_ABORTED);
1262 	}
1263 
1264 	if (aacraid_alloc_command(sc, &cm)) {
1265 		struct aac_event *event;
1266 
1267 		xpt_freeze_simq(sim, 1);
1268 		ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
1269 		ccb->ccb_h.sim_priv.entries[0].ptr = camsc;
1270 		event = malloc(sizeof(struct aac_event), M_AACRAIDCAM,
1271 			M_NOWAIT | M_ZERO);
1272 		if (event == NULL) {
1273 			device_printf(sc->aac_dev,
1274 				"Warning, out of memory for event\n");
1275 			return (CAM_REQ_ABORTED);
1276 		}
1277 		event->ev_callback = aac_cam_event;
1278 		event->ev_arg = ccb;
1279 		event->ev_type = AAC_EVENT_CMFREE;
1280 		aacraid_add_event(sc, event);
1281 		return (CAM_REQ_ABORTED);
1282 	}
1283 
1284 	fib = cm->cm_fib;
1285 	cm->cm_timestamp = time_uptime;
1286 	cm->cm_datalen = 0;
1287 
1288 	fib->Header.Size =
1289 		sizeof(struct aac_fib_header) + sizeof(struct aac_vmioctl);
1290 	fib->Header.XferState =
1291 		AAC_FIBSTATE_HOSTOWNED   |
1292 		AAC_FIBSTATE_INITIALISED |
1293 		AAC_FIBSTATE_EMPTY	 |
1294 		AAC_FIBSTATE_FROMHOST	 |
1295 		AAC_FIBSTATE_REXPECTED   |
1296 		AAC_FIBSTATE_NORM	 |
1297 		AAC_FIBSTATE_ASYNC	 |
1298 		AAC_FIBSTATE_FAST_RESPONSE;
1299 	fib->Header.Command = ContainerCommand;
1300 
1301 	vmi = (struct aac_vmioctl *)&fib->data[0];
1302 	bzero(vmi, sizeof(struct aac_vmioctl));
1303 
1304 	vmi->Command = VM_Ioctl;
1305 	vmi->ObjType = FT_DRIVE;
1306 	vmi->MethId = sc->scsi_method_id;
1307 	vmi->ObjId = 0;
1308 	vmi->IoctlCmd = ResetBus;
1309 
1310 	rbc = (struct aac_resetbus *)&vmi->IoctlBuf[0];
1311 	rbc->BusNumber = camsc->inf->BusNumber - 1;
1312 	aac_vmioctl_tole(vmi);
1313 
1314 	if (aacraid_wait_command(cm) != 0) {
1315 		device_printf(sc->aac_dev,"Error sending ResetBus command\n");
1316 		rval = CAM_REQ_ABORTED;
1317 	} else {
1318 		rval = CAM_REQ_CMP;
1319 	}
1320 	aacraid_release_command(cm);
1321 	return (rval);
1322 }
1323 
1324 static u_int32_t
1325 aac_cam_abort_ccb(struct cam_sim *sim, union ccb *ccb)
1326 {
1327 	return (CAM_UA_ABORT);
1328 }
1329 
1330 static u_int32_t
1331 aac_cam_term_io(struct cam_sim *sim, union ccb *ccb)
1332 {
1333 	return (CAM_UA_TERMIO);
1334 }
1335 
1336 static int
1337 aac_load_map_command_sg(struct aac_softc *sc, struct aac_command *cm)
1338 {
1339 	int error;
1340 
1341 	fwprintf(sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "");
1342 	error = bus_dmamap_load(sc->aac_buffer_dmat,
1343 				cm->cm_datamap, cm->cm_data, cm->cm_datalen,
1344 				aacraid_map_command_sg, cm, 0);
1345 	if (error == EINPROGRESS) {
1346 		fwprintf(sc, HBA_FLAGS_DBG_INIT_B, "freezing queue\n");
1347 		sc->flags |= AAC_QUEUE_FRZN;
1348 		error = 0;
1349 	} else if (error != 0) {
1350 		panic("aac_load_map_command_sg: unexpected error %d from "
1351 	     		"busdma", error);
1352 	}
1353 	return(error);
1354 }
1355 
1356 /*
1357  * Start as much queued I/O as possible on the controller
1358  */
1359 void
1360 aacraid_startio(struct aac_softc *sc)
1361 {
1362 	struct aac_command *cm;
1363 
1364 	fwprintf(sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "");
1365 
1366 	for (;;) {
1367 		if (sc->aac_state & AAC_STATE_RESET) {
1368 			fwprintf(sc, HBA_FLAGS_DBG_ERROR_B, "AAC_STATE_RESET");
1369 			break;
1370 		}
1371 		/*
1372 		 * This flag might be set if the card is out of resources.
1373 		 * Checking it here prevents an infinite loop of deferrals.
1374 		 */
1375 		if (sc->flags & AAC_QUEUE_FRZN) {
1376 			fwprintf(sc, HBA_FLAGS_DBG_ERROR_B, "AAC_QUEUE_FRZN");
1377 			break;
1378 		}
1379 
1380 		/*
1381 		 * Try to get a command that's been put off for lack of
1382 		 * resources
1383 		 */
1384 		if ((sc->flags & AAC_FLAGS_SYNC_MODE) && sc->aac_sync_cm)
1385 			break;
1386 		cm = aac_dequeue_ready(sc);
1387 
1388 		/* nothing to do? */
1389 		if (cm == NULL)
1390 			break;
1391 
1392 		/* don't map more than once */
1393 		if (cm->cm_flags & AAC_CMD_MAPPED)
1394 			panic("aac: command %p already mapped", cm);
1395 
1396 		/*
1397 		 * Set up the command to go to the controller.  If there are no
1398 		 * data buffers associated with the command then it can bypass
1399 		 * busdma.
1400 		 */
1401 		if (cm->cm_datalen)
1402 			aac_load_map_command_sg(sc, cm);
1403 		else
1404 			aacraid_map_command_sg(cm, NULL, 0, 0);
1405 	}
1406 }
1407