1 /*- 2 * Copyright (c) 2011 HighPoint Technologies, Inc. 3 * All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer. 10 * 2. Redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution. 13 * 14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 17 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 24 * SUCH DAMAGE. 25 * 26 * $FreeBSD: src/sys/dev/hpt27xx/osm_bsd.c,v 1.1 2011/12/28 23:26:58 delphij Exp $ 27 */ 28 29 #include <dev/raid/hpt27xx/hpt27xx_config.h> 30 31 #include <dev/raid/hpt27xx/os_bsd.h> 32 #include <dev/raid/hpt27xx/hptintf.h> 33 34 static int hpt_probe(device_t dev) 35 { 36 PCI_ID pci_id; 37 HIM *him; 38 int i; 39 PHBA hba; 40 41 for (him = him_list; him; him = him->next) { 42 for (i=0; him->get_supported_device_id(i, &pci_id); i++) { 43 if (him->get_controller_count) 44 him->get_controller_count(&pci_id,0,0); 45 if ((pci_get_vendor(dev) == pci_id.vid) && 46 (pci_get_device(dev) == pci_id.did)){ 47 KdPrint(("hpt_probe: adapter at PCI %d:%d:%d, IRQ %d", 48 pci_get_bus(dev), pci_get_slot(dev), pci_get_function(dev), pci_get_irq(dev) 49 )); 50 device_set_desc(dev, him->name); 51 hba = (PHBA)device_get_softc(dev); 52 memset(hba, 0, sizeof(HBA)); 53 hba->ext_type = EXT_TYPE_HBA; 54 hba->ldm_adapter.him = him; 55 return 0; 56 } 57 } 58 } 59 60 return (ENXIO); 61 } 62 63 static int hpt_attach(device_t dev) 64 { 65 PHBA hba = (PHBA)device_get_softc(dev); 66 HIM *him = hba->ldm_adapter.him; 67 PCI_ID pci_id; 68 HPT_UINT size; 69 PVBUS vbus; 70 PVBUS_EXT vbus_ext; 71 72 KdPrint(("hpt_attach(%d/%d/%d)", pci_get_bus(dev), pci_get_slot(dev), pci_get_function(dev))); 73 74 pci_enable_busmaster(dev); 75 76 pci_id.vid = pci_get_vendor(dev); 77 pci_id.did = pci_get_device(dev); 78 pci_id.rev = pci_get_revid(dev); 79 pci_id.subsys = (HPT_U32)(pci_get_subdevice(dev)) << 16 | pci_get_subvendor(dev); 80 81 size = him->get_adapter_size(&pci_id); 82 hba->ldm_adapter.him_handle = kmalloc(size, M_DEVBUF, M_WAITOK); 83 if (!hba->ldm_adapter.him_handle) 84 return ENXIO; 85 86 hba->pcidev = dev; 87 hba->pciaddr.tree = 0; 88 hba->pciaddr.bus = pci_get_bus(dev); 89 hba->pciaddr.device = pci_get_slot(dev); 90 hba->pciaddr.function = pci_get_function(dev); 91 92 if (!him->create_adapter(&pci_id, hba->pciaddr, hba->ldm_adapter.him_handle, hba)) { 93 kfree(hba->ldm_adapter.him_handle, M_DEVBUF); 94 return -1; 95 } 96 97 os_printk("adapter at PCI %d:%d:%d, IRQ %d", 98 hba->pciaddr.bus, hba->pciaddr.device, hba->pciaddr.function, pci_get_irq(dev)); 99 100 if (!ldm_register_adapter(&hba->ldm_adapter)) { 101 size = ldm_get_vbus_size(); 102 vbus_ext = kmalloc(sizeof(VBUS_EXT) + size, M_DEVBUF, M_WAITOK); 103 memset(vbus_ext, 0, sizeof(VBUS_EXT)); 104 vbus_ext->ext_type = EXT_TYPE_VBUS; 105 ldm_create_vbus((PVBUS)vbus_ext->vbus, vbus_ext); 106 ldm_register_adapter(&hba->ldm_adapter); 107 } 108 109 ldm_for_each_vbus(vbus, vbus_ext) { 110 if (hba->ldm_adapter.vbus==vbus) { 111 hba->vbus_ext = vbus_ext; 112 hba->next = vbus_ext->hba_list; 113 vbus_ext->hba_list = hba; 114 break; 115 } 116 } 117 return 0; 118 } 119 120 /* 121 * Maybe we'd better to use the bus_dmamem_alloc to alloc DMA memory, 122 * but there are some problems currently (alignment, etc). 123 */ 124 static __inline void *__get_free_pages(int order) 125 { 126 /* don't use low memory - other devices may get starved */ 127 return contigmalloc(PAGE_SIZE<<order, 128 M_DEVBUF, M_WAITOK, BUS_SPACE_MAXADDR_24BIT, BUS_SPACE_MAXADDR, PAGE_SIZE, 0); 129 } 130 131 static __inline void free_pages(void *p, int order) 132 { 133 contigfree(p, PAGE_SIZE<<order, M_DEVBUF); 134 } 135 136 static int hpt_alloc_mem(PVBUS_EXT vbus_ext) 137 { 138 PHBA hba; 139 struct freelist *f; 140 HPT_UINT i; 141 void **p; 142 143 for (hba = vbus_ext->hba_list; hba; hba = hba->next) 144 hba->ldm_adapter.him->get_meminfo(hba->ldm_adapter.him_handle); 145 146 ldm_get_mem_info((PVBUS)vbus_ext->vbus, 0); 147 148 for (f=vbus_ext->freelist_head; f; f=f->next) { 149 KdPrint(("%s: %d*%d=%d bytes", 150 f->tag, f->count, f->size, f->count*f->size)); 151 for (i=0; i<f->count; i++) { 152 p = (void **)kmalloc(f->size, M_DEVBUF, M_WAITOK); 153 if (!p) return (ENXIO); 154 *p = f->head; 155 f->head = p; 156 } 157 } 158 159 for (f=vbus_ext->freelist_dma_head; f; f=f->next) { 160 int order, size, j; 161 162 HPT_ASSERT((f->size & (f->alignment-1))==0); 163 164 for (order=0, size=PAGE_SIZE; size<f->size; order++, size<<=1) ; 165 166 KdPrint(("%s: %d*%d=%d bytes, order %d", 167 f->tag, f->count, f->size, f->count*f->size, order)); 168 HPT_ASSERT(f->alignment<=PAGE_SIZE); 169 170 for (i=0; i<f->count;) { 171 p = (void **)__get_free_pages(order); 172 if (!p) return -1; 173 for (j = size/f->size; j && i<f->count; i++,j--) { 174 *p = f->head; 175 *(BUS_ADDRESS *)(p+1) = (BUS_ADDRESS)vtophys(p); 176 f->head = p; 177 p = (void **)((unsigned long)p + f->size); 178 } 179 } 180 } 181 182 HPT_ASSERT(PAGE_SIZE==DMAPOOL_PAGE_SIZE); 183 184 for (i=0; i<os_max_cache_pages; i++) { 185 p = (void **)__get_free_pages(0); 186 if (!p) return -1; 187 HPT_ASSERT(((HPT_UPTR)p & (DMAPOOL_PAGE_SIZE-1))==0); 188 dmapool_put_page((PVBUS)vbus_ext->vbus, p, (BUS_ADDRESS)vtophys(p)); 189 } 190 191 return 0; 192 } 193 194 static void hpt_free_mem(PVBUS_EXT vbus_ext) 195 { 196 struct freelist *f; 197 void *p; 198 int i; 199 BUS_ADDRESS bus; 200 201 for (f=vbus_ext->freelist_head; f; f=f->next) { 202 #if DBG 203 if (f->count!=f->reserved_count) { 204 KdPrint(("memory leak for freelist %s (%d/%d)", f->tag, f->count, f->reserved_count)); 205 } 206 #endif 207 while ((p=freelist_get(f))) 208 kfree(p, M_DEVBUF); 209 } 210 211 for (i=0; i<os_max_cache_pages; i++) { 212 p = dmapool_get_page((PVBUS)vbus_ext->vbus, &bus); 213 HPT_ASSERT(p); 214 free_pages(p, 0); 215 } 216 217 for (f=vbus_ext->freelist_dma_head; f; f=f->next) { 218 int order, size; 219 #if DBG 220 if (f->count!=f->reserved_count) { 221 KdPrint(("memory leak for dma freelist %s (%d/%d)", f->tag, f->count, f->reserved_count)); 222 } 223 #endif 224 for (order=0, size=PAGE_SIZE; size<f->size; order++, size<<=1) ; 225 226 while ((p=freelist_get_dma(f, &bus))) { 227 if (order) 228 free_pages(p, order); 229 else { 230 /* can't free immediately since other blocks in this page may still be in the list */ 231 if (((HPT_UPTR)p & (PAGE_SIZE-1))==0) 232 dmapool_put_page((PVBUS)vbus_ext->vbus, p, bus); 233 } 234 } 235 } 236 237 while ((p = dmapool_get_page((PVBUS)vbus_ext->vbus, &bus))) 238 free_pages(p, 0); 239 } 240 241 static int hpt_init_vbus(PVBUS_EXT vbus_ext) 242 { 243 PHBA hba; 244 245 for (hba = vbus_ext->hba_list; hba; hba = hba->next) 246 if (!hba->ldm_adapter.him->initialize(hba->ldm_adapter.him_handle)) { 247 KdPrint(("fail to initialize %p", hba)); 248 return -1; 249 } 250 251 ldm_initialize_vbus((PVBUS)vbus_ext->vbus, &vbus_ext->hba_list->ldm_adapter); 252 return 0; 253 } 254 255 static void hpt_flush_done(PCOMMAND pCmd) 256 { 257 PVDEV vd = pCmd->target; 258 259 if (mIsArray(vd->type) && vd->u.array.transform && vd!=vd->u.array.transform->target) { 260 vd = vd->u.array.transform->target; 261 HPT_ASSERT(vd); 262 pCmd->target = vd; 263 pCmd->Result = RETURN_PENDING; 264 vdev_queue_cmd(pCmd); 265 return; 266 } 267 268 *(int *)pCmd->priv = 1; 269 wakeup(pCmd); 270 } 271 272 /* 273 * flush a vdev (without retry). 274 */ 275 static int hpt_flush_vdev(PVBUS_EXT vbus_ext, PVDEV vd) 276 { 277 PCOMMAND pCmd; 278 int result = 0, done; 279 HPT_UINT count; 280 281 KdPrint(("flusing dev %p", vd)); 282 283 hpt_lock_vbus(vbus_ext); 284 285 if (mIsArray(vd->type) && vd->u.array.transform) 286 count = MAX(vd->u.array.transform->source->cmds_per_request, 287 vd->u.array.transform->target->cmds_per_request); 288 else 289 count = vd->cmds_per_request; 290 291 pCmd = ldm_alloc_cmds(vd->vbus, count); 292 293 if (!pCmd) { 294 hpt_unlock_vbus(vbus_ext); 295 return -1; 296 } 297 298 pCmd->type = CMD_TYPE_FLUSH; 299 pCmd->flags.hard_flush = 1; 300 pCmd->target = vd; 301 pCmd->done = hpt_flush_done; 302 done = 0; 303 pCmd->priv = &done; 304 305 ldm_queue_cmd(pCmd); 306 307 if (!done) { 308 while (hpt_sleep(vbus_ext, pCmd, 0, "hptfls", HPT_OSM_TIMEOUT)) { 309 ldm_reset_vbus(vd->vbus); 310 } 311 } 312 313 KdPrint(("flush result %d", pCmd->Result)); 314 315 if (pCmd->Result!=RETURN_SUCCESS) 316 result = -1; 317 318 ldm_free_cmds(pCmd); 319 320 hpt_unlock_vbus(vbus_ext); 321 322 return result; 323 } 324 325 static void hpt_stop_tasks(PVBUS_EXT vbus_ext); 326 static void hpt_shutdown_vbus(PVBUS_EXT vbus_ext, int howto) 327 { 328 PVBUS vbus = (PVBUS)vbus_ext->vbus; 329 PHBA hba; 330 int i; 331 332 KdPrint(("hpt_shutdown_vbus")); 333 334 /* stop all ctl tasks and disable the worker taskqueue */ 335 hpt_stop_tasks(vbus_ext); 336 vbus_ext->worker.ta_context = NULL; 337 338 /* flush devices */ 339 for (i=0; i<osm_max_targets; i++) { 340 PVDEV vd = ldm_find_target(vbus, i); 341 if (vd) { 342 /* retry once */ 343 if (hpt_flush_vdev(vbus_ext, vd)) 344 hpt_flush_vdev(vbus_ext, vd); 345 } 346 } 347 348 hpt_lock_vbus(vbus_ext); 349 ldm_shutdown(vbus); 350 hpt_unlock_vbus(vbus_ext); 351 352 ldm_release_vbus(vbus); 353 354 for (hba=vbus_ext->hba_list; hba; hba=hba->next) 355 bus_teardown_intr(hba->pcidev, hba->irq_res, hba->irq_handle); 356 357 hpt_free_mem(vbus_ext); 358 359 while ((hba=vbus_ext->hba_list)) { 360 vbus_ext->hba_list = hba->next; 361 kfree(hba->ldm_adapter.him_handle, M_DEVBUF); 362 } 363 364 kfree(vbus_ext, M_DEVBUF); 365 KdPrint(("hpt_shutdown_vbus done")); 366 } 367 368 static void __hpt_do_tasks(PVBUS_EXT vbus_ext) 369 { 370 OSM_TASK *tasks; 371 372 tasks = vbus_ext->tasks; 373 vbus_ext->tasks = NULL; 374 375 while (tasks) { 376 OSM_TASK *t = tasks; 377 tasks = t->next; 378 t->next = NULL; 379 t->func(vbus_ext->vbus, t->data); 380 } 381 } 382 383 static void hpt_do_tasks(PVBUS_EXT vbus_ext, int pending) 384 { 385 if(vbus_ext){ 386 hpt_lock_vbus(vbus_ext); 387 __hpt_do_tasks(vbus_ext); 388 hpt_unlock_vbus(vbus_ext); 389 } 390 } 391 392 static void hpt_action(struct cam_sim *sim, union ccb *ccb); 393 static void hpt_poll(struct cam_sim *sim); 394 static void hpt_async(void * callback_arg, u_int32_t code, struct cam_path * path, void * arg); 395 static void hpt_pci_intr(void *arg); 396 397 static __inline POS_CMDEXT cmdext_get(PVBUS_EXT vbus_ext) 398 { 399 POS_CMDEXT p = vbus_ext->cmdext_list; 400 if (p) 401 vbus_ext->cmdext_list = p->next; 402 return p; 403 } 404 405 static __inline void cmdext_put(POS_CMDEXT p) 406 { 407 p->next = p->vbus_ext->cmdext_list; 408 p->vbus_ext->cmdext_list = p; 409 } 410 411 static void hpt_timeout(void *arg) 412 { 413 PCOMMAND pCmd = (PCOMMAND)arg; 414 POS_CMDEXT ext = (POS_CMDEXT)pCmd->priv; 415 416 KdPrint(("pCmd %p timeout", pCmd)); 417 418 ldm_reset_vbus((PVBUS)ext->vbus_ext->vbus); 419 } 420 421 static void os_cmddone(PCOMMAND pCmd) 422 { 423 POS_CMDEXT ext = (POS_CMDEXT)pCmd->priv; 424 union ccb *ccb = ext->ccb; 425 426 KdPrint(("os_cmddone(%p, %d)", pCmd, pCmd->Result)); 427 428 callout_stop(ccb->ccb_h.timeout_ch); 429 430 switch(pCmd->Result) { 431 case RETURN_SUCCESS: 432 ccb->ccb_h.status = CAM_REQ_CMP; 433 break; 434 case RETURN_BAD_DEVICE: 435 ccb->ccb_h.status = CAM_DEV_NOT_THERE; 436 break; 437 case RETURN_DEVICE_BUSY: 438 ccb->ccb_h.status = CAM_BUSY; 439 break; 440 case RETURN_INVALID_REQUEST: 441 ccb->ccb_h.status = CAM_REQ_INVALID; 442 break; 443 case RETURN_SELECTION_TIMEOUT: 444 ccb->ccb_h.status = CAM_SEL_TIMEOUT; 445 break; 446 case RETURN_RETRY: 447 ccb->ccb_h.status = CAM_BUSY; 448 break; 449 default: 450 ccb->ccb_h.status = CAM_SCSI_STATUS_ERROR; 451 break; 452 } 453 454 if (pCmd->flags.data_in) { 455 bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map, BUS_DMASYNC_POSTREAD); 456 } 457 else if (pCmd->flags.data_out) { 458 bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map, BUS_DMASYNC_POSTWRITE); 459 } 460 461 bus_dmamap_unload(ext->vbus_ext->io_dmat, ext->dma_map); 462 463 cmdext_put(ext); 464 ldm_free_cmds(pCmd); 465 xpt_done(ccb); 466 } 467 468 static int os_buildsgl(PCOMMAND pCmd, PSG pSg, int logical) 469 { 470 POS_CMDEXT ext = (POS_CMDEXT)pCmd->priv; 471 union ccb *ccb = ext->ccb; 472 bus_dma_segment_t *sgList = (bus_dma_segment_t *)ccb->csio.data_ptr; 473 int idx; 474 475 if(logical) { 476 if (ccb->ccb_h.flags & CAM_DATA_PHYS) 477 panic("physical address unsupported"); 478 479 if (ccb->ccb_h.flags & CAM_SCATTER_VALID) { 480 if (ccb->ccb_h.flags & CAM_SG_LIST_PHYS) 481 panic("physical address unsupported"); 482 483 for (idx = 0; idx < ccb->csio.sglist_cnt; idx++) { 484 os_set_sgptr(&pSg[idx], (HPT_U8 *)(HPT_UPTR)sgList[idx].ds_addr); 485 pSg[idx].size = sgList[idx].ds_len; 486 pSg[idx].eot = (idx==ccb->csio.sglist_cnt-1)? 1 : 0; 487 } 488 } 489 else { 490 os_set_sgptr(pSg, (HPT_U8 *)ccb->csio.data_ptr); 491 pSg->size = ccb->csio.dxfer_len; 492 pSg->eot = 1; 493 } 494 return TRUE; 495 } 496 497 /* since we have provided physical sg, nobody will ask us to build physical sg */ 498 HPT_ASSERT(0); 499 return FALSE; 500 } 501 502 static void hpt_io_dmamap_callback(void *arg, bus_dma_segment_t *segs, int nsegs, int error) 503 { 504 PCOMMAND pCmd = (PCOMMAND)arg; 505 POS_CMDEXT ext = (POS_CMDEXT)pCmd->priv; 506 PSG psg = pCmd->psg; 507 int idx; 508 509 HPT_ASSERT(pCmd->flags.physical_sg); 510 511 if (error || nsegs == 0) 512 panic("busdma error"); 513 514 HPT_ASSERT(nsegs<=os_max_sg_descriptors); 515 516 for (idx = 0; idx < nsegs; idx++, psg++) { 517 psg->addr.bus = segs[idx].ds_addr; 518 psg->size = segs[idx].ds_len; 519 psg->eot = 0; 520 } 521 psg[-1].eot = 1; 522 523 if (pCmd->flags.data_in) { 524 bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map, BUS_DMASYNC_PREREAD); 525 } 526 else if (pCmd->flags.data_out) { 527 bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map, BUS_DMASYNC_PREWRITE); 528 } 529 530 callout_reset(ext->ccb->ccb_h.timeout_ch, HPT_OSM_TIMEOUT, 531 hpt_timeout, pCmd); 532 ldm_queue_cmd(pCmd); 533 } 534 535 static void hpt_scsi_io(PVBUS_EXT vbus_ext, union ccb *ccb) 536 { 537 PVBUS vbus = (PVBUS)vbus_ext->vbus; 538 PVDEV vd; 539 PCOMMAND pCmd; 540 POS_CMDEXT ext; 541 HPT_U8 *cdb; 542 543 if (ccb->ccb_h.flags & CAM_CDB_POINTER) 544 cdb = ccb->csio.cdb_io.cdb_ptr; 545 else 546 cdb = ccb->csio.cdb_io.cdb_bytes; 547 548 KdPrint(("hpt_scsi_io: ccb %x id %d lun %d cdb %x-%x-%x", 549 ccb, 550 ccb->ccb_h.target_id, ccb->ccb_h.target_lun, 551 *(HPT_U32 *)&cdb[0], *(HPT_U32 *)&cdb[4], *(HPT_U32 *)&cdb[8] 552 )); 553 554 /* ccb->ccb_h.path_id is not our bus id - don't check it */ 555 if (ccb->ccb_h.target_lun != 0 || 556 ccb->ccb_h.target_id >= osm_max_targets || 557 (ccb->ccb_h.flags & CAM_CDB_PHYS)) 558 { 559 ccb->ccb_h.status = CAM_TID_INVALID; 560 xpt_done(ccb); 561 return; 562 } 563 564 vd = ldm_find_target(vbus, ccb->ccb_h.target_id); 565 566 if (!vd) { 567 ccb->ccb_h.status = CAM_TID_INVALID; 568 xpt_done(ccb); 569 return; 570 } 571 572 switch (cdb[0]) { 573 case TEST_UNIT_READY: 574 case START_STOP_UNIT: 575 case SYNCHRONIZE_CACHE: 576 ccb->ccb_h.status = CAM_REQ_CMP; 577 break; 578 579 case INQUIRY: 580 { 581 PINQUIRYDATA inquiryData; 582 memset(ccb->csio.data_ptr, 0, ccb->csio.dxfer_len); 583 inquiryData = (PINQUIRYDATA)ccb->csio.data_ptr; 584 585 inquiryData->AdditionalLength = 31; 586 inquiryData->CommandQueue = 1; 587 memcpy(&inquiryData->VendorId, "HPT ", 8); 588 memcpy(&inquiryData->ProductId, "DISK 0_0 ", 16); 589 590 if (vd->target_id / 10) { 591 inquiryData->ProductId[7] = (vd->target_id % 100) / 10 + '0'; 592 inquiryData->ProductId[8] = (vd->target_id % 100) % 10 + '0'; 593 } 594 else 595 inquiryData->ProductId[7] = (vd->target_id % 100) % 10 + '0'; 596 597 memcpy(&inquiryData->ProductRevisionLevel, "4.00", 4); 598 599 ccb->ccb_h.status = CAM_REQ_CMP; 600 } 601 break; 602 603 case READ_CAPACITY: 604 { 605 HPT_U8 *rbuf = ccb->csio.data_ptr; 606 HPT_U32 cap; 607 608 if (vd->capacity>0xfffffffful) 609 cap = 0xfffffffful; 610 else 611 cap = vd->capacity - 1; 612 613 rbuf[0] = (HPT_U8)(cap>>24); 614 rbuf[1] = (HPT_U8)(cap>>16); 615 rbuf[2] = (HPT_U8)(cap>>8); 616 rbuf[3] = (HPT_U8)cap; 617 rbuf[4] = 0; 618 rbuf[5] = 0; 619 rbuf[6] = 2; 620 rbuf[7] = 0; 621 622 ccb->ccb_h.status = CAM_REQ_CMP; 623 break; 624 } 625 626 case SERVICE_ACTION_IN: 627 { 628 HPT_U8 *rbuf = ccb->csio.data_ptr; 629 HPT_U64 cap = vd->capacity - 1; 630 631 rbuf[0] = (HPT_U8)(cap>>56); 632 rbuf[1] = (HPT_U8)(cap>>48); 633 rbuf[2] = (HPT_U8)(cap>>40); 634 rbuf[3] = (HPT_U8)(cap>>32); 635 rbuf[4] = (HPT_U8)(cap>>24); 636 rbuf[5] = (HPT_U8)(cap>>16); 637 rbuf[6] = (HPT_U8)(cap>>8); 638 rbuf[7] = (HPT_U8)cap; 639 rbuf[8] = 0; 640 rbuf[9] = 0; 641 rbuf[10] = 2; 642 rbuf[11] = 0; 643 644 ccb->ccb_h.status = CAM_REQ_CMP; 645 break; 646 } 647 648 case READ_6: 649 case READ_10: 650 case READ_16: 651 case WRITE_6: 652 case WRITE_10: 653 case WRITE_16: 654 case 0x13: 655 case 0x2f: 656 case 0x8f: /* VERIFY_16 */ 657 { 658 pCmd = ldm_alloc_cmds(vbus, vd->cmds_per_request); 659 if(!pCmd){ 660 KdPrint(("Failed to allocate command!")); 661 ccb->ccb_h.status = CAM_BUSY; 662 break; 663 } 664 665 switch (cdb[0]) { 666 case READ_6: 667 case WRITE_6: 668 case 0x13: 669 pCmd->uCmd.Ide.Lba = ((HPT_U32)cdb[1] << 16) | ((HPT_U32)cdb[2] << 8) | (HPT_U32)cdb[3]; 670 pCmd->uCmd.Ide.nSectors = (HPT_U16) cdb[4]; 671 break; 672 case READ_16: 673 case WRITE_16: 674 case 0x8f: /* VERIFY_16 */ 675 { 676 HPT_U64 block = 677 ((HPT_U64)cdb[2]<<56) | 678 ((HPT_U64)cdb[3]<<48) | 679 ((HPT_U64)cdb[4]<<40) | 680 ((HPT_U64)cdb[5]<<32) | 681 ((HPT_U64)cdb[6]<<24) | 682 ((HPT_U64)cdb[7]<<16) | 683 ((HPT_U64)cdb[8]<<8) | 684 ((HPT_U64)cdb[9]); 685 pCmd->uCmd.Ide.Lba = block; 686 pCmd->uCmd.Ide.nSectors = (HPT_U16)cdb[13] | ((HPT_U16)cdb[12]<<8); 687 break; 688 } 689 690 default: 691 pCmd->uCmd.Ide.Lba = (HPT_U32)cdb[5] | ((HPT_U32)cdb[4] << 8) | ((HPT_U32)cdb[3] << 16) | ((HPT_U32)cdb[2] << 24); 692 pCmd->uCmd.Ide.nSectors = (HPT_U16) cdb[8] | ((HPT_U16)cdb[7]<<8); 693 break; 694 } 695 696 switch (cdb[0]) { 697 case READ_6: 698 case READ_10: 699 case READ_16: 700 pCmd->flags.data_in = 1; 701 break; 702 case WRITE_6: 703 case WRITE_10: 704 case WRITE_16: 705 pCmd->flags.data_out = 1; 706 break; 707 } 708 pCmd->priv = ext = cmdext_get(vbus_ext); 709 HPT_ASSERT(ext); 710 ext->ccb = ccb; 711 pCmd->target = vd; 712 pCmd->done = os_cmddone; 713 pCmd->buildsgl = os_buildsgl; 714 715 pCmd->psg = ext->psg; 716 717 if (ccb->ccb_h.flags & CAM_SCATTER_VALID) { 718 int idx; 719 bus_dma_segment_t *sgList = (bus_dma_segment_t *)ccb->csio.data_ptr; 720 721 if (ccb->ccb_h.flags & CAM_SG_LIST_PHYS) 722 pCmd->flags.physical_sg = 1; 723 724 for (idx = 0; idx < ccb->csio.sglist_cnt; idx++) { 725 pCmd->psg[idx].addr.bus = sgList[idx].ds_addr; 726 pCmd->psg[idx].size = sgList[idx].ds_len; 727 pCmd->psg[idx].eot = (idx==ccb->csio.sglist_cnt-1)? 1 : 0; 728 } 729 730 callout_reset(ccb->ccb_h.timeout_ch, HPT_OSM_TIMEOUT, 731 hpt_timeout, pCmd); 732 ldm_queue_cmd(pCmd); 733 } 734 else { 735 int error; 736 pCmd->flags.physical_sg = 1; 737 error = bus_dmamap_load(vbus_ext->io_dmat, 738 ext->dma_map, 739 ccb->csio.data_ptr, ccb->csio.dxfer_len, 740 hpt_io_dmamap_callback, pCmd, 741 BUS_DMA_WAITOK 742 ); 743 KdPrint(("bus_dmamap_load return %d", error)); 744 if (error && error!=EINPROGRESS) { 745 os_printk("bus_dmamap_load error %d", error); 746 cmdext_put(ext); 747 ldm_free_cmds(pCmd); 748 ccb->ccb_h.status = CAM_REQ_CMP_ERR; 749 xpt_done(ccb); 750 } 751 } 752 return; 753 } 754 755 default: 756 ccb->ccb_h.status = CAM_REQ_INVALID; 757 break; 758 } 759 760 xpt_done(ccb); 761 return; 762 } 763 764 static void hpt_action(struct cam_sim *sim, union ccb *ccb) 765 { 766 PVBUS_EXT vbus_ext = (PVBUS_EXT)cam_sim_softc(sim); 767 768 KdPrint(("hpt_action(fn=%d, id=%d)", ccb->ccb_h.func_code, ccb->ccb_h.target_id)); 769 770 switch (ccb->ccb_h.func_code) { 771 772 case XPT_SCSI_IO: 773 hpt_lock_vbus(vbus_ext); 774 hpt_scsi_io(vbus_ext, ccb); 775 hpt_unlock_vbus(vbus_ext); 776 return; 777 778 case XPT_RESET_BUS: 779 hpt_lock_vbus(vbus_ext); 780 ldm_reset_vbus((PVBUS)vbus_ext->vbus); 781 hpt_unlock_vbus(vbus_ext); 782 break; 783 784 case XPT_GET_TRAN_SETTINGS: 785 case XPT_SET_TRAN_SETTINGS: 786 ccb->ccb_h.status = CAM_FUNC_NOTAVAIL; 787 break; 788 789 case XPT_CALC_GEOMETRY: 790 ccb->ccg.heads = 255; 791 ccb->ccg.secs_per_track = 63; 792 ccb->ccg.cylinders = ccb->ccg.volume_size / (ccb->ccg.heads * ccb->ccg.secs_per_track); 793 ccb->ccb_h.status = CAM_REQ_CMP; 794 break; 795 796 case XPT_PATH_INQ: 797 { 798 struct ccb_pathinq *cpi = &ccb->cpi; 799 800 cpi->version_num = 1; 801 cpi->hba_inquiry = PI_SDTR_ABLE; 802 cpi->target_sprt = 0; 803 cpi->hba_misc = PIM_NOBUSRESET; 804 cpi->hba_eng_cnt = 0; 805 cpi->max_target = osm_max_targets; 806 cpi->max_lun = 0; 807 cpi->unit_number = cam_sim_unit(sim); 808 cpi->bus_id = cam_sim_bus(sim); 809 cpi->initiator_id = osm_max_targets; 810 cpi->base_transfer_speed = 3300; 811 812 strncpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN); 813 strncpy(cpi->hba_vid, "HPT ", HBA_IDLEN); 814 strncpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN); 815 cpi->transport = XPORT_SPI; 816 cpi->transport_version = 2; 817 cpi->protocol = PROTO_SCSI; 818 cpi->protocol_version = SCSI_REV_2; 819 cpi->maxio = HPT27XX_DFLTPHYS; 820 cpi->ccb_h.status = CAM_REQ_CMP; 821 break; 822 } 823 824 default: 825 ccb->ccb_h.status = CAM_REQ_INVALID; 826 break; 827 } 828 829 xpt_done(ccb); 830 return; 831 } 832 833 static void hpt_pci_intr(void *arg) 834 { 835 PVBUS_EXT vbus_ext = (PVBUS_EXT)arg; 836 hpt_lock_vbus(vbus_ext); 837 ldm_intr((PVBUS)vbus_ext->vbus); 838 hpt_unlock_vbus(vbus_ext); 839 } 840 841 static void hpt_poll(struct cam_sim *sim) 842 { 843 hpt_pci_intr(cam_sim_softc(sim)); 844 } 845 846 static void hpt_async(void * callback_arg, u_int32_t code, struct cam_path * path, void * arg) 847 { 848 KdPrint(("hpt_async")); 849 } 850 851 static int hpt_shutdown(device_t dev) 852 { 853 KdPrint(("hpt_shutdown(dev=%p)", dev)); 854 return 0; 855 } 856 857 static int hpt_detach(device_t dev) 858 { 859 /* we don't allow the driver to be unloaded. */ 860 return EBUSY; 861 } 862 863 static void hpt_ioctl_done(struct _IOCTL_ARG *arg) 864 { 865 arg->ioctl_cmnd = NULL; 866 wakeup(arg); 867 } 868 869 static void __hpt_do_ioctl(PVBUS_EXT vbus_ext, IOCTL_ARG *ioctl_args) 870 { 871 ioctl_args->result = -1; 872 ioctl_args->done = hpt_ioctl_done; 873 ioctl_args->ioctl_cmnd = (void *)1; 874 875 hpt_lock_vbus(vbus_ext); 876 ldm_ioctl((PVBUS)vbus_ext->vbus, ioctl_args); 877 878 while (ioctl_args->ioctl_cmnd) { 879 if (hpt_sleep(vbus_ext, ioctl_args, 0, "hptctl", HPT_OSM_TIMEOUT)==0) 880 break; 881 ldm_reset_vbus((PVBUS)vbus_ext->vbus); 882 __hpt_do_tasks(vbus_ext); 883 } 884 885 /* KdPrint(("ioctl %x result %d", ioctl_args->dwIoControlCode, ioctl_args->result)); */ 886 887 hpt_unlock_vbus(vbus_ext); 888 } 889 890 static void hpt_do_ioctl(IOCTL_ARG *ioctl_args) 891 { 892 PVBUS vbus; 893 PVBUS_EXT vbus_ext; 894 895 ldm_for_each_vbus(vbus, vbus_ext) { 896 __hpt_do_ioctl(vbus_ext, ioctl_args); 897 if (ioctl_args->result!=HPT_IOCTL_RESULT_WRONG_VBUS) 898 return; 899 } 900 } 901 902 #define HPT_DO_IOCTL(code, inbuf, insize, outbuf, outsize) ({\ 903 IOCTL_ARG arg;\ 904 arg.dwIoControlCode = code;\ 905 arg.lpInBuffer = inbuf;\ 906 arg.lpOutBuffer = outbuf;\ 907 arg.nInBufferSize = insize;\ 908 arg.nOutBufferSize = outsize;\ 909 arg.lpBytesReturned = NULL;\ 910 hpt_do_ioctl(&arg);\ 911 arg.result;\ 912 }) 913 914 #define DEVICEID_VALID(id) ((id) && ((HPT_U32)(id)!=0xffffffff)) 915 916 static int hpt_get_logical_devices(DEVICEID * pIds, int nMaxCount) 917 { 918 int i; 919 HPT_U32 count = nMaxCount-1; 920 921 if (HPT_DO_IOCTL(HPT_IOCTL_GET_LOGICAL_DEVICES, 922 &count, sizeof(HPT_U32), pIds, sizeof(DEVICEID)*nMaxCount)) 923 return -1; 924 925 nMaxCount = (int)pIds[0]; 926 for (i=0; i<nMaxCount; i++) pIds[i] = pIds[i+1]; 927 return nMaxCount; 928 } 929 930 static int hpt_get_device_info_v3(DEVICEID id, PLOGICAL_DEVICE_INFO_V3 pInfo) 931 { 932 return HPT_DO_IOCTL(HPT_IOCTL_GET_DEVICE_INFO_V3, 933 &id, sizeof(DEVICEID), pInfo, sizeof(LOGICAL_DEVICE_INFO_V3)); 934 } 935 936 /* not belong to this file logically, but we want to use ioctl interface */ 937 static int __hpt_stop_tasks(PVBUS_EXT vbus_ext, DEVICEID id) 938 { 939 LOGICAL_DEVICE_INFO_V3 devinfo; 940 int i, result; 941 DEVICEID param[2] = { id, 0 }; 942 943 if (hpt_get_device_info_v3(id, &devinfo)) 944 return -1; 945 946 if (devinfo.Type!=LDT_ARRAY) 947 return -1; 948 949 if (devinfo.u.array.Flags & ARRAY_FLAG_REBUILDING) 950 param[1] = AS_REBUILD_ABORT; 951 else if (devinfo.u.array.Flags & ARRAY_FLAG_VERIFYING) 952 param[1] = AS_VERIFY_ABORT; 953 else if (devinfo.u.array.Flags & ARRAY_FLAG_INITIALIZING) 954 param[1] = AS_INITIALIZE_ABORT; 955 else if (devinfo.u.array.Flags & ARRAY_FLAG_TRANSFORMING) 956 param[1] = AS_TRANSFORM_ABORT; 957 else 958 return -1; 959 960 KdPrint(("SET_ARRAY_STATE(%x, %d)", param[0], param[1])); 961 result = HPT_DO_IOCTL(HPT_IOCTL_SET_ARRAY_STATE, 962 param, sizeof(param), 0, 0); 963 964 for (i=0; i<devinfo.u.array.nDisk; i++) 965 if (DEVICEID_VALID(devinfo.u.array.Members[i])) 966 __hpt_stop_tasks(vbus_ext, devinfo.u.array.Members[i]); 967 968 return result; 969 } 970 971 static void hpt_stop_tasks(PVBUS_EXT vbus_ext) 972 { 973 DEVICEID ids[32]; 974 int i, count; 975 976 count = hpt_get_logical_devices((DEVICEID *)&ids, sizeof(ids)/sizeof(ids[0])); 977 978 for (i=0; i<count; i++) 979 __hpt_stop_tasks(vbus_ext, ids[i]); 980 } 981 982 static d_open_t hpt_open; 983 static d_close_t hpt_close; 984 static d_ioctl_t hpt_ioctl; 985 static void hpt_bus_scan_cb(struct cam_periph *periph, union ccb *ccb); 986 static int hpt_rescan_bus(void); 987 988 static struct dev_ops hpt_ops = { 989 { driver_name, 0, 0 }, 990 .d_open = hpt_open, 991 .d_close = hpt_close, 992 .d_ioctl = hpt_ioctl, 993 }; 994 995 static struct intr_config_hook hpt_ich; 996 997 /* 998 * hpt_final_init will be called after all hpt_attach. 999 */ 1000 static void hpt_final_init(void *dummy) 1001 { 1002 int i; 1003 PVBUS_EXT vbus_ext; 1004 PVBUS vbus; 1005 PHBA hba; 1006 1007 /* Clear the config hook */ 1008 config_intrhook_disestablish(&hpt_ich); 1009 1010 /* allocate memory */ 1011 i = 0; 1012 ldm_for_each_vbus(vbus, vbus_ext) { 1013 if (hpt_alloc_mem(vbus_ext)) { 1014 os_printk("out of memory"); 1015 return; 1016 } 1017 i++; 1018 } 1019 1020 if (!i) { 1021 os_printk("no controller detected."); 1022 return; 1023 } 1024 1025 /* initializing hardware */ 1026 ldm_for_each_vbus(vbus, vbus_ext) { 1027 /* make timer available here */ 1028 callout_init(&vbus_ext->timer); 1029 if (hpt_init_vbus(vbus_ext)) { 1030 os_printk("fail to initialize hardware"); 1031 break; /* FIXME */ 1032 } 1033 } 1034 1035 /* register CAM interface */ 1036 ldm_for_each_vbus(vbus, vbus_ext) { 1037 struct cam_devq *devq; 1038 struct ccb_setasync *ccb; 1039 1040 lockinit(&vbus_ext->lock, "hptsleeplock", 0, LK_CANRECURSE); 1041 if (bus_dma_tag_create(NULL,/* parent */ 1042 4, /* alignment */ 1043 BUS_SPACE_MAXADDR_32BIT+1, /* boundary */ 1044 BUS_SPACE_MAXADDR, /* lowaddr */ 1045 BUS_SPACE_MAXADDR, /* highaddr */ 1046 NULL, NULL, /* filter, filterarg */ 1047 PAGE_SIZE * (os_max_sg_descriptors-1), /* maxsize */ 1048 os_max_sg_descriptors, /* nsegments */ 1049 0x10000, /* maxsegsize */ 1050 BUS_DMA_WAITOK, /* flags */ 1051 &vbus_ext->io_dmat /* tag */)) 1052 { 1053 return ; 1054 } 1055 1056 for (i=0; i<os_max_queue_comm; i++) { 1057 POS_CMDEXT ext = (POS_CMDEXT)kmalloc(sizeof(OS_CMDEXT), M_DEVBUF, M_WAITOK); 1058 if (!ext) { 1059 os_printk("Can't alloc cmdext(%d)", i); 1060 return ; 1061 } 1062 ext->vbus_ext = vbus_ext; 1063 ext->next = vbus_ext->cmdext_list; 1064 vbus_ext->cmdext_list = ext; 1065 1066 if (bus_dmamap_create(vbus_ext->io_dmat, 0, &ext->dma_map)) { 1067 os_printk("Can't create dma map(%d)", i); 1068 return ; 1069 } 1070 } 1071 1072 if ((devq = cam_simq_alloc(os_max_queue_comm)) == NULL) { 1073 os_printk("cam_simq_alloc failed"); 1074 return ; 1075 } 1076 1077 vbus_ext->sim = cam_sim_alloc(hpt_action, hpt_poll, driver_name, 1078 vbus_ext, 0, &sim_mplock, os_max_queue_comm, /*tagged*/8, devq); 1079 cam_simq_release(devq); 1080 1081 if (!vbus_ext->sim) { 1082 os_printk("cam_sim_alloc failed"); 1083 return ; 1084 } 1085 1086 if (xpt_bus_register(vbus_ext->sim, 0) != CAM_SUCCESS) { 1087 os_printk("xpt_bus_register failed"); 1088 cam_sim_free(vbus_ext->sim); 1089 vbus_ext->sim = NULL; 1090 return ; 1091 } 1092 1093 if (xpt_create_path(&vbus_ext->path, /*periph */ NULL, 1094 cam_sim_path(vbus_ext->sim), CAM_TARGET_WILDCARD, 1095 CAM_LUN_WILDCARD) != CAM_REQ_CMP) 1096 { 1097 os_printk("xpt_create_path failed"); 1098 xpt_bus_deregister(cam_sim_path(vbus_ext->sim)); 1099 cam_sim_free(vbus_ext->sim); 1100 vbus_ext->sim = NULL; 1101 return ; 1102 } 1103 1104 ccb = &xpt_alloc_ccb()->csa; 1105 xpt_setup_ccb(&ccb->ccb_h, vbus_ext->path, /*priority*/5); 1106 ccb->ccb_h.func_code = XPT_SASYNC_CB; 1107 ccb->event_enable = AC_LOST_DEVICE; 1108 ccb->callback = hpt_async; 1109 ccb->callback_arg = vbus_ext; 1110 xpt_action((union ccb *)ccb); 1111 xpt_free_ccb(&ccb->ccb_h); 1112 1113 for (hba = vbus_ext->hba_list; hba; hba = hba->next) { 1114 int rid = 0; 1115 if ((hba->irq_res = bus_alloc_resource(hba->pcidev, 1116 SYS_RES_IRQ, &rid, 0, ~0ul, 1, RF_SHAREABLE | RF_ACTIVE)) == NULL) 1117 { 1118 os_printk("can't allocate interrupt"); 1119 return ; 1120 } 1121 1122 if (bus_setup_intr(hba->pcidev, hba->irq_res, 0, 1123 hpt_pci_intr, vbus_ext, &hba->irq_handle, NULL)) 1124 { 1125 os_printk("can't set up interrupt"); 1126 return ; 1127 } 1128 hba->ldm_adapter.him->intr_control(hba->ldm_adapter.him_handle, HPT_TRUE); 1129 1130 } 1131 1132 vbus_ext->shutdown_eh = EVENTHANDLER_REGISTER(shutdown_final, 1133 hpt_shutdown_vbus, vbus_ext, SHUTDOWN_PRI_DEFAULT); 1134 if (!vbus_ext->shutdown_eh) 1135 os_printk("Shutdown event registration failed"); 1136 } 1137 1138 ldm_for_each_vbus(vbus, vbus_ext) { 1139 TASK_INIT(&vbus_ext->worker, 0, (task_fn_t *)hpt_do_tasks, vbus_ext); 1140 if (vbus_ext->tasks) 1141 TASK_ENQUEUE(&vbus_ext->worker); 1142 } 1143 1144 make_dev(&hpt_ops, DRIVER_MINOR, UID_ROOT, GID_OPERATOR, 1145 S_IRUSR | S_IWUSR, driver_name); 1146 } 1147 1148 #if defined(KLD_MODULE) 1149 1150 typedef struct driverlink *driverlink_t; 1151 struct driverlink { 1152 kobj_class_t driver; 1153 TAILQ_ENTRY(driverlink) link; /* list of drivers in devclass */ 1154 }; 1155 1156 typedef TAILQ_HEAD(driver_list, driverlink) driver_list_t; 1157 1158 struct devclass { 1159 TAILQ_ENTRY(devclass) link; 1160 devclass_t parent; /* parent in devclass hierarchy */ 1161 driver_list_t drivers; /* bus devclasses store drivers for bus */ 1162 char *name; 1163 device_t *devices; /* array of devices indexed by unit */ 1164 int maxunit; /* size of devices array */ 1165 }; 1166 1167 static void override_kernel_driver(void) 1168 { 1169 driverlink_t dl, dlfirst; 1170 driver_t *tmpdriver; 1171 devclass_t dc = devclass_find("pci"); 1172 1173 if (dc){ 1174 dlfirst = TAILQ_FIRST(&dc->drivers); 1175 for (dl = dlfirst; dl; dl = TAILQ_NEXT(dl, link)) { 1176 if(strcmp(dl->driver->name, driver_name) == 0) { 1177 tmpdriver=dl->driver; 1178 dl->driver=dlfirst->driver; 1179 dlfirst->driver=tmpdriver; 1180 break; 1181 } 1182 } 1183 } 1184 } 1185 1186 #else 1187 #define override_kernel_driver() 1188 #endif 1189 1190 static void hpt_init(void *dummy) 1191 { 1192 if (bootverbose) 1193 os_printk("%s %s", driver_name_long, driver_ver); 1194 1195 override_kernel_driver(); 1196 init_config(); 1197 1198 hpt_ich.ich_func = hpt_final_init; 1199 hpt_ich.ich_arg = NULL; 1200 hpt_ich.ich_desc = "hpt27xx"; 1201 if (config_intrhook_establish(&hpt_ich) != 0) { 1202 kprintf("%s: cannot establish configuration hook\n", 1203 driver_name_long); 1204 } 1205 1206 } 1207 SYSINIT(hptinit, SI_SUB_CONFIGURE, SI_ORDER_FIRST, hpt_init, NULL); 1208 1209 /* 1210 * CAM driver interface 1211 */ 1212 static device_method_t driver_methods[] = { 1213 /* Device interface */ 1214 DEVMETHOD(device_probe, hpt_probe), 1215 DEVMETHOD(device_attach, hpt_attach), 1216 DEVMETHOD(device_detach, hpt_detach), 1217 DEVMETHOD(device_shutdown, hpt_shutdown), 1218 DEVMETHOD_END 1219 }; 1220 1221 static driver_t hpt_pci_driver = { 1222 driver_name, 1223 driver_methods, 1224 sizeof(HBA) 1225 }; 1226 1227 static devclass_t hpt_devclass; 1228 1229 #ifndef TARGETNAME 1230 #error "no TARGETNAME found" 1231 #endif 1232 1233 /* use this to make TARGETNAME be expanded */ 1234 #define __DRIVER_MODULE(p1, p2, p3, p4, p5, p6) DRIVER_MODULE(p1, p2, p3, p4, p5, p6) 1235 #define __MODULE_VERSION(p1, p2) MODULE_VERSION(p1, p2) 1236 #define __MODULE_DEPEND(p1, p2, p3, p4, p5) MODULE_DEPEND(p1, p2, p3, p4, p5) 1237 __DRIVER_MODULE(TARGETNAME, pci, hpt_pci_driver, hpt_devclass, NULL, NULL); 1238 __MODULE_VERSION(TARGETNAME, 1); 1239 __MODULE_DEPEND(TARGETNAME, cam, 1, 1, 1); 1240 1241 typedef struct cdev * ioctl_dev_t; 1242 1243 typedef struct thread * ioctl_thread_t; 1244 1245 static int hpt_open(struct dev_open_args *ap) 1246 { 1247 return 0; 1248 } 1249 1250 static int hpt_close(struct dev_close_args *ap) 1251 { 1252 return 0; 1253 } 1254 1255 static int hpt_ioctl(struct dev_ioctl_args *ap) 1256 { 1257 u_long cmd = ap->a_cmd; 1258 caddr_t data = ap->a_data; 1259 PHPT_IOCTL_PARAM piop=(PHPT_IOCTL_PARAM)data; 1260 IOCTL_ARG ioctl_args; 1261 HPT_U32 bytesReturned; 1262 1263 switch (cmd){ 1264 case HPT_DO_IOCONTROL: 1265 { 1266 if (piop->Magic == HPT_IOCTL_MAGIC || piop->Magic == HPT_IOCTL_MAGIC32) { 1267 KdPrint(("ioctl=%x in=%p len=%d out=%p len=%d\n", 1268 piop->dwIoControlCode, 1269 piop->lpInBuffer, 1270 piop->nInBufferSize, 1271 piop->lpOutBuffer, 1272 piop->nOutBufferSize)); 1273 1274 memset(&ioctl_args, 0, sizeof(ioctl_args)); 1275 1276 ioctl_args.dwIoControlCode = piop->dwIoControlCode; 1277 ioctl_args.nInBufferSize = piop->nInBufferSize; 1278 ioctl_args.nOutBufferSize = piop->nOutBufferSize; 1279 ioctl_args.lpBytesReturned = &bytesReturned; 1280 1281 if (ioctl_args.nInBufferSize) { 1282 ioctl_args.lpInBuffer = kmalloc(ioctl_args.nInBufferSize, M_DEVBUF, M_WAITOK); 1283 if (!ioctl_args.lpInBuffer) 1284 goto invalid; 1285 if (copyin((void*)piop->lpInBuffer, 1286 ioctl_args.lpInBuffer, piop->nInBufferSize)) 1287 goto invalid; 1288 } 1289 1290 if (ioctl_args.nOutBufferSize) { 1291 ioctl_args.lpOutBuffer = kmalloc(ioctl_args.nOutBufferSize, M_DEVBUF, M_WAITOK); 1292 if (!ioctl_args.lpOutBuffer) 1293 goto invalid; 1294 } 1295 1296 get_mplock(); 1297 1298 hpt_do_ioctl(&ioctl_args); 1299 1300 rel_mplock(); 1301 1302 if (ioctl_args.result==HPT_IOCTL_RESULT_OK) { 1303 if (piop->nOutBufferSize) { 1304 if (copyout(ioctl_args.lpOutBuffer, 1305 (void*)piop->lpOutBuffer, piop->nOutBufferSize)) 1306 goto invalid; 1307 } 1308 if (piop->lpBytesReturned) { 1309 if (copyout(&bytesReturned, 1310 (void*)piop->lpBytesReturned, sizeof(HPT_U32))) 1311 goto invalid; 1312 } 1313 if (ioctl_args.lpInBuffer) kfree(ioctl_args.lpInBuffer, M_DEVBUF); 1314 if (ioctl_args.lpOutBuffer) kfree(ioctl_args.lpOutBuffer, M_DEVBUF); 1315 return 0; 1316 } 1317 invalid: 1318 if (ioctl_args.lpInBuffer) kfree(ioctl_args.lpInBuffer, M_DEVBUF); 1319 if (ioctl_args.lpOutBuffer) kfree(ioctl_args.lpOutBuffer, M_DEVBUF); 1320 return EFAULT; 1321 } 1322 return EFAULT; 1323 } 1324 1325 case HPT_SCAN_BUS: 1326 { 1327 return hpt_rescan_bus(); 1328 } 1329 default: 1330 KdPrint(("invalid command!")); 1331 return EFAULT; 1332 } 1333 1334 } 1335 1336 static int hpt_rescan_bus(void) 1337 { 1338 struct cam_path *path; 1339 union ccb *ccb; 1340 PVBUS vbus; 1341 PVBUS_EXT vbus_ext; 1342 1343 get_mplock(); 1344 1345 ldm_for_each_vbus(vbus, vbus_ext) { 1346 if (xpt_create_path(&path, xpt_periph, cam_sim_path(vbus_ext->sim), 1347 CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP) { 1348 rel_mplock(); 1349 return(EIO); 1350 } 1351 ccb = xpt_alloc_ccb(); 1352 bzero(ccb, sizeof(union ccb)); 1353 xpt_setup_ccb(&ccb->ccb_h, path, 5); 1354 ccb->ccb_h.func_code = XPT_SCAN_BUS; 1355 ccb->ccb_h.cbfcnp = hpt_bus_scan_cb; 1356 ccb->crcn.flags = CAM_FLAG_NONE; 1357 xpt_action(ccb); 1358 } 1359 1360 rel_mplock(); 1361 1362 return(0); 1363 } 1364 1365 static void hpt_bus_scan_cb(struct cam_periph *periph, union ccb *ccb) 1366 { 1367 if (ccb->ccb_h.status != CAM_REQ_CMP) 1368 KdPrint(("cam_scan_callback: failure status = %x",ccb->ccb_h.status)); 1369 else 1370 KdPrint(("Scan bus successfully!")); 1371 1372 xpt_free_path(ccb->ccb_h.path); 1373 xpt_free_ccb(&ccb->ccb_h); 1374 } 1375