1 /* $NetBSD: wdc_obio.c,v 1.22 2002/10/02 05:30:43 thorpej Exp $ */ 2 3 /*- 4 * Copyright (c) 1998 The NetBSD Foundation, Inc. 5 * All rights reserved. 6 * 7 * This code is derived from software contributed to The NetBSD Foundation 8 * by Charles M. Hannum and by Onno van der Linden. 9 * 10 * Redistribution and use in source and binary forms, with or without 11 * modification, are permitted provided that the following conditions 12 * are met: 13 * 1. Redistributions of source code must retain the above copyright 14 * notice, this list of conditions and the following disclaimer. 15 * 2. Redistributions in binary form must reproduce the above copyright 16 * notice, this list of conditions and the following disclaimer in the 17 * documentation and/or other materials provided with the distribution. 18 * 3. All advertising materials mentioning features or use of this software 19 * must display the following acknowledgement: 20 * This product includes software developed by the NetBSD 21 * Foundation, Inc. and its contributors. 22 * 4. Neither the name of The NetBSD Foundation nor the names of its 23 * contributors may be used to endorse or promote products derived 24 * from this software without specific prior written permission. 25 * 26 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS 27 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED 28 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 29 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS 30 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 31 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 32 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 33 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 34 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 35 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 36 * POSSIBILITY OF SUCH DAMAGE. 37 */ 38 39 #include <sys/param.h> 40 #include <sys/systm.h> 41 #include <sys/device.h> 42 #include <sys/malloc.h> 43 44 #include <uvm/uvm_extern.h> 45 46 #include <machine/bus.h> 47 #include <machine/autoconf.h> 48 49 #include <dev/ata/atareg.h> 50 #include <dev/ata/atavar.h> 51 #include <dev/ic/wdcvar.h> 52 53 #include <dev/ofw/openfirm.h> 54 55 #include <macppc/dev/dbdma.h> 56 57 #define WDC_REG_NPORTS 8 58 #define WDC_AUXREG_OFFSET 0x16 59 #define WDC_DEFAULT_PIO_IRQ 13 /* XXX */ 60 #define WDC_DEFAULT_DMA_IRQ 2 /* XXX */ 61 62 #define WDC_OPTIONS_DMA 0x01 63 64 /* 65 * XXX This code currently doesn't even try to allow 32-bit data port use. 66 */ 67 68 struct wdc_obio_softc { 69 struct wdc_softc sc_wdcdev; 70 struct channel_softc *wdc_chanptr; 71 struct channel_softc wdc_channel; 72 dbdma_regmap_t *sc_dmareg; 73 dbdma_command_t *sc_dmacmd; 74 u_int sc_dmaconf[2]; /* per target value of CONFIG_REG */ 75 void *sc_ih; 76 }; 77 78 int wdc_obio_probe __P((struct device *, struct cfdata *, void *)); 79 void wdc_obio_attach __P((struct device *, struct device *, void *)); 80 int wdc_obio_detach __P((struct device *, int)); 81 int wdc_obio_dma_init __P((void *, int, int, void *, size_t, int)); 82 void wdc_obio_dma_start __P((void *, int, int)); 83 int wdc_obio_dma_finish __P((void *, int, int, int)); 84 85 static void wdc_obio_select __P((struct channel_softc *, int)); 86 static void adjust_timing __P((struct channel_softc *)); 87 static void ata4_adjust_timing __P((struct channel_softc *)); 88 89 CFATTACH_DECL(wdc_obio, sizeof(struct wdc_obio_softc), 90 wdc_obio_probe, wdc_obio_attach, wdc_obio_detach, wdcactivate); 91 92 int 93 wdc_obio_probe(parent, match, aux) 94 struct device *parent; 95 struct cfdata *match; 96 void *aux; 97 { 98 struct confargs *ca = aux; 99 char compat[32]; 100 101 /* XXX should not use name */ 102 if (strcmp(ca->ca_name, "ATA") == 0 || 103 strcmp(ca->ca_name, "ata") == 0 || 104 strcmp(ca->ca_name, "ata0") == 0 || 105 strcmp(ca->ca_name, "ide") == 0) 106 return 1; 107 108 memset(compat, 0, sizeof(compat)); 109 OF_getprop(ca->ca_node, "compatible", compat, sizeof(compat)); 110 if (strcmp(compat, "heathrow-ata") == 0 || 111 strcmp(compat, "keylargo-ata") == 0) 112 return 1; 113 114 return 0; 115 } 116 117 void 118 wdc_obio_attach(parent, self, aux) 119 struct device *parent, *self; 120 void *aux; 121 { 122 struct wdc_obio_softc *sc = (void *)self; 123 struct confargs *ca = aux; 124 struct channel_softc *chp = &sc->wdc_channel; 125 int intr; 126 int use_dma = 0; 127 char path[80]; 128 129 if (sc->sc_wdcdev.sc_dev.dv_cfdata->cf_flags & WDC_OPTIONS_DMA) { 130 if (ca->ca_nreg >= 16 || ca->ca_nintr == -1) 131 use_dma = 1; /* XXX Don't work yet. */ 132 } 133 134 if (ca->ca_nintr >= 4 && ca->ca_nreg >= 8) { 135 intr = ca->ca_intr[0]; 136 printf(" irq %d", intr); 137 } else if (ca->ca_nintr == -1) { 138 intr = WDC_DEFAULT_PIO_IRQ; 139 printf(" irq property not found; using %d", intr); 140 } else { 141 printf(": couldn't get irq property\n"); 142 return; 143 } 144 145 if (use_dma) 146 printf(": DMA transfer"); 147 148 printf("\n"); 149 150 chp->cmd_iot = chp->ctl_iot = 151 macppc_make_bus_space_tag(ca->ca_baseaddr + ca->ca_reg[0], 4); 152 153 if (bus_space_map(chp->cmd_iot, 0, WDC_REG_NPORTS, 0, &chp->cmd_ioh) || 154 bus_space_subregion(chp->cmd_iot, chp->cmd_ioh, 155 WDC_AUXREG_OFFSET, 1, &chp->ctl_ioh)) { 156 printf("%s: couldn't map registers\n", 157 sc->sc_wdcdev.sc_dev.dv_xname); 158 return; 159 } 160 #if 0 161 chp->data32iot = chp->cmd_iot; 162 chp->data32ioh = chp->cmd_ioh; 163 #endif 164 165 sc->sc_ih = intr_establish(intr, IST_LEVEL, IPL_BIO, wdcintr, chp); 166 167 if (use_dma) { 168 sc->sc_dmacmd = dbdma_alloc(sizeof(dbdma_command_t) * 20); 169 sc->sc_dmareg = mapiodev(ca->ca_baseaddr + ca->ca_reg[2], 170 ca->ca_reg[3]); 171 sc->sc_wdcdev.cap |= WDC_CAPABILITY_DMA; 172 sc->sc_wdcdev.DMA_cap = 2; 173 if (strcmp(ca->ca_name, "ata-4") == 0) { 174 sc->sc_wdcdev.cap |= WDC_CAPABILITY_UDMA; 175 sc->sc_wdcdev.UDMA_cap = 4; 176 sc->sc_wdcdev.set_modes = ata4_adjust_timing; 177 } else { 178 sc->sc_wdcdev.set_modes = adjust_timing; 179 } 180 #ifdef notyet 181 /* Minimum cycle time is 150ns (DMA MODE 1) on ohare. */ 182 if (ohare) { 183 sc->sc_wdcdev.PIO_cap = 3; 184 sc->sc_wdcdev.DMA_cap = 1; 185 } 186 #endif 187 } else { 188 /* all non-dma controllers can use adjust_timing */ 189 sc->sc_wdcdev.set_modes = adjust_timing; 190 } 191 192 sc->sc_wdcdev.PIO_cap = 4; 193 sc->sc_wdcdev.cap |= WDC_CAPABILITY_DATA16 | WDC_CAPABILITY_MODE; 194 sc->wdc_chanptr = chp; 195 sc->sc_wdcdev.channels = &sc->wdc_chanptr; 196 sc->sc_wdcdev.nchannels = 1; 197 sc->sc_wdcdev.dma_arg = sc; 198 sc->sc_wdcdev.dma_init = wdc_obio_dma_init; 199 sc->sc_wdcdev.dma_start = wdc_obio_dma_start; 200 sc->sc_wdcdev.dma_finish = wdc_obio_dma_finish; 201 chp->channel = 0; 202 chp->wdc = &sc->sc_wdcdev; 203 chp->ch_queue = malloc(sizeof(struct channel_queue), 204 M_DEVBUF, M_NOWAIT); 205 if (chp->ch_queue == NULL) { 206 printf("%s: can't allocate memory for command queue", 207 sc->sc_wdcdev.sc_dev.dv_xname); 208 return; 209 } 210 211 #define OHARE_FEATURE_REG 0xf3000038 212 213 /* XXX Enable wdc1 by feature reg. */ 214 memset(path, 0, sizeof(path)); 215 OF_package_to_path(ca->ca_node, path, sizeof(path)); 216 if (strcmp(path, "/bandit@F2000000/ohare@10/ata@21000") == 0) { 217 u_int x; 218 219 x = in32rb(OHARE_FEATURE_REG); 220 x |= 8; 221 out32rb(OHARE_FEATURE_REG, x); 222 } 223 224 wdcattach(chp); 225 sc->sc_wdcdev.set_modes(chp); 226 227 } 228 229 /* Multiword DMA transfer timings */ 230 struct ide_timings { 231 int cycle; /* minimum cycle time [ns] */ 232 int active; /* minimum command active time [ns] */ 233 }; 234 static struct ide_timings pio_timing[5] = { 235 { 600, 180 }, /* Mode 0 */ 236 { 390, 150 }, /* 1 */ 237 { 240, 105 }, /* 2 */ 238 { 180, 90 }, /* 3 */ 239 { 120, 75 } /* 4 */ 240 }; 241 static struct ide_timings dma_timing[3] = { 242 { 480, 240 }, /* Mode 0 */ 243 { 165, 90 }, /* Mode 1 */ 244 { 120, 75 } /* Mode 2 */ 245 }; 246 247 static struct ide_timings udma_timing[5] = { 248 {120, 180}, /* Mode 0 */ 249 { 90, 150}, /* Mode 1 */ 250 { 60, 120}, /* Mode 2 */ 251 { 45, 90}, /* Mode 3 */ 252 { 30, 90} /* Mode 4 */ 253 }; 254 255 #define TIME_TO_TICK(time) howmany((time), 30) 256 #define PIO_REC_OFFSET 4 257 #define PIO_REC_MIN 1 258 #define PIO_ACT_MIN 1 259 #define DMA_REC_OFFSET 1 260 #define DMA_REC_MIN 1 261 #define DMA_ACT_MIN 1 262 263 #define ATA4_TIME_TO_TICK(time) howmany((time), 15) /* 15 ns clock */ 264 265 #define CONFIG_REG (0x200 >> 4) /* IDE access timing register */ 266 267 void 268 wdc_obio_select(chp, drive) 269 struct channel_softc *chp; 270 int drive; 271 { 272 struct wdc_obio_softc *sc = (struct wdc_obio_softc *)chp->wdc; 273 bus_space_write_4(chp->cmd_iot, chp->cmd_ioh, 274 CONFIG_REG, sc->sc_dmaconf[drive]); 275 } 276 277 void 278 adjust_timing(chp) 279 struct channel_softc *chp; 280 { 281 struct wdc_obio_softc *sc = (struct wdc_obio_softc *)chp->wdc; 282 int drive; 283 int min_cycle, min_active; 284 int cycle_tick, act_tick, inact_tick, half_tick; 285 286 for (drive = 0; drive < 2; drive++) { 287 u_int conf = 0; 288 struct ata_drive_datas *drvp; 289 290 drvp = &chp->ch_drive[drive]; 291 /* set up pio mode timings */ 292 if (drvp->drive_flags & DRIVE) { 293 int piomode = drvp->PIO_mode; 294 min_cycle = pio_timing[piomode].cycle; 295 min_active = pio_timing[piomode].active; 296 297 cycle_tick = TIME_TO_TICK(min_cycle); 298 act_tick = TIME_TO_TICK(min_active); 299 if (act_tick < PIO_ACT_MIN) 300 act_tick = PIO_ACT_MIN; 301 inact_tick = cycle_tick - act_tick - PIO_REC_OFFSET; 302 if (inact_tick < PIO_REC_MIN) 303 inact_tick = PIO_REC_MIN; 304 /* mask: 0x000007ff */ 305 conf |= (inact_tick << 5) | act_tick; 306 } 307 /* Set up dma mode timings */ 308 if (drvp->drive_flags & DRIVE_DMA) { 309 int dmamode = drvp->DMA_mode; 310 min_cycle = dma_timing[dmamode].cycle; 311 min_active = dma_timing[dmamode].active; 312 cycle_tick = TIME_TO_TICK(min_cycle); 313 act_tick = TIME_TO_TICK(min_active); 314 inact_tick = cycle_tick - act_tick - DMA_REC_OFFSET; 315 if (inact_tick < DMA_REC_MIN) 316 inact_tick = DMA_REC_MIN; 317 half_tick = 0; /* XXX */ 318 /* mask: 0xfffff800 */ 319 conf |= 320 (half_tick << 21) | 321 (inact_tick << 16) | (act_tick << 11); 322 } 323 #ifdef DEBUG 324 if (conf) { 325 printf("conf[%d] = 0x%x, cyc = %d (%d ns), act = %d (%d ns), inact = %d\n", 326 drive, conf, cycle_tick, min_cycle, act_tick, min_active, inact_tick); 327 } 328 #endif 329 sc->sc_dmaconf[drive] = conf; 330 } 331 sc->sc_wdcdev.cap &= ~WDC_CAPABILITY_SELECT; 332 sc->sc_wdcdev.select = 0; 333 if (sc->sc_dmaconf[0]) { 334 wdc_obio_select(chp,0); 335 if (sc->sc_dmaconf[1] && (sc->sc_dmaconf[0] != sc->sc_dmaconf[1])) { 336 sc->sc_wdcdev.select = wdc_obio_select; 337 sc->sc_wdcdev.cap |= WDC_CAPABILITY_SELECT; 338 } 339 } else if (sc->sc_dmaconf[1]) { 340 wdc_obio_select(chp,1); 341 } 342 wdc_print_modes(chp); 343 } 344 345 void 346 ata4_adjust_timing(chp) 347 struct channel_softc *chp; 348 { 349 struct wdc_obio_softc *sc = (struct wdc_obio_softc *)chp->wdc; 350 int drive; 351 int min_cycle, min_active; 352 int cycle_tick, act_tick, inact_tick; 353 354 for (drive = 0; drive < 2; drive++) { 355 u_int conf = 0; 356 struct ata_drive_datas *drvp; 357 358 drvp = &chp->ch_drive[drive]; 359 /* set up pio mode timings */ 360 361 if (drvp->drive_flags & DRIVE) { 362 int piomode = drvp->PIO_mode; 363 min_cycle = pio_timing[piomode].cycle; 364 min_active = pio_timing[piomode].active; 365 366 cycle_tick = ATA4_TIME_TO_TICK(min_cycle); 367 act_tick = ATA4_TIME_TO_TICK(min_active); 368 inact_tick = cycle_tick - act_tick; 369 /* mask: 0x000003ff */ 370 conf |= (inact_tick << 5) | act_tick; 371 } 372 /* set up dma mode timings */ 373 if (drvp->drive_flags & DRIVE_DMA) { 374 int dmamode = drvp->DMA_mode; 375 min_cycle = dma_timing[dmamode].cycle; 376 min_active = dma_timing[dmamode].active; 377 cycle_tick = ATA4_TIME_TO_TICK(min_cycle); 378 act_tick = ATA4_TIME_TO_TICK(min_active); 379 inact_tick = cycle_tick - act_tick; 380 /* mask: 0x001ffc00 */ 381 conf |= (act_tick << 10) | (inact_tick << 15); 382 } 383 /* set up udma mode timings */ 384 if (drvp->drive_flags & DRIVE_UDMA) { 385 int udmamode = drvp->UDMA_mode; 386 min_cycle = udma_timing[udmamode].cycle; 387 min_active = udma_timing[udmamode].active; 388 act_tick = ATA4_TIME_TO_TICK(min_active); 389 cycle_tick = ATA4_TIME_TO_TICK(min_cycle); 390 /* mask: 0x1ff00000 */ 391 conf |= (cycle_tick << 21) | (act_tick << 25) | 0x100000; 392 } 393 #ifdef DEBUG 394 if (conf) { 395 printf("ata4 conf[%d] = 0x%x, cyc = %d (%d ns), act = %d (%d ns), inact = %d\n", 396 drive, conf, cycle_tick, min_cycle, act_tick, min_active, inact_tick); 397 } 398 #endif 399 sc->sc_dmaconf[drive] = conf; 400 } 401 sc->sc_wdcdev.cap &= ~WDC_CAPABILITY_SELECT; 402 sc->sc_wdcdev.select = 0; 403 if (sc->sc_dmaconf[0]) { 404 wdc_obio_select(chp,0); 405 if (sc->sc_dmaconf[1] && (sc->sc_dmaconf[0] != sc->sc_dmaconf[1])) { 406 sc->sc_wdcdev.select = wdc_obio_select; 407 sc->sc_wdcdev.cap |= WDC_CAPABILITY_SELECT; 408 } 409 } else if (sc->sc_dmaconf[1]) { 410 wdc_obio_select(chp,1); 411 } 412 wdc_print_modes(chp); 413 } 414 415 int 416 wdc_obio_detach(self, flags) 417 struct device *self; 418 int flags; 419 { 420 struct wdc_obio_softc *sc = (void *)self; 421 int error; 422 423 if ((error = wdcdetach(self, flags)) != 0) 424 return error; 425 426 intr_disestablish(sc->sc_ih); 427 428 free(sc->wdc_channel.ch_queue, M_DEVBUF); 429 430 /* Unmap our i/o space. */ 431 bus_space_unmap(chp->cmd_iot, chp->cmd_ioh, WDC_REG_NPORTS); 432 433 /* Unmap DMA registers. */ 434 /* XXX unmapiodev(sc->sc_dmareg); */ 435 /* XXX free(sc->sc_dmacmd); */ 436 437 return 0; 438 } 439 440 int 441 wdc_obio_dma_init(v, channel, drive, databuf, datalen, read) 442 void *v; 443 void *databuf; 444 size_t datalen; 445 int read; 446 { 447 struct wdc_obio_softc *sc = v; 448 vaddr_t va = (vaddr_t)databuf; 449 dbdma_command_t *cmdp; 450 u_int cmd, offset; 451 452 cmdp = sc->sc_dmacmd; 453 cmd = read ? DBDMA_CMD_IN_MORE : DBDMA_CMD_OUT_MORE; 454 455 offset = va & PGOFSET; 456 457 /* if va is not page-aligned, setup the first page */ 458 if (offset != 0) { 459 int rest = NBPG - offset; /* the rest of the page */ 460 461 if (datalen > rest) { /* if continues to next page */ 462 DBDMA_BUILD(cmdp, cmd, 0, rest, vtophys(va), 463 DBDMA_INT_NEVER, DBDMA_WAIT_NEVER, 464 DBDMA_BRANCH_NEVER); 465 datalen -= rest; 466 va += rest; 467 cmdp++; 468 } 469 } 470 471 /* now va is page-aligned */ 472 while (datalen > NBPG) { 473 DBDMA_BUILD(cmdp, cmd, 0, NBPG, vtophys(va), 474 DBDMA_INT_NEVER, DBDMA_WAIT_NEVER, DBDMA_BRANCH_NEVER); 475 datalen -= NBPG; 476 va += NBPG; 477 cmdp++; 478 } 479 480 /* the last page (datalen <= NBPG here) */ 481 cmd = read ? DBDMA_CMD_IN_LAST : DBDMA_CMD_OUT_LAST; 482 DBDMA_BUILD(cmdp, cmd, 0, datalen, vtophys(va), 483 DBDMA_INT_NEVER, DBDMA_WAIT_NEVER, DBDMA_BRANCH_NEVER); 484 cmdp++; 485 486 DBDMA_BUILD(cmdp, DBDMA_CMD_STOP, 0, 0, 0, 487 DBDMA_INT_NEVER, DBDMA_WAIT_NEVER, DBDMA_BRANCH_NEVER); 488 489 return 0; 490 } 491 492 void 493 wdc_obio_dma_start(v, channel, drive) 494 void *v; 495 int channel, drive; 496 { 497 struct wdc_obio_softc *sc = v; 498 499 dbdma_start(sc->sc_dmareg, sc->sc_dmacmd); 500 } 501 502 int 503 wdc_obio_dma_finish(v, channel, drive, read) 504 void *v; 505 int channel, drive; 506 int read; 507 { 508 struct wdc_obio_softc *sc = v; 509 510 dbdma_stop(sc->sc_dmareg); 511 return 0; 512 } 513