1 /* $NetBSD: isp_sbus.c,v 1.70 2007/10/19 12:01:11 ad Exp $ */ 2 /* 3 * SBus specific probe and attach routines for Qlogic ISP SCSI adapters. 4 * 5 * Copyright (C) 1997, 1998, 1999 National Aeronautics & Space Administration 6 * All rights reserved. 7 * 8 * Additional Copyright (C) 2000-2007 by Matthew Jacob 9 * All rights reserved. 10 * 11 * Redistribution and use in source and binary forms, with or without 12 * modification, are permitted provided that the following conditions 13 * are met: 14 * 1. Redistributions of source code must retain the above copyright 15 * notice, this list of conditions and the following disclaimer. 16 * 2. Redistributions in binary form must reproduce the above copyright 17 * notice, this list of conditions and the following disclaimer in the 18 * documentation and/or other materials provided with the distribution. 19 * 3. The name of the author may not be used to endorse or promote products 20 * derived from this software without specific prior written permission 21 * 22 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 23 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 24 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 25 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, 26 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT 27 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 28 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 29 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 30 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 31 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 32 * 33 */ 34 35 #include <sys/cdefs.h> 36 __KERNEL_RCSID(0, "$NetBSD: isp_sbus.c,v 1.70 2007/10/19 12:01:11 ad Exp $"); 37 38 #include <sys/param.h> 39 #include <sys/systm.h> 40 #include <sys/device.h> 41 #include <sys/kernel.h> 42 #include <sys/malloc.h> 43 #include <sys/queue.h> 44 #include <dev/ic/isp_netbsd.h> 45 #include <sys/intr.h> 46 #include <machine/autoconf.h> 47 #include <dev/sbus/sbusvar.h> 48 #include <sys/reboot.h> 49 50 static void isp_sbus_reset0(struct ispsoftc *); 51 static void isp_sbus_reset1(struct ispsoftc *); 52 static int isp_sbus_intr(void *); 53 static int 54 isp_sbus_rd_isr(struct ispsoftc *, uint32_t *, uint16_t *, uint16_t *); 55 static uint32_t isp_sbus_rd_reg(struct ispsoftc *, int); 56 static void isp_sbus_wr_reg (struct ispsoftc *, int, uint32_t); 57 static int isp_sbus_mbxdma(struct ispsoftc *); 58 static int isp_sbus_dmasetup(struct ispsoftc *, XS_T *, ispreq_t *, uint32_t *, 59 uint32_t); 60 static void isp_sbus_dmateardown(struct ispsoftc *, XS_T *, uint32_t); 61 62 #ifndef ISP_DISABLE_FW 63 #include <dev/microcode/isp/asm_sbus.h> 64 #else 65 #define ISP_1000_RISC_CODE NULL 66 #endif 67 68 static const struct ispmdvec mdvec = { 69 isp_sbus_rd_isr, 70 isp_sbus_rd_reg, 71 isp_sbus_wr_reg, 72 isp_sbus_mbxdma, 73 isp_sbus_dmasetup, 74 isp_sbus_dmateardown, 75 isp_sbus_reset0, 76 isp_sbus_reset1, 77 NULL, 78 ISP_1000_RISC_CODE, 79 0, 80 0 81 }; 82 83 struct isp_sbussoftc { 84 struct ispsoftc sbus_isp; 85 struct sbusdev sbus_sd; 86 sdparam sbus_dev; 87 bus_space_tag_t sbus_bustag; 88 bus_space_handle_t sbus_reg; 89 int sbus_node; 90 int sbus_pri; 91 struct ispmdvec sbus_mdvec; 92 bus_dmamap_t *sbus_dmamap; 93 int16_t sbus_poff[_NREG_BLKS]; 94 }; 95 96 97 static int isp_match(struct device *, struct cfdata *, void *); 98 static void isp_sbus_attach(struct device *, struct device *, void *); 99 CFATTACH_DECL(isp_sbus, sizeof (struct isp_sbussoftc), 100 isp_match, isp_sbus_attach, NULL, NULL); 101 102 static int 103 isp_match(struct device *parent, struct cfdata *cf, void *aux) 104 { 105 int rv; 106 #ifdef DEBUG 107 static int oneshot = 1; 108 #endif 109 struct sbus_attach_args *sa = aux; 110 111 rv = (strcmp(cf->cf_name, sa->sa_name) == 0 || 112 strcmp("PTI,ptisp", sa->sa_name) == 0 || 113 strcmp("ptisp", sa->sa_name) == 0 || 114 strcmp("SUNW,isp", sa->sa_name) == 0 || 115 strcmp("QLGC,isp", sa->sa_name) == 0); 116 #ifdef DEBUG 117 if (rv && oneshot) { 118 oneshot = 0; 119 printf("Qlogic ISP Driver, NetBSD (sbus) Platform Version " 120 "%d.%d Core Version %d.%d\n", 121 ISP_PLATFORM_VERSION_MAJOR, ISP_PLATFORM_VERSION_MINOR, 122 ISP_CORE_VERSION_MAJOR, ISP_CORE_VERSION_MINOR); 123 } 124 #endif 125 return (rv); 126 } 127 128 129 static void 130 isp_sbus_attach(struct device *parent, struct device *self, void *aux) 131 { 132 int freq, ispburst, sbusburst; 133 struct sbus_attach_args *sa = aux; 134 struct isp_sbussoftc *sbc = (struct isp_sbussoftc *) self; 135 struct ispsoftc *isp = &sbc->sbus_isp; 136 137 printf(" for %s\n", sa->sa_name); 138 139 sbc->sbus_bustag = sa->sa_bustag; 140 if (sa->sa_nintr != 0) 141 sbc->sbus_pri = sa->sa_pri; 142 sbc->sbus_mdvec = mdvec; 143 144 if (sa->sa_npromvaddrs) { 145 sbus_promaddr_to_handle(sa->sa_bustag, 146 sa->sa_promvaddrs[0], &sbc->sbus_reg); 147 } else { 148 if (sbus_bus_map(sa->sa_bustag, sa->sa_slot, sa->sa_offset, 149 sa->sa_size, 0, &sbc->sbus_reg) != 0) { 150 printf("%s: cannot map registers\n", self->dv_xname); 151 return; 152 } 153 } 154 sbc->sbus_node = sa->sa_node; 155 156 freq = prom_getpropint(sa->sa_node, "clock-frequency", 0); 157 if (freq) { 158 /* 159 * Convert from HZ to MHz, rounding up. 160 */ 161 freq = (freq + 500000)/1000000; 162 } 163 sbc->sbus_mdvec.dv_clock = freq; 164 165 /* 166 * Now figure out what the proper burst sizes, etc., to use. 167 * Unfortunately, there is no ddi_dma_burstsizes here which 168 * walks up the tree finding the limiting burst size node (if 169 * any). 170 */ 171 sbusburst = ((struct sbus_softc *)parent)->sc_burst; 172 if (sbusburst == 0) 173 sbusburst = SBUS_BURST_32 - 1; 174 ispburst = prom_getpropint(sa->sa_node, "burst-sizes", -1); 175 if (ispburst == -1) { 176 ispburst = sbusburst; 177 } 178 ispburst &= sbusburst; 179 ispburst &= ~(1 << 7); 180 ispburst &= ~(1 << 6); 181 sbc->sbus_mdvec.dv_conf1 = 0; 182 if (ispburst & (1 << 5)) { 183 sbc->sbus_mdvec.dv_conf1 = BIU_SBUS_CONF1_FIFO_32; 184 } else if (ispburst & (1 << 4)) { 185 sbc->sbus_mdvec.dv_conf1 = BIU_SBUS_CONF1_FIFO_16; 186 } else if (ispburst & (1 << 3)) { 187 sbc->sbus_mdvec.dv_conf1 = 188 BIU_SBUS_CONF1_BURST8 | BIU_SBUS_CONF1_FIFO_8; 189 } 190 if (sbc->sbus_mdvec.dv_conf1) { 191 sbc->sbus_mdvec.dv_conf1 |= BIU_BURST_ENABLE; 192 } 193 194 isp->isp_mdvec = &sbc->sbus_mdvec; 195 isp->isp_bustype = ISP_BT_SBUS; 196 isp->isp_type = ISP_HA_SCSI_UNKNOWN; 197 isp->isp_param = &sbc->sbus_dev; 198 isp->isp_dmatag = sa->sa_dmatag; 199 MEMZERO(isp->isp_param, sizeof (sdparam)); 200 201 sbc->sbus_poff[BIU_BLOCK >> _BLK_REG_SHFT] = BIU_REGS_OFF; 202 sbc->sbus_poff[MBOX_BLOCK >> _BLK_REG_SHFT] = SBUS_MBOX_REGS_OFF; 203 sbc->sbus_poff[SXP_BLOCK >> _BLK_REG_SHFT] = SBUS_SXP_REGS_OFF; 204 sbc->sbus_poff[RISC_BLOCK >> _BLK_REG_SHFT] = SBUS_RISC_REGS_OFF; 205 sbc->sbus_poff[DMA_BLOCK >> _BLK_REG_SHFT] = DMA_REGS_OFF; 206 207 /* Establish interrupt channel */ 208 bus_intr_establish(sbc->sbus_bustag, sbc->sbus_pri, IPL_BIO, 209 isp_sbus_intr, sbc); 210 sbus_establish(&sbc->sbus_sd, &sbc->sbus_isp.isp_osinfo._dev); 211 212 /* 213 * Set up logging levels. 214 */ 215 #ifdef ISP_LOGDEFAULT 216 isp->isp_dblev = ISP_LOGDEFAULT; 217 #else 218 isp->isp_dblev = ISP_LOGWARN|ISP_LOGERR; 219 if (bootverbose) 220 isp->isp_dblev |= ISP_LOGCONFIG|ISP_LOGINFO; 221 #ifdef SCSIDEBUG 222 isp->isp_dblev |= ISP_LOGDEBUG1|ISP_LOGDEBUG2; 223 #endif 224 #ifdef DEBUG 225 isp->isp_dblev |= ISP_LOGDEBUG0; 226 #endif 227 #endif 228 229 isp->isp_confopts = device_cfdata(self)->cf_flags; 230 isp->isp_role = ISP_DEFAULT_ROLES; 231 232 /* 233 * There's no tool on sparc to set NVRAM for ISPs, so ignore it. 234 */ 235 isp->isp_confopts |= ISP_CFG_NONVRAM; 236 237 /* 238 * Mark things if we're a PTI SBus adapter. 239 */ 240 if (strcmp("PTI,ptisp", sa->sa_name) == 0 || 241 strcmp("ptisp", sa->sa_name) == 0) { 242 SDPARAM(isp)->isp_ptisp = 1; 243 } 244 ISP_LOCK(isp); 245 isp_reset(isp); 246 if (isp->isp_state != ISP_RESETSTATE) { 247 ISP_UNLOCK(isp); 248 return; 249 } 250 ISP_ENABLE_INTS(isp); 251 isp_init(isp); 252 if (isp->isp_state != ISP_INITSTATE) { 253 isp_uninit(isp); 254 ISP_UNLOCK(isp); 255 return; 256 } 257 258 /* 259 * do generic attach. 260 */ 261 ISP_UNLOCK(isp); 262 isp_attach(isp); 263 if (isp->isp_state != ISP_RUNSTATE) { 264 ISP_LOCK(isp); 265 isp_uninit(isp); 266 ISP_UNLOCK(isp); 267 } 268 } 269 270 271 static void 272 isp_sbus_reset0(struct ispsoftc *isp) 273 { 274 ISP_DISABLE_INTS(isp); 275 } 276 277 static void 278 isp_sbus_reset1(struct ispsoftc *isp) 279 { 280 ISP_ENABLE_INTS(isp); 281 } 282 283 static int 284 isp_sbus_intr(void *arg) 285 { 286 uint32_t isr; 287 uint16_t sema, mbox; 288 struct ispsoftc *isp = arg; 289 290 if (ISP_READ_ISR(isp, &isr, &sema, &mbox) == 0) { 291 isp->isp_intbogus++; 292 return (0); 293 } else { 294 struct isp_sbussoftc *sbc = arg; 295 sbc->sbus_isp.isp_osinfo.onintstack = 1; 296 isp_intr(isp, isr, sema, mbox); 297 sbc->sbus_isp.isp_osinfo.onintstack = 0; 298 return (1); 299 } 300 } 301 302 #define IspVirt2Off(a, x) \ 303 (((struct isp_sbussoftc *)a)->sbus_poff[((x) & _BLK_REG_MASK) >> \ 304 _BLK_REG_SHFT] + ((x) & 0xff)) 305 306 #define BXR2(sbc, off) \ 307 bus_space_read_2(sbc->sbus_bustag, sbc->sbus_reg, off) 308 309 static int 310 isp_sbus_rd_isr(struct ispsoftc *isp, uint32_t *isrp, 311 uint16_t *semap, uint16_t *mbp) 312 { 313 struct isp_sbussoftc *sbc = (struct isp_sbussoftc *) isp; 314 uint32_t isr; 315 uint16_t sema; 316 317 isr = BXR2(sbc, IspVirt2Off(isp, BIU_ISR)); 318 sema = BXR2(sbc, IspVirt2Off(isp, BIU_SEMA)); 319 isp_prt(isp, ISP_LOGDEBUG3, "ISR 0x%x SEMA 0x%x", isr, sema); 320 isr &= INT_PENDING_MASK(isp); 321 sema &= BIU_SEMA_LOCK; 322 if (isr == 0 && sema == 0) { 323 return (0); 324 } 325 *isrp = isr; 326 if ((*semap = sema) != 0) { 327 *mbp = BXR2(sbc, IspVirt2Off(isp, OUTMAILBOX0)); 328 } 329 return (1); 330 } 331 332 static uint32_t 333 isp_sbus_rd_reg(struct ispsoftc *isp, int regoff) 334 { 335 struct isp_sbussoftc *sbc = (struct isp_sbussoftc *) isp; 336 int offset = sbc->sbus_poff[(regoff & _BLK_REG_MASK) >> _BLK_REG_SHFT]; 337 offset += (regoff & 0xff); 338 return (bus_space_read_2(sbc->sbus_bustag, sbc->sbus_reg, offset)); 339 } 340 341 static void 342 isp_sbus_wr_reg(struct ispsoftc *isp, int regoff, uint32_t val) 343 { 344 struct isp_sbussoftc *sbc = (struct isp_sbussoftc *) isp; 345 int offset = sbc->sbus_poff[(regoff & _BLK_REG_MASK) >> _BLK_REG_SHFT]; 346 offset += (regoff & 0xff); 347 bus_space_write_2(sbc->sbus_bustag, sbc->sbus_reg, offset, val); 348 } 349 350 static int 351 isp_sbus_mbxdma(struct ispsoftc *isp) 352 { 353 struct isp_sbussoftc *sbc = (struct isp_sbussoftc *) isp; 354 bus_dma_segment_t reqseg, rspseg; 355 int reqrs, rsprs, i, progress; 356 size_t n; 357 bus_size_t len; 358 359 if (isp->isp_rquest_dma) 360 return (0); 361 362 n = isp->isp_maxcmds * sizeof (XS_T *); 363 isp->isp_xflist = (XS_T **) malloc(n, M_DEVBUF, M_WAITOK); 364 if (isp->isp_xflist == NULL) { 365 isp_prt(isp, ISP_LOGERR, "cannot alloc xflist array"); 366 return (1); 367 } 368 MEMZERO(isp->isp_xflist, n); 369 n = sizeof (bus_dmamap_t) * isp->isp_maxcmds; 370 sbc->sbus_dmamap = (bus_dmamap_t *) malloc(n, M_DEVBUF, M_WAITOK); 371 if (sbc->sbus_dmamap == NULL) { 372 free(isp->isp_xflist, M_DEVBUF); 373 isp->isp_xflist = NULL; 374 isp_prt(isp, ISP_LOGERR, "cannot alloc dmamap array"); 375 return (1); 376 } 377 for (i = 0; i < isp->isp_maxcmds; i++) { 378 /* Allocate a DMA handle */ 379 if (bus_dmamap_create(isp->isp_dmatag, MAXPHYS, 1, MAXPHYS, 380 1 << 24, BUS_DMA_NOWAIT, &sbc->sbus_dmamap[i]) != 0) { 381 isp_prt(isp, ISP_LOGERR, "cmd DMA maps create error"); 382 break; 383 } 384 } 385 if (i < isp->isp_maxcmds) { 386 while (--i >= 0) { 387 bus_dmamap_destroy(isp->isp_dmatag, 388 sbc->sbus_dmamap[i]); 389 } 390 free(isp->isp_xflist, M_DEVBUF); 391 free(sbc->sbus_dmamap, M_DEVBUF); 392 isp->isp_xflist = NULL; 393 sbc->sbus_dmamap = NULL; 394 return (1); 395 } 396 397 /* 398 * Allocate and map the request and response queues 399 */ 400 progress = 0; 401 len = ISP_QUEUE_SIZE(RQUEST_QUEUE_LEN(isp)); 402 if (bus_dmamem_alloc(isp->isp_dmatag, len, 0, 0, &reqseg, 1, &reqrs, 403 BUS_DMA_NOWAIT)) { 404 goto dmafail; 405 } 406 progress++; 407 if (bus_dmamem_map(isp->isp_dmatag, &reqseg, reqrs, len, 408 (void *)&isp->isp_rquest, BUS_DMA_NOWAIT|BUS_DMA_COHERENT)) { 409 goto dmafail; 410 } 411 progress++; 412 if (bus_dmamap_create(isp->isp_dmatag, len, 1, len, 1 << 24, 413 BUS_DMA_NOWAIT, &isp->isp_rqdmap) != 0) { 414 goto dmafail; 415 } 416 progress++; 417 if (bus_dmamap_load(isp->isp_dmatag, isp->isp_rqdmap, 418 isp->isp_rquest, len, NULL, BUS_DMA_NOWAIT) != 0) { 419 goto dmafail; 420 } 421 progress++; 422 isp->isp_rquest_dma = isp->isp_rqdmap->dm_segs[0].ds_addr; 423 424 len = ISP_QUEUE_SIZE(RESULT_QUEUE_LEN(isp)); 425 if (bus_dmamem_alloc(isp->isp_dmatag, len, 0, 0, &rspseg, 1, &rsprs, 426 BUS_DMA_NOWAIT)) { 427 goto dmafail; 428 } 429 progress++; 430 if (bus_dmamem_map(isp->isp_dmatag, &rspseg, rsprs, len, 431 (void *)&isp->isp_result, BUS_DMA_NOWAIT|BUS_DMA_COHERENT)) { 432 goto dmafail; 433 } 434 progress++; 435 if (bus_dmamap_create(isp->isp_dmatag, len, 1, len, 1 << 24, 436 BUS_DMA_NOWAIT, &isp->isp_rsdmap) != 0) { 437 goto dmafail; 438 } 439 progress++; 440 if (bus_dmamap_load(isp->isp_dmatag, isp->isp_rsdmap, 441 isp->isp_result, len, NULL, BUS_DMA_NOWAIT) != 0) { 442 goto dmafail; 443 } 444 isp->isp_result_dma = isp->isp_rsdmap->dm_segs[0].ds_addr; 445 446 return (0); 447 448 dmafail: 449 isp_prt(isp, ISP_LOGERR, "Mailbox DMA Setup Failure"); 450 451 if (progress >= 8) { 452 bus_dmamap_unload(isp->isp_dmatag, isp->isp_rsdmap); 453 } 454 if (progress >= 7) { 455 bus_dmamap_destroy(isp->isp_dmatag, isp->isp_rsdmap); 456 } 457 if (progress >= 6) { 458 bus_dmamem_unmap(isp->isp_dmatag, 459 isp->isp_result, ISP_QUEUE_SIZE(RESULT_QUEUE_LEN(isp))); 460 } 461 if (progress >= 5) { 462 bus_dmamem_free(isp->isp_dmatag, &rspseg, rsprs); 463 } 464 465 if (progress >= 4) { 466 bus_dmamap_unload(isp->isp_dmatag, isp->isp_rqdmap); 467 } 468 if (progress >= 3) { 469 bus_dmamap_destroy(isp->isp_dmatag, isp->isp_rqdmap); 470 } 471 if (progress >= 2) { 472 bus_dmamem_unmap(isp->isp_dmatag, 473 isp->isp_rquest, ISP_QUEUE_SIZE(RQUEST_QUEUE_LEN(isp))); 474 } 475 if (progress >= 1) { 476 bus_dmamem_free(isp->isp_dmatag, &reqseg, reqrs); 477 } 478 479 for (i = 0; i < isp->isp_maxcmds; i++) { 480 bus_dmamap_destroy(isp->isp_dmatag, sbc->sbus_dmamap[i]); 481 } 482 free(sbc->sbus_dmamap, M_DEVBUF); 483 free(isp->isp_xflist, M_DEVBUF); 484 isp->isp_xflist = NULL; 485 sbc->sbus_dmamap = NULL; 486 return (1); 487 } 488 489 /* 490 * Map a DMA request. 491 * We're guaranteed that rq->req_handle is a value from 1 to isp->isp_maxcmds. 492 */ 493 494 static int 495 isp_sbus_dmasetup(struct ispsoftc *isp, XS_T *xs, ispreq_t *rq, 496 uint32_t *nxtip, uint32_t optr) 497 { 498 struct isp_sbussoftc *sbc = (struct isp_sbussoftc *) isp; 499 bus_dmamap_t dmap; 500 ispreq_t *qep; 501 int error, cansleep = (xs->xs_control & XS_CTL_NOSLEEP) == 0; 502 int in = (xs->xs_control & XS_CTL_DATA_IN) != 0; 503 504 qep = (ispreq_t *) ISP_QUEUE_ENTRY(isp->isp_rquest, isp->isp_reqidx); 505 if (xs->datalen == 0) { 506 rq->req_seg_count = 1; 507 goto mbxsync; 508 } 509 510 dmap = sbc->sbus_dmamap[isp_handle_index(rq->req_handle)]; 511 if (dmap->dm_nsegs != 0) { 512 panic("%s: DMA map already allocated", isp->isp_name); 513 /* NOTREACHED */ 514 } 515 error = bus_dmamap_load(isp->isp_dmatag, dmap, xs->data, xs->datalen, 516 NULL, (cansleep ? BUS_DMA_WAITOK : BUS_DMA_NOWAIT) | 517 BUS_DMA_STREAMING); 518 if (error != 0) { 519 XS_SETERR(xs, HBA_BOTCH); 520 if (error == EAGAIN || error == ENOMEM) 521 return (CMD_EAGAIN); 522 else 523 return (CMD_COMPLETE); 524 } 525 526 bus_dmamap_sync(isp->isp_dmatag, dmap, 0, xs->datalen, 527 in? BUS_DMASYNC_PREREAD : BUS_DMASYNC_PREWRITE); 528 529 if (in) { 530 rq->req_flags |= REQFLAG_DATA_IN; 531 } else { 532 rq->req_flags |= REQFLAG_DATA_OUT; 533 } 534 535 if (XS_CDBLEN(xs) > 12) { 536 uint32_t onxti; 537 ispcontreq_t local, *crq = &local, *cqe; 538 539 onxti = *nxtip; 540 cqe = (ispcontreq_t *) ISP_QUEUE_ENTRY(isp->isp_rquest, onxti); 541 *nxtip = ISP_NXT_QENTRY(onxti, RQUEST_QUEUE_LEN(isp)); 542 if (*nxtip == optr) { 543 isp_prt(isp, ISP_LOGDEBUG0, "Request Queue Overflow++"); 544 bus_dmamap_unload(isp->isp_dmatag, dmap); 545 XS_SETERR(xs, HBA_BOTCH); 546 return (CMD_EAGAIN); 547 } 548 rq->req_seg_count = 2; 549 MEMZERO((void *)crq, sizeof (*crq)); 550 crq->req_header.rqs_entry_count = 1; 551 crq->req_header.rqs_entry_type = RQSTYPE_DATASEG; 552 crq->req_dataseg[0].ds_count = xs->datalen; 553 crq->req_dataseg[0].ds_base = dmap->dm_segs[0].ds_addr; 554 isp_put_cont_req(isp, crq, cqe); 555 MEMORYBARRIER(isp, SYNC_REQUEST, onxti, QENTRY_LEN); 556 } else { 557 rq->req_seg_count = 1; 558 rq->req_dataseg[0].ds_count = xs->datalen; 559 rq->req_dataseg[0].ds_base = dmap->dm_segs[0].ds_addr; 560 } 561 562 mbxsync: 563 if (XS_CDBLEN(xs) > 12) { 564 isp_put_extended_request(isp, 565 (ispextreq_t *)rq, (ispextreq_t *) qep); 566 } else { 567 isp_put_request(isp, rq, qep); 568 } 569 return (CMD_QUEUED); 570 } 571 572 static void 573 isp_sbus_dmateardown(struct ispsoftc *isp, XS_T *xs, uint32_t handle) 574 { 575 struct isp_sbussoftc *sbc = (struct isp_sbussoftc *) isp; 576 bus_dmamap_t dmap; 577 578 dmap = sbc->sbus_dmamap[isp_handle_index(handle)]; 579 580 if (dmap->dm_nsegs == 0) { 581 panic("%s: DMA map not already allocated", isp->isp_name); 582 /* NOTREACHED */ 583 } 584 bus_dmamap_sync(isp->isp_dmatag, dmap, 0, 585 xs->datalen, (xs->xs_control & XS_CTL_DATA_IN)? 586 BUS_DMASYNC_POSTREAD : BUS_DMASYNC_POSTWRITE); 587 bus_dmamap_unload(isp->isp_dmatag, dmap); 588 } 589