1 /* $NetBSD: ciss.c,v 1.41 2019/11/10 21:16:35 chs Exp $ */ 2 /* $OpenBSD: ciss.c,v 1.68 2013/05/30 16:15:02 deraadt Exp $ */ 3 4 /* 5 * Copyright (c) 2005,2006 Michael Shalayeff 6 * All rights reserved. 7 * 8 * Permission to use, copy, modify, and distribute this software for any 9 * purpose with or without fee is hereby granted, provided that the above 10 * copyright notice and this permission notice appear in all copies. 11 * 12 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES 13 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF 14 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR 15 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES 16 * WHATSOEVER RESULTING FROM LOSS OF MIND, USE, DATA OR PROFITS, WHETHER IN 17 * AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT 18 * OF OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. 19 */ 20 21 #include <sys/cdefs.h> 22 __KERNEL_RCSID(0, "$NetBSD: ciss.c,v 1.41 2019/11/10 21:16:35 chs Exp $"); 23 24 #include "bio.h" 25 26 /* #define CISS_DEBUG */ 27 28 #include <sys/param.h> 29 #include <sys/systm.h> 30 #include <sys/buf.h> 31 #include <sys/ioctl.h> 32 #include <sys/device.h> 33 #include <sys/kernel.h> 34 #include <sys/malloc.h> 35 #include <sys/proc.h> 36 37 #include <sys/bus.h> 38 39 #include <dev/scsipi/scsi_all.h> 40 #include <dev/scsipi/scsi_disk.h> 41 #include <dev/scsipi/scsiconf.h> 42 #include <dev/scsipi/scsipi_all.h> 43 44 #include <dev/ic/cissreg.h> 45 #include <dev/ic/cissvar.h> 46 47 #if NBIO > 0 48 #include <dev/biovar.h> 49 #endif /* NBIO > 0 */ 50 51 #ifdef CISS_DEBUG 52 #define CISS_DPRINTF(m,a) if (ciss_debug & (m)) printf a 53 #define CISS_D_CMD 0x0001 54 #define CISS_D_INTR 0x0002 55 #define CISS_D_MISC 0x0004 56 #define CISS_D_DMA 0x0008 57 #define CISS_D_IOCTL 0x0010 58 #define CISS_D_ERR 0x0020 59 int ciss_debug = 0 60 | CISS_D_CMD 61 | CISS_D_INTR 62 | CISS_D_MISC 63 | CISS_D_DMA 64 | CISS_D_IOCTL 65 | CISS_D_ERR 66 ; 67 #else 68 #define CISS_DPRINTF(m,a) /* m, a */ 69 #endif 70 71 static void ciss_scsi_cmd(struct scsipi_channel *chan, 72 scsipi_adapter_req_t req, void *arg); 73 static int ciss_scsi_ioctl(struct scsipi_channel *chan, u_long cmd, 74 void *addr, int flag, struct proc *p); 75 static void cissminphys(struct buf *bp); 76 77 #if 0 78 static void ciss_scsi_raw_cmd(struct scsipi_channel *chan, 79 scsipi_adapter_req_t req, void *arg); 80 #endif 81 82 static int ciss_sync(struct ciss_softc *sc); 83 static void ciss_heartbeat(void *v); 84 static void ciss_shutdown(void *v); 85 86 static struct ciss_ccb *ciss_get_ccb(struct ciss_softc *sc); 87 static void ciss_put_ccb(struct ciss_ccb *ccb); 88 static int ciss_cmd(struct ciss_ccb *ccb, int flags, int wait); 89 static int ciss_done(struct ciss_ccb *ccb); 90 static int ciss_error(struct ciss_ccb *ccb); 91 struct ciss_ld *ciss_pdscan(struct ciss_softc *sc, int ld); 92 static int ciss_inq(struct ciss_softc *sc, struct ciss_inquiry *inq); 93 int ciss_ldid(struct ciss_softc *, int, struct ciss_ldid *); 94 int ciss_ldstat(struct ciss_softc *, int, struct ciss_ldstat *); 95 static int ciss_ldmap(struct ciss_softc *sc); 96 int ciss_pdid(struct ciss_softc *, u_int8_t, struct ciss_pdid *, int); 97 98 #if NBIO > 0 99 int ciss_ioctl(device_t, u_long, void *); 100 int ciss_ioctl_vol(struct ciss_softc *, struct bioc_vol *); 101 int ciss_blink(struct ciss_softc *, int, int, int, struct ciss_blink *); 102 int ciss_create_sensors(struct ciss_softc *); 103 void ciss_sensor_refresh(struct sysmon_envsys *, envsys_data_t *); 104 #endif /* NBIO > 0 */ 105 106 static struct ciss_ccb * 107 ciss_get_ccb(struct ciss_softc *sc) 108 { 109 struct ciss_ccb *ccb; 110 111 mutex_enter(&sc->sc_mutex); 112 if ((ccb = TAILQ_LAST(&sc->sc_free_ccb, ciss_queue_head))) { 113 TAILQ_REMOVE(&sc->sc_free_ccb, ccb, ccb_link); 114 ccb->ccb_state = CISS_CCB_READY; 115 } 116 mutex_exit(&sc->sc_mutex); 117 return ccb; 118 } 119 120 static void 121 ciss_put_ccb(struct ciss_ccb *ccb) 122 { 123 struct ciss_softc *sc = ccb->ccb_sc; 124 125 ccb->ccb_state = CISS_CCB_FREE; 126 mutex_enter(&sc->sc_mutex); 127 TAILQ_INSERT_TAIL(&sc->sc_free_ccb, ccb, ccb_link); 128 mutex_exit(&sc->sc_mutex); 129 } 130 131 int 132 ciss_attach(struct ciss_softc *sc) 133 { 134 struct ciss_ccb *ccb; 135 struct ciss_cmd *cmd; 136 struct ciss_inquiry *inq; 137 bus_dma_segment_t seg[1]; 138 int error, i, total, rseg, maxfer; 139 paddr_t pa; 140 141 bus_space_read_region_4(sc->sc_iot, sc->cfg_ioh, sc->cfgoff, 142 (u_int32_t *)&sc->cfg, sizeof(sc->cfg) / 4); 143 144 if (sc->cfg.signature != CISS_SIGNATURE) { 145 aprint_error(": bad sign 0x%08x\n", sc->cfg.signature); 146 return -1; 147 } 148 149 if (!(sc->cfg.methods & CISS_METH_SIMPL)) { 150 aprint_error(": not simple 0x%08x\n", sc->cfg.methods); 151 return -1; 152 } 153 154 sc->cfg.rmethod = CISS_METH_SIMPL; 155 sc->cfg.paddr_lim = 0; /* 32bit addrs */ 156 sc->cfg.int_delay = 0; /* disable coalescing */ 157 sc->cfg.int_count = 0; 158 strlcpy(sc->cfg.hostname, "HUMPPA", sizeof(sc->cfg.hostname)); 159 sc->cfg.driverf |= CISS_DRV_PRF; /* enable prefetch */ 160 if (!sc->cfg.maxsg) 161 sc->cfg.maxsg = MAXPHYS / PAGE_SIZE + 1; 162 163 bus_space_write_region_4(sc->sc_iot, sc->cfg_ioh, sc->cfgoff, 164 (u_int32_t *)&sc->cfg, sizeof(sc->cfg) / 4); 165 bus_space_barrier(sc->sc_iot, sc->cfg_ioh, sc->cfgoff, sizeof(sc->cfg), 166 BUS_SPACE_BARRIER_READ|BUS_SPACE_BARRIER_WRITE); 167 168 bus_space_write_4(sc->sc_iot, sc->sc_ioh, CISS_IDB, CISS_IDB_CFG); 169 bus_space_barrier(sc->sc_iot, sc->sc_ioh, CISS_IDB, 4, 170 BUS_SPACE_BARRIER_WRITE); 171 for (i = 1000; i--; DELAY(1000)) { 172 /* XXX maybe IDB is really 64bit? - hp dl380 needs this */ 173 (void)bus_space_read_4(sc->sc_iot, sc->sc_ioh, CISS_IDB + 4); 174 if (!(bus_space_read_4(sc->sc_iot, sc->sc_ioh, CISS_IDB) & CISS_IDB_CFG)) 175 break; 176 bus_space_barrier(sc->sc_iot, sc->sc_ioh, CISS_IDB, 4, 177 BUS_SPACE_BARRIER_READ); 178 } 179 180 if (bus_space_read_4(sc->sc_iot, sc->sc_ioh, CISS_IDB) & CISS_IDB_CFG) { 181 printf(": cannot set config\n"); 182 return -1; 183 } 184 185 bus_space_read_region_4(sc->sc_iot, sc->cfg_ioh, sc->cfgoff, 186 (u_int32_t *)&sc->cfg, sizeof(sc->cfg) / 4); 187 188 if (!(sc->cfg.amethod & CISS_METH_SIMPL)) { 189 printf(": cannot simplify 0x%08x\n", sc->cfg.amethod); 190 return -1; 191 } 192 193 /* i'm ready for you and i hope you're ready for me */ 194 for (i = 30000; i--; DELAY(1000)) { 195 if (bus_space_read_4(sc->sc_iot, sc->cfg_ioh, sc->cfgoff + 196 offsetof(struct ciss_config, amethod)) & CISS_METH_READY) 197 break; 198 bus_space_barrier(sc->sc_iot, sc->cfg_ioh, sc->cfgoff + 199 offsetof(struct ciss_config, amethod), 4, 200 BUS_SPACE_BARRIER_READ); 201 } 202 203 if (!(bus_space_read_4(sc->sc_iot, sc->cfg_ioh, sc->cfgoff + 204 offsetof(struct ciss_config, amethod)) & CISS_METH_READY)) { 205 aprint_error(": she never came ready for me 0x%08x\n", 206 sc->cfg.amethod); 207 return -1; 208 } 209 210 mutex_init(&sc->sc_mutex, MUTEX_DEFAULT, IPL_VM); 211 mutex_init(&sc->sc_mutex_scratch, MUTEX_DEFAULT, IPL_VM); 212 cv_init(&sc->sc_condvar, "ciss_cmd"); 213 sc->maxcmd = sc->cfg.maxcmd; 214 sc->maxsg = sc->cfg.maxsg; 215 if (sc->maxsg > MAXPHYS / PAGE_SIZE + 1) 216 sc->maxsg = MAXPHYS / PAGE_SIZE + 1; 217 i = sizeof(struct ciss_ccb) + 218 sizeof(ccb->ccb_cmd.sgl[0]) * (sc->maxsg - 1); 219 for (sc->ccblen = 0x10; sc->ccblen < i; sc->ccblen <<= 1); 220 221 total = sc->ccblen * sc->maxcmd; 222 if ((error = bus_dmamem_alloc(sc->sc_dmat, total, PAGE_SIZE, 0, 223 sc->cmdseg, 1, &rseg, BUS_DMA_NOWAIT))) { 224 aprint_error(": cannot allocate CCBs (%d)\n", error); 225 return -1; 226 } 227 228 if ((error = bus_dmamem_map(sc->sc_dmat, sc->cmdseg, rseg, total, 229 (void **)&sc->ccbs, BUS_DMA_NOWAIT))) { 230 aprint_error(": cannot map CCBs (%d)\n", error); 231 return -1; 232 } 233 memset(sc->ccbs, 0, total); 234 235 if ((error = bus_dmamap_create(sc->sc_dmat, total, 1, 236 total, 0, BUS_DMA_NOWAIT | BUS_DMA_ALLOCNOW, &sc->cmdmap))) { 237 aprint_error(": cannot create CCBs dmamap (%d)\n", error); 238 bus_dmamem_free(sc->sc_dmat, sc->cmdseg, 1); 239 return -1; 240 } 241 242 if ((error = bus_dmamap_load(sc->sc_dmat, sc->cmdmap, sc->ccbs, total, 243 NULL, BUS_DMA_NOWAIT))) { 244 aprint_error(": cannot load CCBs dmamap (%d)\n", error); 245 bus_dmamem_free(sc->sc_dmat, sc->cmdseg, 1); 246 bus_dmamap_destroy(sc->sc_dmat, sc->cmdmap); 247 return -1; 248 } 249 250 TAILQ_INIT(&sc->sc_ccbq); 251 TAILQ_INIT(&sc->sc_ccbdone); 252 TAILQ_INIT(&sc->sc_free_ccb); 253 254 maxfer = sc->maxsg * PAGE_SIZE; 255 for (i = 0; total > 0 && i < sc->maxcmd; i++, total -= sc->ccblen) { 256 ccb = (struct ciss_ccb *) ((char *)sc->ccbs + i * sc->ccblen); 257 cmd = &ccb->ccb_cmd; 258 pa = sc->cmdseg[0].ds_addr + i * sc->ccblen; 259 260 ccb->ccb_sc = sc; 261 ccb->ccb_cmdpa = pa + offsetof(struct ciss_ccb, ccb_cmd); 262 ccb->ccb_state = CISS_CCB_FREE; 263 264 cmd->id = htole32(i << 2); 265 cmd->id_hi = htole32(0); 266 cmd->sgin = sc->maxsg; 267 cmd->sglen = htole16((u_int16_t)cmd->sgin); 268 cmd->err_len = htole32(sizeof(ccb->ccb_err)); 269 pa += offsetof(struct ciss_ccb, ccb_err); 270 cmd->err_pa = htole64((u_int64_t)pa); 271 272 if ((error = bus_dmamap_create(sc->sc_dmat, maxfer, sc->maxsg, 273 maxfer, 0, BUS_DMA_NOWAIT | BUS_DMA_ALLOCNOW, 274 &ccb->ccb_dmamap))) 275 break; 276 277 TAILQ_INSERT_TAIL(&sc->sc_free_ccb, ccb, ccb_link); 278 } 279 280 if (i < sc->maxcmd) { 281 aprint_error(": cannot create ccb#%d dmamap (%d)\n", i, error); 282 if (i == 0) { 283 /* TODO leaking cmd's dmamaps and shitz */ 284 bus_dmamem_free(sc->sc_dmat, sc->cmdseg, 1); 285 bus_dmamap_destroy(sc->sc_dmat, sc->cmdmap); 286 return -1; 287 } 288 } 289 290 if ((error = bus_dmamem_alloc(sc->sc_dmat, PAGE_SIZE, PAGE_SIZE, 0, 291 seg, 1, &rseg, BUS_DMA_NOWAIT))) { 292 aprint_error(": cannot allocate scratch buffer (%d)\n", error); 293 return -1; 294 } 295 296 if ((error = bus_dmamem_map(sc->sc_dmat, seg, rseg, PAGE_SIZE, 297 (void **)&sc->scratch, BUS_DMA_NOWAIT))) { 298 aprint_error(": cannot map scratch buffer (%d)\n", error); 299 return -1; 300 } 301 memset(sc->scratch, 0, PAGE_SIZE); 302 sc->sc_waitflag = XS_CTL_NOSLEEP; /* can't sleep yet */ 303 304 mutex_enter(&sc->sc_mutex_scratch); /* is this really needed? */ 305 inq = sc->scratch; 306 if (ciss_inq(sc, inq)) { 307 aprint_error(": adapter inquiry failed\n"); 308 mutex_exit(&sc->sc_mutex_scratch); 309 bus_dmamem_free(sc->sc_dmat, sc->cmdseg, 1); 310 bus_dmamap_destroy(sc->sc_dmat, sc->cmdmap); 311 return -1; 312 } 313 314 if (!(inq->flags & CISS_INQ_BIGMAP)) { 315 aprint_error(": big map is not supported, flags=0x%x\n", 316 inq->flags); 317 mutex_exit(&sc->sc_mutex_scratch); 318 bus_dmamem_free(sc->sc_dmat, sc->cmdseg, 1); 319 bus_dmamap_destroy(sc->sc_dmat, sc->cmdmap); 320 return -1; 321 } 322 323 sc->maxunits = inq->numld; 324 sc->nbus = inq->nscsi_bus; 325 sc->ndrives = inq->buswidth ? inq->buswidth : 256; 326 aprint_normal(": %d LD%s, HW rev %d, FW %4.4s/%4.4s", 327 inq->numld, inq->numld == 1? "" : "s", 328 inq->hw_rev, inq->fw_running, inq->fw_stored); 329 330 if (sc->cfg.methods & CISS_METH_FIFO64) 331 aprint_normal(", 64bit fifo"); 332 else if (sc->cfg.methods & CISS_METH_FIFO64_RRO) 333 aprint_normal(", 64bit fifo rro"); 334 aprint_normal("\n"); 335 336 mutex_exit(&sc->sc_mutex_scratch); 337 338 callout_init(&sc->sc_hb, 0); 339 callout_setfunc(&sc->sc_hb, ciss_heartbeat, sc); 340 callout_schedule(&sc->sc_hb, hz * 3); 341 342 /* map LDs */ 343 if (ciss_ldmap(sc)) { 344 aprint_error_dev(sc->sc_dev, "adapter LD map failed\n"); 345 bus_dmamem_free(sc->sc_dmat, sc->cmdseg, 1); 346 bus_dmamap_destroy(sc->sc_dmat, sc->cmdmap); 347 return -1; 348 } 349 350 sc->sc_lds = malloc(sc->maxunits * sizeof(*sc->sc_lds), 351 M_DEVBUF, M_WAITOK | M_ZERO); 352 353 sc->sc_flush = CISS_FLUSH_ENABLE; 354 if (!(sc->sc_sh = shutdownhook_establish(ciss_shutdown, sc))) { 355 aprint_error_dev(sc->sc_dev, 356 "unable to establish shutdown hook\n"); 357 bus_dmamem_free(sc->sc_dmat, sc->cmdseg, 1); 358 bus_dmamap_destroy(sc->sc_dmat, sc->cmdmap); 359 return -1; 360 } 361 362 sc->sc_channel.chan_adapter = &sc->sc_adapter; 363 sc->sc_channel.chan_bustype = &scsi_bustype; 364 sc->sc_channel.chan_channel = 0; 365 sc->sc_channel.chan_ntargets = sc->maxunits; 366 sc->sc_channel.chan_nluns = 1; /* ciss doesn't really have SCSI luns */ 367 sc->sc_channel.chan_openings = sc->maxcmd; 368 #if NBIO > 0 369 /* XXX Reserve some ccb's for sensor and bioctl. */ 370 if (sc->sc_channel.chan_openings > 2) 371 sc->sc_channel.chan_openings -= 2; 372 #endif 373 sc->sc_channel.chan_flags = 0; 374 sc->sc_channel.chan_id = sc->maxunits; 375 376 sc->sc_adapter.adapt_dev = sc->sc_dev; 377 sc->sc_adapter.adapt_openings = sc->sc_channel.chan_openings; 378 sc->sc_adapter.adapt_max_periph = uimin(sc->sc_adapter.adapt_openings, 256); 379 sc->sc_adapter.adapt_request = ciss_scsi_cmd; 380 sc->sc_adapter.adapt_minphys = cissminphys; 381 sc->sc_adapter.adapt_ioctl = ciss_scsi_ioctl; 382 sc->sc_adapter.adapt_nchannels = 1; 383 config_found(sc->sc_dev, &sc->sc_channel, scsiprint); 384 385 #if 0 386 sc->sc_link_raw.adapter_softc = sc; 387 sc->sc_link.openings = sc->sc_channel.chan_openings; 388 sc->sc_link_raw.adapter = &ciss_raw_switch; 389 sc->sc_link_raw.adapter_target = sc->ndrives; 390 sc->sc_link_raw.adapter_buswidth = sc->ndrives; 391 config_found(sc->sc_dev, &sc->sc_channel, scsiprint); 392 #endif 393 394 #if NBIO > 0 395 /* now map all the physdevs into their lds */ 396 /* XXX currently we assign all of them into ld0 */ 397 for (i = 0; i < sc->maxunits && i < 1; i++) 398 if (!(sc->sc_lds[i] = ciss_pdscan(sc, i))) { 399 sc->sc_waitflag = 0; /* we can sleep now */ 400 return 0; 401 } 402 403 if (bio_register(sc->sc_dev, ciss_ioctl) != 0) 404 aprint_error_dev(sc->sc_dev, "controller registration failed"); 405 else 406 sc->sc_ioctl = ciss_ioctl; 407 if (ciss_create_sensors(sc) != 0) 408 aprint_error_dev(sc->sc_dev, "unable to create sensors"); 409 #endif 410 sc->sc_waitflag = 0; /* we can sleep now */ 411 412 return 0; 413 } 414 415 static void 416 ciss_shutdown(void *v) 417 { 418 struct ciss_softc *sc = v; 419 420 sc->sc_flush = CISS_FLUSH_DISABLE; 421 /* timeout_del(&sc->sc_hb); */ 422 ciss_sync(sc); 423 } 424 425 static void 426 cissminphys(struct buf *bp) 427 { 428 #if 0 /* TODO */ 429 #define CISS_MAXFER (PAGE_SIZE * (sc->maxsg + 1)) 430 if (bp->b_bcount > CISS_MAXFER) 431 bp->b_bcount = CISS_MAXFER; 432 #endif 433 minphys(bp); 434 } 435 436 static struct ciss_ccb * 437 ciss_poll1(struct ciss_softc *sc) 438 { 439 struct ciss_ccb *ccb; 440 uint32_t id; 441 442 if (!(bus_space_read_4(sc->sc_iot, sc->sc_ioh, CISS_ISR) & sc->iem)) { 443 CISS_DPRINTF(CISS_D_CMD, ("N")); 444 return NULL; 445 } 446 447 if (sc->cfg.methods & CISS_METH_FIFO64) { 448 if (bus_space_read_4(sc->sc_iot, sc->sc_ioh, CISS_OUTQ64_HI) == 449 0xffffffff) { 450 CISS_DPRINTF(CISS_D_CMD, ("Q")); 451 return NULL; 452 } 453 id = bus_space_read_4(sc->sc_iot, sc->sc_ioh, CISS_OUTQ64_LO); 454 } else if (sc->cfg.methods & CISS_METH_FIFO64_RRO) { 455 id = bus_space_read_4(sc->sc_iot, sc->sc_ioh, CISS_OUTQ64_LO); 456 if (id == 0xffffffff) { 457 CISS_DPRINTF(CISS_D_CMD, ("Q")); 458 return NULL; 459 } 460 (void)bus_space_read_4(sc->sc_iot, sc->sc_ioh, CISS_OUTQ64_HI); 461 } else { 462 id = bus_space_read_4(sc->sc_iot, sc->sc_ioh, CISS_OUTQ); 463 if (id == 0xffffffff) { 464 CISS_DPRINTF(CISS_D_CMD, ("Q")); 465 return NULL; 466 } 467 } 468 469 CISS_DPRINTF(CISS_D_CMD, ("got=0x%x ", id)); 470 ccb = (struct ciss_ccb *) ((char *)sc->ccbs + (id >> 2) * sc->ccblen); 471 ccb->ccb_cmd.id = htole32(id); 472 ccb->ccb_cmd.id_hi = htole32(0); 473 return ccb; 474 } 475 476 static int 477 ciss_poll(struct ciss_softc *sc, struct ciss_ccb *ccb, int ms) 478 { 479 struct ciss_ccb *ccb1; 480 481 ms /= 10; 482 483 while (ms-- > 0) { 484 DELAY(10); 485 ccb1 = ciss_poll1(sc); 486 if (ccb1 == NULL) 487 continue; 488 ciss_done(ccb1); 489 if (ccb1 == ccb) 490 return 0; 491 } 492 493 return ETIMEDOUT; 494 } 495 496 static int 497 ciss_wait(struct ciss_softc *sc, struct ciss_ccb *ccb, int ms) 498 { 499 int tohz, etick; 500 501 tohz = mstohz(ms); 502 if (tohz == 0) 503 tohz = 1; 504 etick = hardclock_ticks + tohz; 505 506 for (;;) { 507 ccb->ccb_state = CISS_CCB_POLL; 508 CISS_DPRINTF(CISS_D_CMD, ("cv_timedwait(%d) ", tohz)); 509 mutex_enter(&sc->sc_mutex); 510 if (cv_timedwait(&sc->sc_condvar, &sc->sc_mutex, tohz) 511 == EWOULDBLOCK) { 512 mutex_exit(&sc->sc_mutex); 513 return EWOULDBLOCK; 514 } 515 mutex_exit(&sc->sc_mutex); 516 if (ccb->ccb_state == CISS_CCB_ONQ) { 517 ciss_done(ccb); 518 return 0; 519 } 520 tohz = etick - hardclock_ticks; 521 if (tohz <= 0) 522 return EWOULDBLOCK; 523 CISS_DPRINTF(CISS_D_CMD, ("T")); 524 } 525 } 526 527 /* 528 * submit a command and optionally wait for completition. 529 * wait arg abuses XS_CTL_POLL|XS_CTL_NOSLEEP flags to request 530 * to wait (XS_CTL_POLL) and to allow tsleep() (!XS_CTL_NOSLEEP) 531 * instead of busy loop waiting 532 */ 533 static int 534 ciss_cmd(struct ciss_ccb *ccb, int flags, int wait) 535 { 536 struct ciss_softc *sc = ccb->ccb_sc; 537 struct ciss_cmd *cmd = &ccb->ccb_cmd; 538 bus_dmamap_t dmap = ccb->ccb_dmamap; 539 u_int64_t addr; 540 int i, error = 0; 541 542 if (ccb->ccb_state != CISS_CCB_READY) { 543 printf("%s: ccb %d not ready state=0x%x\n", device_xname(sc->sc_dev), 544 cmd->id, ccb->ccb_state); 545 return (EINVAL); 546 } 547 548 if (ccb->ccb_data) { 549 bus_dma_segment_t *sgd; 550 551 if ((error = bus_dmamap_load(sc->sc_dmat, dmap, ccb->ccb_data, 552 ccb->ccb_len, NULL, flags))) { 553 if (error == EFBIG) 554 printf("more than %d dma segs\n", sc->maxsg); 555 else 556 printf("error %d loading dma map\n", error); 557 ciss_put_ccb(ccb); 558 return (error); 559 } 560 cmd->sgin = dmap->dm_nsegs; 561 562 sgd = dmap->dm_segs; 563 CISS_DPRINTF(CISS_D_DMA, ("data=%p/%zu<%#" PRIxPADDR "/%zu", 564 ccb->ccb_data, ccb->ccb_len, sgd->ds_addr, sgd->ds_len)); 565 566 for (i = 0; i < dmap->dm_nsegs; sgd++, i++) { 567 cmd->sgl[i].addr_lo = htole32(sgd->ds_addr); 568 cmd->sgl[i].addr_hi = 569 htole32((u_int64_t)sgd->ds_addr >> 32); 570 cmd->sgl[i].len = htole32(sgd->ds_len); 571 cmd->sgl[i].flags = htole32(0); 572 if (i) { 573 CISS_DPRINTF(CISS_D_DMA, 574 (",%#" PRIxPADDR "/%zu", sgd->ds_addr, 575 sgd->ds_len)); 576 } 577 } 578 579 CISS_DPRINTF(CISS_D_DMA, ("> ")); 580 581 bus_dmamap_sync(sc->sc_dmat, dmap, 0, dmap->dm_mapsize, 582 BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE); 583 } else 584 cmd->sgin = 0; 585 cmd->sglen = htole16((u_int16_t)cmd->sgin); 586 memset(&ccb->ccb_err, 0, sizeof(ccb->ccb_err)); 587 588 bus_dmamap_sync(sc->sc_dmat, sc->cmdmap, 0, sc->cmdmap->dm_mapsize, 589 BUS_DMASYNC_PREWRITE); 590 591 #ifndef CISS_NO_INTERRUPT_HACK 592 if ((wait & (XS_CTL_POLL|XS_CTL_NOSLEEP)) == (XS_CTL_POLL|XS_CTL_NOSLEEP)) 593 bus_space_write_4(sc->sc_iot, sc->sc_ioh, CISS_IMR, 594 bus_space_read_4(sc->sc_iot, sc->sc_ioh, CISS_IMR) | sc->iem); 595 #endif 596 597 mutex_enter(&sc->sc_mutex); 598 TAILQ_INSERT_TAIL(&sc->sc_ccbq, ccb, ccb_link); 599 mutex_exit(&sc->sc_mutex); 600 ccb->ccb_state = CISS_CCB_ONQ; 601 CISS_DPRINTF(CISS_D_CMD, ("submit=0x%x ", cmd->id)); 602 if (sc->cfg.methods & (CISS_METH_FIFO64|CISS_METH_FIFO64_RRO)) { 603 /* 604 * Write the upper 32bits immediately before the lower 605 * 32bits and set bit 63 to indicate 64bit FIFO mode. 606 */ 607 addr = (u_int64_t)ccb->ccb_cmdpa; 608 bus_space_write_4(sc->sc_iot, sc->sc_ioh, CISS_INQ64_HI, 609 (addr >> 32) | 0x80000000); 610 bus_space_write_4(sc->sc_iot, sc->sc_ioh, CISS_INQ64_LO, 611 addr & 0x00000000ffffffffULL); 612 } else 613 bus_space_write_4(sc->sc_iot, sc->sc_ioh, CISS_INQ, 614 ccb->ccb_cmdpa); 615 616 if (wait & XS_CTL_POLL) { 617 int ms; 618 CISS_DPRINTF(CISS_D_CMD, ("waiting ")); 619 620 ms = ccb->ccb_xs ? ccb->ccb_xs->timeout : 60000; 621 if (wait & XS_CTL_NOSLEEP) 622 error = ciss_poll(sc, ccb, ms); 623 else 624 error = ciss_wait(sc, ccb, ms); 625 626 /* if never got a chance to be done above... */ 627 if (ccb->ccb_state != CISS_CCB_FREE) { 628 KASSERT(error); 629 ccb->ccb_err.cmd_stat = CISS_ERR_TMO; 630 error = ciss_done(ccb); 631 } 632 633 CISS_DPRINTF(CISS_D_CMD, ("done %d:%d", 634 ccb->ccb_err.cmd_stat, ccb->ccb_err.scsi_stat)); 635 } 636 637 #ifndef CISS_NO_INTERRUPT_HACK 638 if ((wait & (XS_CTL_POLL|XS_CTL_NOSLEEP)) == (XS_CTL_POLL|XS_CTL_NOSLEEP)) 639 bus_space_write_4(sc->sc_iot, sc->sc_ioh, CISS_IMR, 640 bus_space_read_4(sc->sc_iot, sc->sc_ioh, CISS_IMR) & ~sc->iem); 641 #endif 642 643 return (error); 644 } 645 646 static int 647 ciss_done(struct ciss_ccb *ccb) 648 { 649 struct ciss_softc *sc = ccb->ccb_sc; 650 struct scsipi_xfer *xs = ccb->ccb_xs; 651 struct ciss_cmd *cmd; 652 int error = 0; 653 654 CISS_DPRINTF(CISS_D_CMD, ("ciss_done(%p) ", ccb)); 655 656 if (ccb->ccb_state != CISS_CCB_ONQ) { 657 printf("%s: unqueued ccb %p ready, state=0x%x\n", 658 device_xname(sc->sc_dev), ccb, ccb->ccb_state); 659 return 1; 660 } 661 662 ccb->ccb_state = CISS_CCB_READY; 663 mutex_enter(&sc->sc_mutex); 664 TAILQ_REMOVE(&sc->sc_ccbq, ccb, ccb_link); 665 mutex_exit(&sc->sc_mutex); 666 667 if (ccb->ccb_cmd.id & CISS_CMD_ERR) 668 error = ciss_error(ccb); 669 670 cmd = &ccb->ccb_cmd; 671 if (ccb->ccb_data) { 672 bus_dmamap_sync(sc->sc_dmat, ccb->ccb_dmamap, 0, 673 ccb->ccb_dmamap->dm_mapsize, (cmd->flags & CISS_CDB_IN) ? 674 BUS_DMASYNC_POSTREAD : BUS_DMASYNC_POSTWRITE); 675 bus_dmamap_unload(sc->sc_dmat, ccb->ccb_dmamap); 676 ccb->ccb_xs = NULL; 677 ccb->ccb_data = NULL; 678 } 679 680 ciss_put_ccb(ccb); 681 682 if (xs) { 683 xs->resid = 0; 684 CISS_DPRINTF(CISS_D_CMD, ("scsipi_done(%p) ", xs)); 685 if (xs->cmd->opcode == INQUIRY) { 686 struct scsipi_inquiry_data *inq; 687 inq = (struct scsipi_inquiry_data *)xs->data; 688 if ((inq->version & SID_ANSII) == 0 && 689 (inq->flags3 & SID_CmdQue) != 0) { 690 inq->version |= 2; 691 } 692 } 693 scsipi_done(xs); 694 } 695 696 return error; 697 } 698 699 static int 700 ciss_error(struct ciss_ccb *ccb) 701 { 702 struct ciss_softc *sc = ccb->ccb_sc; 703 struct ciss_error *err = &ccb->ccb_err; 704 struct scsipi_xfer *xs = ccb->ccb_xs; 705 int rv; 706 707 switch ((rv = le16toh(err->cmd_stat))) { 708 case CISS_ERR_OK: 709 rv = 0; 710 break; 711 712 case CISS_ERR_INVCMD: 713 if (xs == NULL || 714 xs->cmd->opcode != SCSI_SYNCHRONIZE_CACHE_10) 715 printf("%s: invalid cmd 0x%x: 0x%x is not valid @ 0x%x[%d]\n", 716 device_xname(sc->sc_dev), ccb->ccb_cmd.id, 717 err->err_info, err->err_type[3], err->err_type[2]); 718 if (xs) { 719 memset(&xs->sense, 0, sizeof(xs->sense)); 720 xs->sense.scsi_sense.response_code = 721 SSD_RCODE_CURRENT | SSD_RCODE_VALID; 722 xs->sense.scsi_sense.flags = SKEY_ILLEGAL_REQUEST; 723 xs->sense.scsi_sense.asc = 0x24; /* ill field */ 724 xs->sense.scsi_sense.ascq = 0x0; 725 xs->error = XS_SENSE; 726 } 727 rv = EIO; 728 break; 729 730 case CISS_ERR_TMO: 731 xs->error = XS_TIMEOUT; 732 rv = ETIMEDOUT; 733 break; 734 735 case CISS_ERR_UNRUN: 736 /* Underrun */ 737 xs->resid = le32toh(err->resid); 738 CISS_DPRINTF(CISS_D_CMD, (" underrun resid=0x%x ", 739 xs->resid)); 740 rv = EIO; 741 break; 742 default: 743 if (xs) { 744 CISS_DPRINTF(CISS_D_CMD, ("scsi_stat=%x ", err->scsi_stat)); 745 switch (err->scsi_stat) { 746 case SCSI_CHECK: 747 xs->error = XS_SENSE; 748 memcpy(&xs->sense, &err->sense[0], 749 sizeof(xs->sense)); 750 CISS_DPRINTF(CISS_D_CMD, (" sense=%02x %02x %02x %02x ", 751 err->sense[0], err->sense[1], err->sense[2], err->sense[3])); 752 rv = EIO; 753 break; 754 755 case XS_BUSY: 756 xs->error = XS_BUSY; 757 rv = EBUSY; 758 break; 759 760 default: 761 CISS_DPRINTF(CISS_D_ERR, ("%s: " 762 "cmd_stat=%x scsi_stat=0x%x resid=0x%x\n", 763 device_xname(sc->sc_dev), rv, err->scsi_stat, 764 le32toh(err->resid))); 765 printf("ciss driver stuffup in %s:%d: %s()\n", 766 __FILE__, __LINE__, __func__); 767 xs->error = XS_DRIVER_STUFFUP; 768 rv = EIO; 769 break; 770 } 771 xs->resid = le32toh(err->resid); 772 } else 773 rv = EIO; 774 } 775 ccb->ccb_cmd.id &= htole32(~3); 776 777 return rv; 778 } 779 780 static int 781 ciss_inq(struct ciss_softc *sc, struct ciss_inquiry *inq) 782 { 783 struct ciss_ccb *ccb; 784 struct ciss_cmd *cmd; 785 786 ccb = ciss_get_ccb(sc); 787 ccb->ccb_len = sizeof(*inq); 788 ccb->ccb_data = inq; 789 ccb->ccb_xs = NULL; 790 cmd = &ccb->ccb_cmd; 791 cmd->tgt = htole32(CISS_CMD_MODE_PERIPH); 792 cmd->tgt2 = 0; 793 cmd->cdblen = 10; 794 cmd->flags = CISS_CDB_CMD | CISS_CDB_SIMPL | CISS_CDB_IN; 795 cmd->tmo = htole16(0); 796 memset(&cmd->cdb[0], 0, sizeof(cmd->cdb)); 797 cmd->cdb[0] = CISS_CMD_CTRL_GET; 798 cmd->cdb[6] = CISS_CMS_CTRL_CTRL; 799 cmd->cdb[7] = sizeof(*inq) >> 8; /* biiiig endian */ 800 cmd->cdb[8] = sizeof(*inq) & 0xff; 801 802 return ciss_cmd(ccb, BUS_DMA_NOWAIT, XS_CTL_POLL|XS_CTL_NOSLEEP); 803 } 804 805 static int 806 ciss_ldmap(struct ciss_softc *sc) 807 { 808 struct ciss_ccb *ccb; 809 struct ciss_cmd *cmd; 810 struct ciss_ldmap *lmap; 811 int total, rv; 812 813 mutex_enter(&sc->sc_mutex_scratch); 814 lmap = sc->scratch; 815 lmap->size = htobe32(sc->maxunits * sizeof(lmap->map)); 816 total = sizeof(*lmap) + (sc->maxunits - 1) * sizeof(lmap->map); 817 818 ccb = ciss_get_ccb(sc); 819 ccb->ccb_len = total; 820 ccb->ccb_data = lmap; 821 ccb->ccb_xs = NULL; 822 cmd = &ccb->ccb_cmd; 823 cmd->tgt = CISS_CMD_MODE_PERIPH; 824 cmd->tgt2 = 0; 825 cmd->cdblen = 12; 826 cmd->flags = CISS_CDB_CMD | CISS_CDB_SIMPL | CISS_CDB_IN; 827 cmd->tmo = htole16(30); 828 memset(&cmd->cdb[0], 0, sizeof(cmd->cdb)); 829 cmd->cdb[0] = CISS_CMD_LDMAP; 830 cmd->cdb[8] = total >> 8; /* biiiig endian */ 831 cmd->cdb[9] = total & 0xff; 832 833 rv = ciss_cmd(ccb, BUS_DMA_NOWAIT, XS_CTL_POLL|XS_CTL_NOSLEEP); 834 835 if (rv) { 836 mutex_exit(&sc->sc_mutex_scratch); 837 return rv; 838 } 839 840 CISS_DPRINTF(CISS_D_MISC, ("lmap %x:%x\n", 841 lmap->map[0].tgt, lmap->map[0].tgt2)); 842 843 mutex_exit(&sc->sc_mutex_scratch); 844 return 0; 845 } 846 847 static int 848 ciss_sync(struct ciss_softc *sc) 849 { 850 struct ciss_ccb *ccb; 851 struct ciss_cmd *cmd; 852 struct ciss_flush *flush; 853 int rv; 854 855 mutex_enter(&sc->sc_mutex_scratch); 856 flush = sc->scratch; 857 memset(flush, 0, sizeof(*flush)); 858 flush->flush = sc->sc_flush; 859 860 ccb = ciss_get_ccb(sc); 861 ccb->ccb_len = sizeof(*flush); 862 ccb->ccb_data = flush; 863 ccb->ccb_xs = NULL; 864 cmd = &ccb->ccb_cmd; 865 cmd->tgt = CISS_CMD_MODE_PERIPH; 866 cmd->tgt2 = 0; 867 cmd->cdblen = 10; 868 cmd->flags = CISS_CDB_CMD | CISS_CDB_SIMPL | CISS_CDB_OUT; 869 cmd->tmo = 0; 870 memset(&cmd->cdb[0], 0, sizeof(cmd->cdb)); 871 cmd->cdb[0] = CISS_CMD_CTRL_SET; 872 cmd->cdb[6] = CISS_CMS_CTRL_FLUSH; 873 cmd->cdb[7] = sizeof(*flush) >> 8; /* biiiig endian */ 874 cmd->cdb[8] = sizeof(*flush) & 0xff; 875 876 rv = ciss_cmd(ccb, BUS_DMA_NOWAIT, XS_CTL_POLL|XS_CTL_NOSLEEP); 877 mutex_exit(&sc->sc_mutex_scratch); 878 879 return rv; 880 } 881 882 int 883 ciss_ldid(struct ciss_softc *sc, int target, struct ciss_ldid *id) 884 { 885 struct ciss_ccb *ccb; 886 struct ciss_cmd *cmd; 887 888 ccb = ciss_get_ccb(sc); 889 if (ccb == NULL) 890 return ENOMEM; 891 ccb->ccb_len = sizeof(*id); 892 ccb->ccb_data = id; 893 ccb->ccb_xs = NULL; 894 cmd = &ccb->ccb_cmd; 895 cmd->tgt = htole32(CISS_CMD_MODE_PERIPH); 896 cmd->tgt2 = 0; 897 cmd->cdblen = 10; 898 cmd->flags = CISS_CDB_CMD | CISS_CDB_SIMPL | CISS_CDB_IN; 899 cmd->tmo = htole16(0); 900 memset(&cmd->cdb[0], 0, sizeof(cmd->cdb)); 901 cmd->cdb[0] = CISS_CMD_CTRL_GET; 902 cmd->cdb[1] = target; 903 cmd->cdb[6] = CISS_CMS_CTRL_LDIDEXT; 904 cmd->cdb[7] = sizeof(*id) >> 8; /* biiiig endian */ 905 cmd->cdb[8] = sizeof(*id) & 0xff; 906 907 return ciss_cmd(ccb, BUS_DMA_NOWAIT, XS_CTL_POLL | sc->sc_waitflag); 908 } 909 910 int 911 ciss_ldstat(struct ciss_softc *sc, int target, struct ciss_ldstat *stat) 912 { 913 struct ciss_ccb *ccb; 914 struct ciss_cmd *cmd; 915 916 ccb = ciss_get_ccb(sc); 917 if (ccb == NULL) 918 return ENOMEM; 919 ccb->ccb_len = sizeof(*stat); 920 ccb->ccb_data = stat; 921 ccb->ccb_xs = NULL; 922 cmd = &ccb->ccb_cmd; 923 cmd->tgt = htole32(CISS_CMD_MODE_PERIPH); 924 cmd->tgt2 = 0; 925 cmd->cdblen = 10; 926 cmd->flags = CISS_CDB_CMD | CISS_CDB_SIMPL | CISS_CDB_IN; 927 cmd->tmo = htole16(0); 928 memset(&cmd->cdb[0], 0, sizeof(cmd->cdb)); 929 cmd->cdb[0] = CISS_CMD_CTRL_GET; 930 cmd->cdb[1] = target; 931 cmd->cdb[6] = CISS_CMS_CTRL_LDSTAT; 932 cmd->cdb[7] = sizeof(*stat) >> 8; /* biiiig endian */ 933 cmd->cdb[8] = sizeof(*stat) & 0xff; 934 935 return ciss_cmd(ccb, BUS_DMA_NOWAIT, XS_CTL_POLL | sc->sc_waitflag); 936 } 937 938 int 939 ciss_pdid(struct ciss_softc *sc, u_int8_t drv, struct ciss_pdid *id, int wait) 940 { 941 struct ciss_ccb *ccb; 942 struct ciss_cmd *cmd; 943 944 ccb = ciss_get_ccb(sc); 945 if (ccb == NULL) 946 return ENOMEM; 947 ccb->ccb_len = sizeof(*id); 948 ccb->ccb_data = id; 949 ccb->ccb_xs = NULL; 950 cmd = &ccb->ccb_cmd; 951 cmd->tgt = htole32(CISS_CMD_MODE_PERIPH); 952 cmd->tgt2 = 0; 953 cmd->cdblen = 10; 954 cmd->flags = CISS_CDB_CMD | CISS_CDB_SIMPL | CISS_CDB_IN; 955 cmd->tmo = htole16(0); 956 memset(&cmd->cdb[0], 0, sizeof(cmd->cdb)); 957 cmd->cdb[0] = CISS_CMD_CTRL_GET; 958 cmd->cdb[2] = drv; 959 cmd->cdb[6] = CISS_CMS_CTRL_PDID; 960 cmd->cdb[7] = sizeof(*id) >> 8; /* biiiig endian */ 961 cmd->cdb[8] = sizeof(*id) & 0xff; 962 963 return ciss_cmd(ccb, BUS_DMA_NOWAIT, wait); 964 } 965 966 967 struct ciss_ld * 968 ciss_pdscan(struct ciss_softc *sc, int ld) 969 { 970 struct ciss_pdid *pdid; 971 struct ciss_ld *ldp; 972 u_int8_t drv, buf[128]; 973 int i, j, k = 0; 974 975 mutex_enter(&sc->sc_mutex_scratch); 976 pdid = sc->scratch; 977 if (sc->ndrives == 256) { 978 for (i = 0; i < CISS_BIGBIT; i++) 979 if (!ciss_pdid(sc, i, pdid, 980 XS_CTL_POLL|XS_CTL_NOSLEEP) && 981 (pdid->present & CISS_PD_PRESENT)) 982 buf[k++] = i; 983 } else 984 for (i = 0; i < sc->nbus; i++) 985 for (j = 0; j < sc->ndrives; j++) { 986 drv = CISS_BIGBIT + i * sc->ndrives + j; 987 if (!ciss_pdid(sc, drv, pdid, 988 XS_CTL_POLL|XS_CTL_NOSLEEP)) 989 buf[k++] = drv; 990 } 991 mutex_exit(&sc->sc_mutex_scratch); 992 993 if (!k) 994 return NULL; 995 996 ldp = malloc(sizeof(*ldp) + (k-1), M_DEVBUF, M_WAITOK); 997 memset(&ldp->bling, 0, sizeof(ldp->bling)); 998 ldp->ndrives = k; 999 ldp->xname[0] = 0; 1000 memcpy(ldp->tgts, buf, k); 1001 return ldp; 1002 } 1003 1004 #if 0 1005 static void 1006 ciss_scsi_raw_cmd(struct scsipi_channel *chan, scsipi_adapter_req_t req, 1007 void *arg) /* TODO */ 1008 { 1009 struct scsipi_xfer *xs = (struct scsipi_xfer *) arg; 1010 struct ciss_rawsoftc *rsc = device_private( 1011 chan->chan_adapter->adapt_dev); 1012 struct ciss_softc *sc = rsc->sc_softc; 1013 struct ciss_ccb *ccb; 1014 struct ciss_cmd *cmd; 1015 int error; 1016 1017 CISS_DPRINTF(CISS_D_CMD, ("ciss_scsi_raw_cmd ")); 1018 1019 switch (req) 1020 { 1021 case ADAPTER_REQ_RUN_XFER: 1022 if (xs->cmdlen > CISS_MAX_CDB) { 1023 CISS_DPRINTF(CISS_D_CMD, ("CDB too big %p ", xs)); 1024 memset(&xs->sense, 0, sizeof(xs->sense)); 1025 printf("ciss driver stuffup in %s:%d: %s()\n", 1026 __FILE__, __LINE__, __func__); 1027 xs->error = XS_DRIVER_STUFFUP; 1028 scsipi_done(xs); 1029 break; 1030 } 1031 1032 error = 0; 1033 xs->error = XS_NOERROR; 1034 1035 /* TODO check this target has not yet employed w/ any volume */ 1036 1037 ccb = ciss_get_ccb(sc); 1038 cmd = &ccb->ccb_cmd; 1039 ccb->ccb_len = xs->datalen; 1040 ccb->ccb_data = xs->data; 1041 ccb->ccb_xs = xs; 1042 1043 cmd->cdblen = xs->cmdlen; 1044 cmd->flags = CISS_CDB_CMD | CISS_CDB_SIMPL; 1045 if (xs->xs_control & XS_CTL_DATA_IN) 1046 cmd->flags |= CISS_CDB_IN; 1047 else if (xs->xs_control & XS_CTL_DATA_OUT) 1048 cmd->flags |= CISS_CDB_OUT; 1049 cmd->tmo = xs->timeout < 1000? 1 : xs->timeout / 1000; 1050 memset(&cmd->cdb[0], 0, sizeof(cmd->cdb)); 1051 memcpy(&cmd->cdb[0], xs->cmd, CISS_MAX_CDB); 1052 1053 if (ciss_cmd(ccb, BUS_DMA_WAITOK, 1054 xs->xs_control & (XS_CTL_POLL|XS_CTL_NOSLEEP))) { 1055 printf("ciss driver stuffup in %s:%d: %s()\n", 1056 __FILE__, __LINE__, __func__); 1057 xs->error = XS_DRIVER_STUFFUP; 1058 scsipi_done(xs); 1059 break; 1060 } 1061 1062 break; 1063 1064 case ADAPTER_REQ_GROW_RESOURCES: 1065 /* 1066 * Not supported. 1067 */ 1068 break; 1069 1070 case ADAPTER_REQ_SET_XFER_MODE: 1071 /* 1072 * We can't change the transfer mode, but at least let 1073 * scsipi know what the adapter has negociated. 1074 */ 1075 /* Get xfer mode and return it */ 1076 break; 1077 } 1078 } 1079 #endif 1080 1081 static void 1082 ciss_scsi_cmd(struct scsipi_channel *chan, scsipi_adapter_req_t req, 1083 void *arg) 1084 { 1085 struct scsipi_xfer *xs; 1086 struct scsipi_xfer_mode *xm; 1087 struct ciss_softc *sc = device_private(chan->chan_adapter->adapt_dev); 1088 u_int8_t target; 1089 struct ciss_ccb *ccb; 1090 struct ciss_cmd *cmd; 1091 1092 CISS_DPRINTF(CISS_D_CMD, ("ciss_scsi_cmd ")); 1093 1094 switch (req) 1095 { 1096 case ADAPTER_REQ_RUN_XFER: 1097 xs = (struct scsipi_xfer *) arg; 1098 target = xs->xs_periph->periph_target; 1099 CISS_DPRINTF(CISS_D_CMD, ("targ=%d ", target)); 1100 if (xs->cmdlen > CISS_MAX_CDB) { 1101 CISS_DPRINTF(CISS_D_CMD, ("CDB too big %p ", xs)); 1102 memset(&xs->sense, 0, sizeof(xs->sense)); 1103 xs->error = XS_SENSE; 1104 printf("ciss driver stuffup in %s:%d: %s()\n", 1105 __FILE__, __LINE__, __func__); 1106 scsipi_done(xs); 1107 break; 1108 } 1109 1110 xs->error = XS_NOERROR; 1111 1112 /* XXX emulate SYNCHRONIZE_CACHE ??? */ 1113 1114 ccb = ciss_get_ccb(sc); 1115 cmd = &ccb->ccb_cmd; 1116 ccb->ccb_len = xs->datalen; 1117 ccb->ccb_data = xs->data; 1118 ccb->ccb_xs = xs; 1119 cmd->tgt = CISS_CMD_MODE_LD | target; 1120 cmd->tgt2 = 0; 1121 cmd->cdblen = xs->cmdlen; 1122 cmd->flags = CISS_CDB_CMD | CISS_CDB_SIMPL; 1123 if (xs->xs_control & XS_CTL_DATA_IN) 1124 cmd->flags |= CISS_CDB_IN; 1125 else if (xs->xs_control & XS_CTL_DATA_OUT) 1126 cmd->flags |= CISS_CDB_OUT; 1127 cmd->tmo = htole16(xs->timeout < 1000? 1 : xs->timeout / 1000); 1128 memset(&cmd->cdb[0], 0, sizeof(cmd->cdb)); 1129 memcpy(&cmd->cdb[0], xs->cmd, CISS_MAX_CDB); 1130 CISS_DPRINTF(CISS_D_CMD, ("cmd=%02x %02x %02x %02x %02x %02x ", 1131 cmd->cdb[0], cmd->cdb[1], cmd->cdb[2], 1132 cmd->cdb[3], cmd->cdb[4], cmd->cdb[5])); 1133 1134 if (ciss_cmd(ccb, BUS_DMA_WAITOK, 1135 xs->xs_control & (XS_CTL_POLL|XS_CTL_NOSLEEP))) { 1136 printf("ciss driver stuffup in %s:%d: %s()\n", 1137 __FILE__, __LINE__, __func__); 1138 xs->error = XS_DRIVER_STUFFUP; 1139 scsipi_done(xs); 1140 return; 1141 } 1142 1143 break; 1144 case ADAPTER_REQ_GROW_RESOURCES: 1145 /* 1146 * Not supported. 1147 */ 1148 break; 1149 case ADAPTER_REQ_SET_XFER_MODE: 1150 /* 1151 * We can't change the transfer mode, but at least let 1152 * scsipi know what the adapter has negociated. 1153 */ 1154 xm = (struct scsipi_xfer_mode *)arg; 1155 xm->xm_mode |= PERIPH_CAP_TQING; 1156 scsipi_async_event(chan, ASYNC_EVENT_XFER_MODE, xm); 1157 break; 1158 default: 1159 printf("%s: %d %d unsupported\n", __func__, __LINE__, req); 1160 } 1161 } 1162 1163 int 1164 ciss_intr(void *v) 1165 { 1166 struct ciss_softc *sc = v; 1167 struct ciss_ccb *ccb; 1168 u_int32_t id; 1169 bus_size_t reg; 1170 int hit = 0; 1171 1172 CISS_DPRINTF(CISS_D_INTR, ("intr ")); 1173 1174 if (!(bus_space_read_4(sc->sc_iot, sc->sc_ioh, CISS_ISR) & sc->iem)) 1175 return 0; 1176 1177 if (sc->cfg.methods & CISS_METH_FIFO64) 1178 reg = CISS_OUTQ64_HI; 1179 else if (sc->cfg.methods & CISS_METH_FIFO64_RRO) 1180 reg = CISS_OUTQ64_LO; 1181 else 1182 reg = CISS_OUTQ; 1183 while ((id = bus_space_read_4(sc->sc_iot, sc->sc_ioh, reg)) != 1184 0xffffffff) { 1185 if (reg == CISS_OUTQ64_HI) 1186 id = bus_space_read_4(sc->sc_iot, sc->sc_ioh, 1187 CISS_OUTQ64_LO); 1188 else if (reg == CISS_OUTQ64_LO) 1189 (void)bus_space_read_4(sc->sc_iot, sc->sc_ioh, 1190 CISS_OUTQ64_HI); 1191 1192 ccb = (struct ciss_ccb *) ((char *)sc->ccbs + (id >> 2) * sc->ccblen); 1193 ccb->ccb_cmd.id = htole32(id); 1194 ccb->ccb_cmd.id_hi = htole32(0); /* ignore the upper 32bits */ 1195 if (ccb->ccb_state == CISS_CCB_POLL) { 1196 ccb->ccb_state = CISS_CCB_ONQ; 1197 mutex_enter(&sc->sc_mutex); 1198 cv_broadcast(&sc->sc_condvar); 1199 mutex_exit(&sc->sc_mutex); 1200 } else 1201 ciss_done(ccb); 1202 1203 hit = 1; 1204 } 1205 1206 CISS_DPRINTF(CISS_D_INTR, ("exit\n")); 1207 return hit; 1208 } 1209 1210 static void 1211 ciss_heartbeat(void *v) 1212 { 1213 struct ciss_softc *sc = v; 1214 u_int32_t hb; 1215 1216 hb = bus_space_read_4(sc->sc_iot, sc->cfg_ioh, 1217 sc->cfgoff + offsetof(struct ciss_config, heartbeat)); 1218 if (hb == sc->heartbeat) { 1219 sc->fibrillation++; 1220 CISS_DPRINTF(CISS_D_ERR, ("%s: fibrillation #%d (value=%d)\n", 1221 device_xname(sc->sc_dev), sc->fibrillation, hb)); 1222 if (sc->fibrillation >= 11) { 1223 /* No heartbeat for 33 seconds */ 1224 panic("%s: dead", device_xname(sc->sc_dev)); /* XXX reset! */ 1225 } 1226 } else { 1227 sc->heartbeat = hb; 1228 if (sc->fibrillation) { 1229 CISS_DPRINTF(CISS_D_ERR, ("%s: " 1230 "fibrillation ended (value=%d)\n", 1231 device_xname(sc->sc_dev), hb)); 1232 } 1233 sc->fibrillation = 0; 1234 } 1235 1236 callout_schedule(&sc->sc_hb, hz * 3); 1237 } 1238 1239 static int 1240 ciss_scsi_ioctl(struct scsipi_channel *chan, u_long cmd, 1241 void *addr, int flag, struct proc *p) 1242 { 1243 #if NBIO > 0 1244 return ciss_ioctl(chan->chan_adapter->adapt_dev, cmd, addr); 1245 #else 1246 return ENOTTY; 1247 #endif 1248 } 1249 1250 #if NBIO > 0 1251 const int ciss_level[] = { 0, 4, 1, 5, 51, 7 }; 1252 const int ciss_stat[] = { BIOC_SVONLINE, BIOC_SVOFFLINE, BIOC_SVOFFLINE, 1253 BIOC_SVDEGRADED, BIOC_SVREBUILD, BIOC_SVREBUILD, BIOC_SVDEGRADED, 1254 BIOC_SVDEGRADED, BIOC_SVINVALID, BIOC_SVINVALID, BIOC_SVBUILDING, 1255 BIOC_SVOFFLINE, BIOC_SVBUILDING }; 1256 1257 int 1258 ciss_ioctl(device_t dev, u_long cmd, void *addr) 1259 { 1260 struct ciss_softc *sc = device_private(dev); 1261 struct bioc_inq *bi; 1262 struct bioc_disk *bd; 1263 struct bioc_blink *bb; 1264 struct ciss_ldstat *ldstat; 1265 struct ciss_pdid *pdid; 1266 struct ciss_blink *blink; 1267 struct ciss_ld *ldp; 1268 u_int8_t drv; 1269 int ld, pd, error = 0; 1270 1271 switch (cmd) { 1272 case BIOCINQ: 1273 bi = (struct bioc_inq *)addr; 1274 strlcpy(bi->bi_dev, device_xname(sc->sc_dev), sizeof(bi->bi_dev)); 1275 bi->bi_novol = sc->maxunits; 1276 bi->bi_nodisk = sc->sc_lds[0]->ndrives; 1277 break; 1278 1279 case BIOCVOL: 1280 error = ciss_ioctl_vol(sc, (struct bioc_vol *)addr); 1281 break; 1282 1283 case BIOCDISK_NOVOL: 1284 /* 1285 * XXX since we don't know how to associate physical drives with logical drives 1286 * yet, BIOCDISK_NOVOL is equivalent to BIOCDISK to the volume that we've 1287 * associated all physical drives to. 1288 * Maybe assoicate all physical drives to all logical volumes, but only return 1289 * physical drives on one logical volume. Which one? Either 1st volume that 1290 * is degraded, rebuilding, or failed? 1291 */ 1292 bd = (struct bioc_disk *)addr; 1293 bd->bd_volid = 0; 1294 bd->bd_disknovol = true; 1295 /* FALLTHROUGH */ 1296 case BIOCDISK: 1297 bd = (struct bioc_disk *)addr; 1298 if (bd->bd_volid < 0 || bd->bd_volid > sc->maxunits) { 1299 error = EINVAL; 1300 break; 1301 } 1302 ldp = sc->sc_lds[0]; 1303 if (!ldp || (pd = bd->bd_diskid) < 0 || pd > ldp->ndrives) { 1304 error = EINVAL; 1305 break; 1306 } 1307 ldstat = sc->scratch; 1308 if ((error = ciss_ldstat(sc, bd->bd_volid, ldstat))) { 1309 break; 1310 } 1311 bd->bd_status = -1; 1312 if (ldstat->stat == CISS_LD_REBLD && 1313 ldstat->bigrebuild == ldp->tgts[pd]) 1314 bd->bd_status = BIOC_SDREBUILD; 1315 if (ciss_bitset(ldp->tgts[pd] & (~CISS_BIGBIT), 1316 ldstat->bigfailed)) { 1317 bd->bd_status = BIOC_SDFAILED; 1318 bd->bd_size = 0; 1319 bd->bd_channel = (ldp->tgts[pd] & (~CISS_BIGBIT)) / 1320 sc->ndrives; 1321 bd->bd_target = ldp->tgts[pd] % sc->ndrives; 1322 bd->bd_lun = 0; 1323 bd->bd_vendor[0] = '\0'; 1324 bd->bd_serial[0] = '\0'; 1325 bd->bd_procdev[0] = '\0'; 1326 } else { 1327 pdid = sc->scratch; 1328 if ((error = ciss_pdid(sc, ldp->tgts[pd], pdid, 1329 XS_CTL_POLL))) { 1330 bd->bd_status = BIOC_SDFAILED; 1331 bd->bd_size = 0; 1332 bd->bd_channel = (ldp->tgts[pd] & (~CISS_BIGBIT)) / 1333 sc->ndrives; 1334 bd->bd_target = ldp->tgts[pd] % sc->ndrives; 1335 bd->bd_lun = 0; 1336 bd->bd_vendor[0] = '\0'; 1337 bd->bd_serial[0] = '\0'; 1338 bd->bd_procdev[0] = '\0'; 1339 error = 0; 1340 break; 1341 } 1342 if (bd->bd_status < 0) { 1343 if (pdid->config & CISS_PD_SPARE) 1344 bd->bd_status = BIOC_SDHOTSPARE; 1345 else if (pdid->present & CISS_PD_PRESENT) 1346 bd->bd_status = BIOC_SDONLINE; 1347 else 1348 bd->bd_status = BIOC_SDINVALID; 1349 } 1350 bd->bd_size = (u_int64_t)le32toh(pdid->nblocks) * 1351 le16toh(pdid->blksz); 1352 bd->bd_channel = pdid->bus; 1353 bd->bd_target = pdid->target; 1354 bd->bd_lun = 0; 1355 strlcpy(bd->bd_vendor, pdid->model, 1356 sizeof(bd->bd_vendor)); 1357 strlcpy(bd->bd_serial, pdid->serial, 1358 sizeof(bd->bd_serial)); 1359 bd->bd_procdev[0] = '\0'; 1360 } 1361 break; 1362 1363 case BIOCBLINK: 1364 bb = (struct bioc_blink *)addr; 1365 blink = sc->scratch; 1366 error = EINVAL; 1367 /* XXX workaround completely dumb scsi addressing */ 1368 for (ld = 0; ld < sc->maxunits; ld++) { 1369 ldp = sc->sc_lds[ld]; 1370 if (!ldp) 1371 continue; 1372 if (sc->ndrives == 256) 1373 drv = bb->bb_target; 1374 else 1375 drv = CISS_BIGBIT + 1376 bb->bb_channel * sc->ndrives + 1377 bb->bb_target; 1378 for (pd = 0; pd < ldp->ndrives; pd++) 1379 if (ldp->tgts[pd] == drv) 1380 error = ciss_blink(sc, ld, pd, 1381 bb->bb_status, blink); 1382 } 1383 break; 1384 1385 default: 1386 error = EINVAL; 1387 } 1388 1389 return (error); 1390 } 1391 1392 int 1393 ciss_ioctl_vol(struct ciss_softc *sc, struct bioc_vol *bv) 1394 { 1395 struct ciss_ldid *ldid; 1396 struct ciss_ld *ldp; 1397 struct ciss_ldstat *ldstat; 1398 struct ciss_pdid *pdid; 1399 int error = 0; 1400 u_int blks; 1401 1402 if (bv->bv_volid < 0 || bv->bv_volid > sc->maxunits) { 1403 return EINVAL; 1404 } 1405 ldp = sc->sc_lds[bv->bv_volid]; 1406 ldid = sc->scratch; 1407 if ((error = ciss_ldid(sc, bv->bv_volid, ldid))) { 1408 return error; 1409 } 1410 bv->bv_status = BIOC_SVINVALID; 1411 blks = (u_int)le16toh(ldid->nblocks[1]) << 16 | 1412 le16toh(ldid->nblocks[0]); 1413 bv->bv_size = blks * (u_quad_t)le16toh(ldid->blksize); 1414 bv->bv_level = ciss_level[ldid->type]; 1415 /* 1416 * XXX Should only return bv_nodisk for logigal volume that we've associated 1417 * the physical drives to: either the 1st degraded, rebuilding, or failed 1418 * volume else volume 0? 1419 */ 1420 if (ldp) { 1421 bv->bv_nodisk = ldp->ndrives; 1422 strlcpy(bv->bv_dev, ldp->xname, sizeof(bv->bv_dev)); 1423 } 1424 strlcpy(bv->bv_vendor, "CISS", sizeof(bv->bv_vendor)); 1425 ldstat = sc->scratch; 1426 memset(ldstat, 0, sizeof(*ldstat)); 1427 if ((error = ciss_ldstat(sc, bv->bv_volid, ldstat))) { 1428 return error; 1429 } 1430 bv->bv_percent = -1; 1431 bv->bv_seconds = 0; 1432 if (ldstat->stat < sizeof(ciss_stat)/sizeof(ciss_stat[0])) 1433 bv->bv_status = ciss_stat[ldstat->stat]; 1434 if (bv->bv_status == BIOC_SVREBUILD || 1435 bv->bv_status == BIOC_SVBUILDING) { 1436 u_int64_t prog; 1437 1438 ldp = sc->sc_lds[0]; 1439 if (ldp) { 1440 bv->bv_nodisk = ldp->ndrives; 1441 strlcpy(bv->bv_dev, ldp->xname, sizeof(bv->bv_dev)); 1442 } 1443 /* 1444 * XXX ldstat->prog is blocks remaining on physical drive being rebuilt 1445 * blks is only correct for a RAID1 set; RAID5 needs to determine the 1446 * size of the physical device - which we don't yet know. 1447 * ldstat->bigrebuild has physical device target, so could be used with 1448 * pdid to get size. Another way is to save pd information in sc so it's 1449 * easy to reference. 1450 */ 1451 prog = (u_int64_t)((ldstat->prog[3] << 24) | 1452 (ldstat->prog[2] << 16) | (ldstat->prog[1] << 8) | 1453 ldstat->prog[0]); 1454 pdid = sc->scratch; 1455 if (!ciss_pdid(sc, ldstat->bigrebuild, pdid, XS_CTL_POLL)) { 1456 blks = le32toh(pdid->nblocks); 1457 bv->bv_percent = (blks - prog) * 1000ULL / blks; 1458 } 1459 } 1460 return 0; 1461 } 1462 1463 int 1464 ciss_blink(struct ciss_softc *sc, int ld, int pd, int stat, 1465 struct ciss_blink *blink) 1466 { 1467 struct ciss_ccb *ccb; 1468 struct ciss_cmd *cmd; 1469 struct ciss_ld *ldp; 1470 1471 if (ld > sc->maxunits) 1472 return EINVAL; 1473 1474 ldp = sc->sc_lds[ld]; 1475 if (!ldp || pd > ldp->ndrives) 1476 return EINVAL; 1477 1478 ldp->bling.pdtab[ldp->tgts[pd]] = stat == BIOC_SBUNBLINK? 0 : 1479 CISS_BLINK_ALL; 1480 memcpy(blink, &ldp->bling, sizeof(*blink)); 1481 1482 ccb = ciss_get_ccb(sc); 1483 if (ccb == NULL) 1484 return ENOMEM; 1485 ccb->ccb_len = sizeof(*blink); 1486 ccb->ccb_data = blink; 1487 ccb->ccb_xs = NULL; 1488 cmd = &ccb->ccb_cmd; 1489 cmd->tgt = htole32(CISS_CMD_MODE_PERIPH); 1490 cmd->tgt2 = 0; 1491 cmd->cdblen = 10; 1492 cmd->flags = CISS_CDB_CMD | CISS_CDB_SIMPL | CISS_CDB_OUT; 1493 cmd->tmo = htole16(0); 1494 memset(&cmd->cdb[0], 0, sizeof(cmd->cdb)); 1495 cmd->cdb[0] = CISS_CMD_CTRL_SET; 1496 cmd->cdb[6] = CISS_CMS_CTRL_PDBLINK; 1497 cmd->cdb[7] = sizeof(*blink) >> 8; /* biiiig endian */ 1498 cmd->cdb[8] = sizeof(*blink) & 0xff; 1499 1500 return ciss_cmd(ccb, BUS_DMA_NOWAIT, XS_CTL_POLL); 1501 } 1502 1503 int 1504 ciss_create_sensors(struct ciss_softc *sc) 1505 { 1506 int i; 1507 int nsensors = sc->maxunits; 1508 1509 if (nsensors == 0) { 1510 return 0; 1511 } 1512 1513 sc->sc_sme = sysmon_envsys_create(); 1514 sc->sc_sensor = malloc(sizeof(envsys_data_t) * nsensors, 1515 M_DEVBUF, M_WAITOK | M_ZERO); 1516 1517 for (i = 0; i < nsensors; i++) { 1518 sc->sc_sensor[i].units = ENVSYS_DRIVE; 1519 sc->sc_sensor[i].state = ENVSYS_SINVALID; 1520 sc->sc_sensor[i].value_cur = ENVSYS_DRIVE_EMPTY; 1521 /* Enable monitoring for drive state changes */ 1522 sc->sc_sensor[i].flags |= ENVSYS_FMONSTCHANGED; 1523 /* logical drives */ 1524 snprintf(sc->sc_sensor[i].desc, 1525 sizeof(sc->sc_sensor[i].desc), "%s:%d", 1526 device_xname(sc->sc_dev), i); 1527 if (sysmon_envsys_sensor_attach(sc->sc_sme, 1528 &sc->sc_sensor[i])) 1529 goto out; 1530 } 1531 1532 sc->sc_sme->sme_name = device_xname(sc->sc_dev); 1533 sc->sc_sme->sme_cookie = sc; 1534 sc->sc_sme->sme_refresh = ciss_sensor_refresh; 1535 if (sysmon_envsys_register(sc->sc_sme)) { 1536 printf("%s: unable to register with sysmon\n", 1537 device_xname(sc->sc_dev)); 1538 return(1); 1539 } 1540 return (0); 1541 1542 out: 1543 free(sc->sc_sensor, M_DEVBUF); 1544 sysmon_envsys_destroy(sc->sc_sme); 1545 return EINVAL; 1546 } 1547 1548 void 1549 ciss_sensor_refresh(struct sysmon_envsys *sme, envsys_data_t *edata) 1550 { 1551 struct ciss_softc *sc = sme->sme_cookie; 1552 struct bioc_vol bv; 1553 1554 if (edata->sensor >= sc->maxunits) 1555 return; 1556 1557 memset(&bv, 0, sizeof(bv)); 1558 bv.bv_volid = edata->sensor; 1559 if (ciss_ioctl_vol(sc, &bv)) 1560 bv.bv_status = BIOC_SVINVALID; 1561 1562 bio_vol_to_envsys(edata, &bv); 1563 } 1564 #endif /* NBIO > 0 */ 1565