1 /* $NetBSD: icp.c,v 1.31 2012/10/27 17:18:20 chs Exp $ */ 2 3 /*- 4 * Copyright (c) 2002, 2003 The NetBSD Foundation, Inc. 5 * All rights reserved. 6 * 7 * This code is derived from software contributed to The NetBSD Foundation 8 * by Andrew Doran, and by Jason R. Thorpe of Wasabi Systems, Inc. 9 * 10 * Redistribution and use in source and binary forms, with or without 11 * modification, are permitted provided that the following conditions 12 * are met: 13 * 1. Redistributions of source code must retain the above copyright 14 * notice, this list of conditions and the following disclaimer. 15 * 2. Redistributions in binary form must reproduce the above copyright 16 * notice, this list of conditions and the following disclaimer in the 17 * documentation and/or other materials provided with the distribution. 18 * 19 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS 20 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED 21 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 22 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS 23 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 24 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 25 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 26 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 27 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 28 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 29 * POSSIBILITY OF SUCH DAMAGE. 30 */ 31 32 /* 33 * Copyright (c) 1999, 2000 Niklas Hallqvist. All rights reserved. 34 * 35 * Redistribution and use in source and binary forms, with or without 36 * modification, are permitted provided that the following conditions 37 * are met: 38 * 1. Redistributions of source code must retain the above copyright 39 * notice, this list of conditions and the following disclaimer. 40 * 2. Redistributions in binary form must reproduce the above copyright 41 * notice, this list of conditions and the following disclaimer in the 42 * documentation and/or other materials provided with the distribution. 43 * 3. All advertising materials mentioning features or use of this software 44 * must display the following acknowledgement: 45 * This product includes software developed by Niklas Hallqvist. 46 * 4. The name of the author may not be used to endorse or promote products 47 * derived from this software without specific prior written permission. 48 * 49 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 50 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 51 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 52 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, 53 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT 54 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 55 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 56 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 57 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 58 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 59 * 60 * from OpenBSD: gdt_common.c,v 1.12 2001/07/04 06:43:18 niklas Exp 61 */ 62 63 /* 64 * This driver would not have written if it was not for the hardware donations 65 * from both ICP-Vortex and �ko.neT. I want to thank them for their support. 66 * 67 * Re-worked for NetBSD by Andrew Doran. Test hardware kindly supplied by 68 * Intel. 69 * 70 * Support for the ICP-Vortex management tools added by 71 * Jason R. Thorpe of Wasabi Systems, Inc., based on code 72 * provided by Achim Leubner <achim.leubner@intel.com>. 73 * 74 * Additional support for dynamic rescan of cacheservice drives by 75 * Jason R. Thorpe of Wasabi Systems, Inc. 76 */ 77 78 #include <sys/cdefs.h> 79 __KERNEL_RCSID(0, "$NetBSD: icp.c,v 1.31 2012/10/27 17:18:20 chs Exp $"); 80 81 #include <sys/param.h> 82 #include <sys/systm.h> 83 #include <sys/kernel.h> 84 #include <sys/device.h> 85 #include <sys/queue.h> 86 #include <sys/proc.h> 87 #include <sys/buf.h> 88 #include <sys/endian.h> 89 #include <sys/malloc.h> 90 #include <sys/disk.h> 91 92 #include <sys/bswap.h> 93 #include <sys/bus.h> 94 95 #include <dev/pci/pcireg.h> 96 #include <dev/pci/pcivar.h> 97 #include <dev/pci/pcidevs.h> 98 99 #include <dev/ic/icpreg.h> 100 #include <dev/ic/icpvar.h> 101 102 #include <dev/scsipi/scsipi_all.h> 103 #include <dev/scsipi/scsiconf.h> 104 105 #include "locators.h" 106 107 int icp_async_event(struct icp_softc *, int); 108 void icp_ccb_submit(struct icp_softc *icp, struct icp_ccb *ic); 109 void icp_chain(struct icp_softc *); 110 int icp_print(void *, const char *); 111 void icp_watchdog(void *); 112 void icp_ucmd_intr(struct icp_ccb *); 113 void icp_recompute_openings(struct icp_softc *); 114 115 int icp_count; /* total # of controllers, for ioctl interface */ 116 117 /* 118 * Statistics for the ioctl interface to query. 119 * 120 * XXX Global. They should probably be made per-controller 121 * XXX at some point. 122 */ 123 gdt_statist_t icp_stats; 124 125 int 126 icp_init(struct icp_softc *icp, const char *intrstr) 127 { 128 struct icp_attach_args icpa; 129 struct icp_binfo binfo; 130 struct icp_ccb *ic; 131 u_int16_t cdev_cnt; 132 int i, j, state, feat, nsegs, rv; 133 int locs[ICPCF_NLOCS]; 134 135 state = 0; 136 137 if (intrstr != NULL) 138 aprint_normal_dev(icp->icp_dv, "interrupting at %s\n", 139 intrstr); 140 141 SIMPLEQ_INIT(&icp->icp_ccb_queue); 142 SIMPLEQ_INIT(&icp->icp_ccb_freelist); 143 SIMPLEQ_INIT(&icp->icp_ucmd_queue); 144 callout_init(&icp->icp_wdog_callout, 0); 145 146 /* 147 * Allocate a scratch area. 148 */ 149 if (bus_dmamap_create(icp->icp_dmat, ICP_SCRATCH_SIZE, 1, 150 ICP_SCRATCH_SIZE, 0, BUS_DMA_NOWAIT | BUS_DMA_ALLOCNOW, 151 &icp->icp_scr_dmamap) != 0) { 152 aprint_error_dev(icp->icp_dv, "cannot create scratch dmamap\n"); 153 return (1); 154 } 155 state++; 156 157 if (bus_dmamem_alloc(icp->icp_dmat, ICP_SCRATCH_SIZE, PAGE_SIZE, 0, 158 icp->icp_scr_seg, 1, &nsegs, BUS_DMA_NOWAIT) != 0) { 159 aprint_error_dev(icp->icp_dv, "cannot alloc scratch dmamem\n"); 160 goto bail_out; 161 } 162 state++; 163 164 if (bus_dmamem_map(icp->icp_dmat, icp->icp_scr_seg, nsegs, 165 ICP_SCRATCH_SIZE, &icp->icp_scr, 0)) { 166 aprint_error_dev(icp->icp_dv, "cannot map scratch dmamem\n"); 167 goto bail_out; 168 } 169 state++; 170 171 if (bus_dmamap_load(icp->icp_dmat, icp->icp_scr_dmamap, icp->icp_scr, 172 ICP_SCRATCH_SIZE, NULL, BUS_DMA_NOWAIT)) { 173 aprint_error_dev(icp->icp_dv, "cannot load scratch dmamap\n"); 174 goto bail_out; 175 } 176 state++; 177 178 /* 179 * Allocate and initialize the command control blocks. 180 */ 181 ic = malloc(sizeof(*ic) * ICP_NCCBS, M_DEVBUF, M_NOWAIT | M_ZERO); 182 if ((icp->icp_ccbs = ic) == NULL) { 183 aprint_error_dev(icp->icp_dv, "malloc() failed\n"); 184 goto bail_out; 185 } 186 state++; 187 188 for (i = 0; i < ICP_NCCBS; i++, ic++) { 189 /* 190 * The first two command indexes have special meanings, so 191 * we can't use them. 192 */ 193 ic->ic_ident = i + 2; 194 rv = bus_dmamap_create(icp->icp_dmat, ICP_MAX_XFER, 195 ICP_MAXSG, ICP_MAX_XFER, 0, 196 BUS_DMA_NOWAIT | BUS_DMA_ALLOCNOW, 197 &ic->ic_xfer_map); 198 if (rv != 0) 199 break; 200 icp->icp_nccbs++; 201 icp_ccb_free(icp, ic); 202 } 203 #ifdef DIAGNOSTIC 204 if (icp->icp_nccbs != ICP_NCCBS) 205 aprint_error_dev(icp->icp_dv, "%d/%d CCBs usable\n", 206 icp->icp_nccbs, ICP_NCCBS); 207 #endif 208 209 /* 210 * Initalize the controller. 211 */ 212 if (!icp_cmd(icp, ICP_SCREENSERVICE, ICP_INIT, 0, 0, 0)) { 213 aprint_error_dev(icp->icp_dv, "screen service init error %d\n", 214 icp->icp_status); 215 goto bail_out; 216 } 217 218 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INIT, ICP_LINUX_OS, 0, 0)) { 219 aprint_error_dev(icp->icp_dv, "cache service init error %d\n", 220 icp->icp_status); 221 goto bail_out; 222 } 223 224 icp_cmd(icp, ICP_CACHESERVICE, ICP_UNFREEZE_IO, 0, 0, 0); 225 226 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_MOUNT, 0xffff, 1, 0)) { 227 aprint_error_dev(icp->icp_dv, "cache service mount error %d\n", 228 icp->icp_status); 229 goto bail_out; 230 } 231 232 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INIT, ICP_LINUX_OS, 0, 0)) { 233 aprint_error_dev(icp->icp_dv, "cache service post-mount init error %d\n", 234 icp->icp_status); 235 goto bail_out; 236 } 237 cdev_cnt = (u_int16_t)icp->icp_info; 238 icp->icp_fw_vers = icp->icp_service; 239 240 if (!icp_cmd(icp, ICP_SCSIRAWSERVICE, ICP_INIT, 0, 0, 0)) { 241 aprint_error_dev(icp->icp_dv, "raw service init error %d\n", 242 icp->icp_status); 243 goto bail_out; 244 } 245 246 /* 247 * Set/get raw service features (scatter/gather). 248 */ 249 feat = 0; 250 if (icp_cmd(icp, ICP_SCSIRAWSERVICE, ICP_SET_FEAT, ICP_SCATTER_GATHER, 251 0, 0)) 252 if (icp_cmd(icp, ICP_SCSIRAWSERVICE, ICP_GET_FEAT, 0, 0, 0)) 253 feat = icp->icp_info; 254 255 if ((feat & ICP_SCATTER_GATHER) == 0) { 256 #ifdef DIAGNOSTIC 257 aprint_normal_dev(icp->icp_dv, 258 "scatter/gather not supported (raw service)\n"); 259 #endif 260 } else 261 icp->icp_features |= ICP_FEAT_RAWSERVICE; 262 263 /* 264 * Set/get cache service features (scatter/gather). 265 */ 266 feat = 0; 267 if (icp_cmd(icp, ICP_CACHESERVICE, ICP_SET_FEAT, 0, 268 ICP_SCATTER_GATHER, 0)) 269 if (icp_cmd(icp, ICP_CACHESERVICE, ICP_GET_FEAT, 0, 0, 0)) 270 feat = icp->icp_info; 271 272 if ((feat & ICP_SCATTER_GATHER) == 0) { 273 #ifdef DIAGNOSTIC 274 aprint_normal_dev(icp->icp_dv, 275 "scatter/gather not supported (cache service)\n"); 276 #endif 277 } else 278 icp->icp_features |= ICP_FEAT_CACHESERVICE; 279 280 /* 281 * Pull some information from the board and dump. 282 */ 283 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_IOCTL, ICP_BOARD_INFO, 284 ICP_INVALID_CHANNEL, sizeof(struct icp_binfo))) { 285 aprint_error_dev(icp->icp_dv, "unable to retrive board info\n"); 286 goto bail_out; 287 } 288 memcpy(&binfo, icp->icp_scr, sizeof(binfo)); 289 290 aprint_normal_dev(icp->icp_dv, 291 "model <%s>, firmware <%s>, %d channel(s), %dMB memory\n", 292 binfo.bi_type_string, binfo.bi_raid_string, 293 binfo.bi_chan_count, le32toh(binfo.bi_memsize) >> 20); 294 295 /* 296 * Determine the number of devices, and number of openings per 297 * device. 298 */ 299 if (icp->icp_features & ICP_FEAT_CACHESERVICE) { 300 for (j = 0; j < cdev_cnt && j < ICP_MAX_HDRIVES; j++) { 301 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INFO, j, 0, 302 0)) 303 continue; 304 305 icp->icp_cdr[j].cd_size = icp->icp_info; 306 if (icp->icp_cdr[j].cd_size != 0) 307 icp->icp_ndevs++; 308 309 if (icp_cmd(icp, ICP_CACHESERVICE, ICP_DEVTYPE, j, 0, 310 0)) 311 icp->icp_cdr[j].cd_type = icp->icp_info; 312 } 313 } 314 315 if (icp->icp_features & ICP_FEAT_RAWSERVICE) { 316 icp->icp_nchan = binfo.bi_chan_count; 317 icp->icp_ndevs += icp->icp_nchan; 318 } 319 320 icp_recompute_openings(icp); 321 322 /* 323 * Attach SCSI channels. 324 */ 325 if (icp->icp_features & ICP_FEAT_RAWSERVICE) { 326 struct icp_ioc_version *iv; 327 struct icp_rawioc *ri; 328 struct icp_getch *gc; 329 330 iv = (struct icp_ioc_version *)icp->icp_scr; 331 iv->iv_version = htole32(ICP_IOC_NEWEST); 332 iv->iv_listents = ICP_MAXBUS; 333 iv->iv_firstchan = 0; 334 iv->iv_lastchan = ICP_MAXBUS - 1; 335 iv->iv_listoffset = htole32(sizeof(*iv)); 336 337 if (icp_cmd(icp, ICP_CACHESERVICE, ICP_IOCTL, 338 ICP_IOCHAN_RAW_DESC, ICP_INVALID_CHANNEL, 339 sizeof(*iv) + ICP_MAXBUS * sizeof(*ri))) { 340 ri = (struct icp_rawioc *)(iv + 1); 341 for (j = 0; j < binfo.bi_chan_count; j++, ri++) 342 icp->icp_bus_id[j] = ri->ri_procid; 343 } else { 344 /* 345 * Fall back to the old method. 346 */ 347 gc = (struct icp_getch *)icp->icp_scr; 348 349 for (j = 0; j < binfo.bi_chan_count; j++) { 350 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_IOCTL, 351 ICP_SCSI_CHAN_CNT | ICP_L_CTRL_PATTERN, 352 ICP_IO_CHANNEL | ICP_INVALID_CHANNEL, 353 sizeof(*gc))) { 354 aprint_error_dev(icp->icp_dv, 355 "unable to get chan info"); 356 goto bail_out; 357 } 358 icp->icp_bus_id[j] = gc->gc_scsiid; 359 } 360 } 361 362 for (j = 0; j < binfo.bi_chan_count; j++) { 363 if (icp->icp_bus_id[j] > ICP_MAXID_FC) 364 icp->icp_bus_id[j] = ICP_MAXID_FC; 365 366 icpa.icpa_unit = j + ICPA_UNIT_SCSI; 367 368 locs[ICPCF_UNIT] = j + ICPA_UNIT_SCSI; 369 370 icp->icp_children[icpa.icpa_unit] = 371 config_found_sm_loc(icp->icp_dv, "icp", locs, 372 &icpa, icp_print, config_stdsubmatch); 373 } 374 } 375 376 /* 377 * Attach cache devices. 378 */ 379 if (icp->icp_features & ICP_FEAT_CACHESERVICE) { 380 for (j = 0; j < cdev_cnt && j < ICP_MAX_HDRIVES; j++) { 381 if (icp->icp_cdr[j].cd_size == 0) 382 continue; 383 384 icpa.icpa_unit = j; 385 386 locs[ICPCF_UNIT] = j; 387 388 icp->icp_children[icpa.icpa_unit] = 389 config_found_sm_loc(icp->icp_dv, "icp", locs, 390 &icpa, icp_print, config_stdsubmatch); 391 } 392 } 393 394 /* 395 * Start the watchdog. 396 */ 397 icp_watchdog(icp); 398 399 /* 400 * Count the controller, and we're done! 401 */ 402 if (icp_count++ == 0) 403 mutex_init(&icp_ioctl_mutex, MUTEX_DEFAULT, IPL_NONE); 404 405 return (0); 406 407 bail_out: 408 if (state > 4) 409 for (j = 0; j < i; j++) 410 bus_dmamap_destroy(icp->icp_dmat, 411 icp->icp_ccbs[j].ic_xfer_map); 412 if (state > 3) 413 free(icp->icp_ccbs, M_DEVBUF); 414 if (state > 2) 415 bus_dmamap_unload(icp->icp_dmat, icp->icp_scr_dmamap); 416 if (state > 1) 417 bus_dmamem_unmap(icp->icp_dmat, icp->icp_scr, 418 ICP_SCRATCH_SIZE); 419 if (state > 0) 420 bus_dmamem_free(icp->icp_dmat, icp->icp_scr_seg, nsegs); 421 bus_dmamap_destroy(icp->icp_dmat, icp->icp_scr_dmamap); 422 423 return (1); 424 } 425 426 void 427 icp_register_servicecb(struct icp_softc *icp, int unit, 428 const struct icp_servicecb *cb) 429 { 430 431 icp->icp_servicecb[unit] = cb; 432 } 433 434 void 435 icp_rescan(struct icp_softc *icp, int unit) 436 { 437 struct icp_attach_args icpa; 438 u_int newsize, newtype; 439 int locs[ICPCF_NLOCS]; 440 441 /* 442 * NOTE: It is very important that the queue be frozen and not 443 * commands running when this is called. The ioctl mutex must 444 * also be held. 445 */ 446 447 KASSERT(icp->icp_qfreeze != 0); 448 KASSERT(icp->icp_running == 0); 449 KASSERT(unit < ICP_MAX_HDRIVES); 450 451 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INFO, unit, 0, 0)) { 452 #ifdef ICP_DEBUG 453 printf("%s: rescan: unit %d ICP_INFO failed -> 0x%04x\n", 454 device_xname(icp->icp_dv), unit, icp->icp_status); 455 #endif 456 goto gone; 457 } 458 if ((newsize = icp->icp_info) == 0) { 459 #ifdef ICP_DEBUG 460 printf("%s: rescan: unit %d has zero size\n", 461 device_xname(icp->icp_dv), unit); 462 #endif 463 gone: 464 /* 465 * Host drive is no longer present; detach if a child 466 * is currently there. 467 */ 468 if (icp->icp_cdr[unit].cd_size != 0) 469 icp->icp_ndevs--; 470 icp->icp_cdr[unit].cd_size = 0; 471 if (icp->icp_children[unit] != NULL) { 472 (void) config_detach(icp->icp_children[unit], 473 DETACH_FORCE); 474 icp->icp_children[unit] = NULL; 475 } 476 return; 477 } 478 479 if (icp_cmd(icp, ICP_CACHESERVICE, ICP_DEVTYPE, unit, 0, 0)) 480 newtype = icp->icp_info; 481 else { 482 #ifdef ICP_DEBUG 483 printf("%s: rescan: unit %d ICP_DEVTYPE failed\n", 484 device_xname(icp->icp_dv), unit); 485 #endif 486 newtype = 0; /* XXX? */ 487 } 488 489 #ifdef ICP_DEBUG 490 printf("%s: rescan: unit %d old %u/%u, new %u/%u\n", 491 device_xname(icp->icp_dv), unit, icp->icp_cdr[unit].cd_size, 492 icp->icp_cdr[unit].cd_type, newsize, newtype); 493 #endif 494 495 /* 496 * If the type or size changed, detach any old child (if it exists) 497 * and attach a new one. 498 */ 499 if (icp->icp_children[unit] == NULL || 500 newsize != icp->icp_cdr[unit].cd_size || 501 newtype != icp->icp_cdr[unit].cd_type) { 502 if (icp->icp_cdr[unit].cd_size == 0) 503 icp->icp_ndevs++; 504 icp->icp_cdr[unit].cd_size = newsize; 505 icp->icp_cdr[unit].cd_type = newtype; 506 if (icp->icp_children[unit] != NULL) 507 (void) config_detach(icp->icp_children[unit], 508 DETACH_FORCE); 509 510 icpa.icpa_unit = unit; 511 512 locs[ICPCF_UNIT] = unit; 513 514 icp->icp_children[unit] = config_found_sm_loc(icp->icp_dv, 515 "icp", locs, &icpa, icp_print, config_stdsubmatch); 516 } 517 518 icp_recompute_openings(icp); 519 } 520 521 void 522 icp_rescan_all(struct icp_softc *icp) 523 { 524 int unit; 525 u_int16_t cdev_cnt; 526 527 /* 528 * This is the old method of rescanning the host drives. We 529 * start by reinitializing the cache service. 530 */ 531 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INIT, ICP_LINUX_OS, 0, 0)) { 532 printf("%s: unable to re-initialize cache service for rescan\n", 533 device_xname(icp->icp_dv)); 534 return; 535 } 536 cdev_cnt = (u_int16_t) icp->icp_info; 537 538 /* For each host drive, do the new-style rescan. */ 539 for (unit = 0; unit < cdev_cnt && unit < ICP_MAX_HDRIVES; unit++) 540 icp_rescan(icp, unit); 541 542 /* Now detach anything in the slots after cdev_cnt. */ 543 for (; unit < ICP_MAX_HDRIVES; unit++) { 544 if (icp->icp_cdr[unit].cd_size != 0) { 545 #ifdef ICP_DEBUG 546 printf("%s: rescan all: unit %d < new cdev_cnt (%d)\n", 547 device_xname(icp->icp_dv), unit, cdev_cnt); 548 #endif 549 icp->icp_ndevs--; 550 icp->icp_cdr[unit].cd_size = 0; 551 if (icp->icp_children[unit] != NULL) { 552 (void) config_detach(icp->icp_children[unit], 553 DETACH_FORCE); 554 icp->icp_children[unit] = NULL; 555 } 556 } 557 } 558 559 icp_recompute_openings(icp); 560 } 561 562 void 563 icp_recompute_openings(struct icp_softc *icp) 564 { 565 int unit, openings; 566 567 if (icp->icp_ndevs != 0) 568 openings = 569 (icp->icp_nccbs - ICP_NCCB_RESERVE) / icp->icp_ndevs; 570 else 571 openings = 0; 572 if (openings == icp->icp_openings) 573 return; 574 icp->icp_openings = openings; 575 576 #ifdef ICP_DEBUG 577 printf("%s: %d device%s, %d openings per device\n", 578 device_xname(icp->icp_dv), icp->icp_ndevs, 579 icp->icp_ndevs == 1 ? "" : "s", icp->icp_openings); 580 #endif 581 582 for (unit = 0; unit < ICP_MAX_HDRIVES + ICP_MAXBUS; unit++) { 583 if (icp->icp_children[unit] != NULL) 584 (*icp->icp_servicecb[unit]->iscb_openings)( 585 icp->icp_children[unit], icp->icp_openings); 586 } 587 } 588 589 void 590 icp_watchdog(void *cookie) 591 { 592 struct icp_softc *icp; 593 int s; 594 595 icp = cookie; 596 597 s = splbio(); 598 icp_intr(icp); 599 if (ICP_HAS_WORK(icp)) 600 icp_ccb_enqueue(icp, NULL); 601 splx(s); 602 603 callout_reset(&icp->icp_wdog_callout, hz * ICP_WATCHDOG_FREQ, 604 icp_watchdog, icp); 605 } 606 607 int 608 icp_print(void *aux, const char *pnp) 609 { 610 struct icp_attach_args *icpa; 611 const char *str; 612 613 icpa = (struct icp_attach_args *)aux; 614 615 if (pnp != NULL) { 616 if (icpa->icpa_unit < ICPA_UNIT_SCSI) 617 str = "block device"; 618 else 619 str = "SCSI channel"; 620 aprint_normal("%s at %s", str, pnp); 621 } 622 aprint_normal(" unit %d", icpa->icpa_unit); 623 624 return (UNCONF); 625 } 626 627 int 628 icp_async_event(struct icp_softc *icp, int service) 629 { 630 631 if (service == ICP_SCREENSERVICE) { 632 if (icp->icp_status == ICP_S_MSG_REQUEST) { 633 /* XXX */ 634 } 635 } else { 636 if ((icp->icp_fw_vers & 0xff) >= 0x1a) { 637 icp->icp_evt.size = 0; 638 icp->icp_evt.eu.async.ionode = 639 device_unit(icp->icp_dv); 640 icp->icp_evt.eu.async.status = icp->icp_status; 641 /* 642 * Severity and event string are filled in by the 643 * hardware interface interrupt handler. 644 */ 645 printf("%s: %s\n", device_xname(icp->icp_dv), 646 icp->icp_evt.event_string); 647 } else { 648 icp->icp_evt.size = sizeof(icp->icp_evt.eu.async); 649 icp->icp_evt.eu.async.ionode = 650 device_unit(icp->icp_dv); 651 icp->icp_evt.eu.async.service = service; 652 icp->icp_evt.eu.async.status = icp->icp_status; 653 icp->icp_evt.eu.async.info = icp->icp_info; 654 /* XXXJRT FIX THIS */ 655 *(u_int32_t *) icp->icp_evt.eu.async.scsi_coord = 656 icp->icp_info2; 657 } 658 icp_store_event(icp, GDT_ES_ASYNC, service, &icp->icp_evt); 659 } 660 661 return (0); 662 } 663 664 int 665 icp_intr(void *cookie) 666 { 667 struct icp_softc *icp; 668 struct icp_intr_ctx ctx; 669 struct icp_ccb *ic; 670 671 icp = cookie; 672 673 ctx.istatus = (*icp->icp_get_status)(icp); 674 if (!ctx.istatus) { 675 icp->icp_status = ICP_S_NO_STATUS; 676 return (0); 677 } 678 679 (*icp->icp_intr)(icp, &ctx); 680 681 icp->icp_status = ctx.cmd_status; 682 icp->icp_service = ctx.service; 683 icp->icp_info = ctx.info; 684 icp->icp_info2 = ctx.info2; 685 686 switch (ctx.istatus) { 687 case ICP_ASYNCINDEX: 688 icp_async_event(icp, ctx.service); 689 return (1); 690 691 case ICP_SPEZINDEX: 692 aprint_error_dev(icp->icp_dv, "uninitialized or unknown service (%d/%d)\n", 693 ctx.info, ctx.info2); 694 icp->icp_evt.size = sizeof(icp->icp_evt.eu.driver); 695 icp->icp_evt.eu.driver.ionode = device_unit(icp->icp_dv); 696 icp_store_event(icp, GDT_ES_DRIVER, 4, &icp->icp_evt); 697 return (1); 698 } 699 700 if ((ctx.istatus - 2) > icp->icp_nccbs) 701 panic("icp_intr: bad command index returned"); 702 703 ic = &icp->icp_ccbs[ctx.istatus - 2]; 704 ic->ic_status = icp->icp_status; 705 706 if ((ic->ic_flags & IC_ALLOCED) == 0) { 707 /* XXX ICP's "iir" driver just sends an event here. */ 708 panic("icp_intr: inactive CCB identified"); 709 } 710 711 /* 712 * Try to protect ourselves from the running command count already 713 * being 0 (e.g. if a polled command times out). 714 */ 715 KDASSERT(icp->icp_running != 0); 716 if (--icp->icp_running == 0 && 717 (icp->icp_flags & ICP_F_WAIT_FREEZE) != 0) { 718 icp->icp_flags &= ~ICP_F_WAIT_FREEZE; 719 wakeup(&icp->icp_qfreeze); 720 } 721 722 switch (icp->icp_status) { 723 case ICP_S_BSY: 724 #ifdef ICP_DEBUG 725 printf("%s: ICP_S_BSY received\n", device_xname(icp->icp_dv)); 726 #endif 727 if (__predict_false((ic->ic_flags & IC_UCMD) != 0)) 728 SIMPLEQ_INSERT_HEAD(&icp->icp_ucmd_queue, ic, ic_chain); 729 else 730 SIMPLEQ_INSERT_HEAD(&icp->icp_ccb_queue, ic, ic_chain); 731 break; 732 733 default: 734 ic->ic_flags |= IC_COMPLETE; 735 736 if ((ic->ic_flags & IC_WAITING) != 0) 737 wakeup(ic); 738 else if (ic->ic_intr != NULL) 739 (*ic->ic_intr)(ic); 740 741 if (ICP_HAS_WORK(icp)) 742 icp_ccb_enqueue(icp, NULL); 743 744 break; 745 } 746 747 return (1); 748 } 749 750 struct icp_ucmd_ctx { 751 gdt_ucmd_t *iu_ucmd; 752 u_int32_t iu_cnt; 753 }; 754 755 void 756 icp_ucmd_intr(struct icp_ccb *ic) 757 { 758 struct icp_softc *icp = device_private(ic->ic_dv); 759 struct icp_ucmd_ctx *iu = ic->ic_context; 760 gdt_ucmd_t *ucmd = iu->iu_ucmd; 761 762 ucmd->status = icp->icp_status; 763 ucmd->info = icp->icp_info; 764 765 if (iu->iu_cnt != 0) { 766 bus_dmamap_sync(icp->icp_dmat, 767 icp->icp_scr_dmamap, 768 ICP_SCRATCH_UCMD, iu->iu_cnt, 769 BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE); 770 memcpy(ucmd->data, 771 (char *)icp->icp_scr + ICP_SCRATCH_UCMD, iu->iu_cnt); 772 } 773 774 icp->icp_ucmd_ccb = NULL; 775 776 ic->ic_flags |= IC_COMPLETE; 777 wakeup(ic); 778 } 779 780 /* 781 * NOTE: We assume that it is safe to sleep here! 782 */ 783 int 784 icp_cmd(struct icp_softc *icp, u_int8_t service, u_int16_t opcode, 785 u_int32_t arg1, u_int32_t arg2, u_int32_t arg3) 786 { 787 struct icp_ioctlcmd *icmd; 788 struct icp_cachecmd *cc; 789 struct icp_rawcmd *rc; 790 int retries, rv; 791 struct icp_ccb *ic; 792 793 retries = ICP_RETRIES; 794 795 do { 796 ic = icp_ccb_alloc_wait(icp); 797 memset(&ic->ic_cmd, 0, sizeof(ic->ic_cmd)); 798 ic->ic_cmd.cmd_opcode = htole16(opcode); 799 800 switch (service) { 801 case ICP_CACHESERVICE: 802 if (opcode == ICP_IOCTL) { 803 icmd = &ic->ic_cmd.cmd_packet.ic; 804 icmd->ic_subfunc = htole16(arg1); 805 icmd->ic_channel = htole32(arg2); 806 icmd->ic_bufsize = htole32(arg3); 807 icmd->ic_addr = 808 htole32(icp->icp_scr_seg[0].ds_addr); 809 810 bus_dmamap_sync(icp->icp_dmat, 811 icp->icp_scr_dmamap, 0, arg3, 812 BUS_DMASYNC_PREWRITE | 813 BUS_DMASYNC_PREREAD); 814 } else { 815 cc = &ic->ic_cmd.cmd_packet.cc; 816 cc->cc_deviceno = htole16(arg1); 817 cc->cc_blockno = htole32(arg2); 818 } 819 break; 820 821 case ICP_SCSIRAWSERVICE: 822 rc = &ic->ic_cmd.cmd_packet.rc; 823 rc->rc_direction = htole32(arg1); 824 rc->rc_bus = arg2; 825 rc->rc_target = arg3; 826 rc->rc_lun = arg3 >> 8; 827 break; 828 } 829 830 ic->ic_service = service; 831 ic->ic_cmdlen = sizeof(ic->ic_cmd); 832 rv = icp_ccb_poll(icp, ic, 10000); 833 834 switch (service) { 835 case ICP_CACHESERVICE: 836 if (opcode == ICP_IOCTL) { 837 bus_dmamap_sync(icp->icp_dmat, 838 icp->icp_scr_dmamap, 0, arg3, 839 BUS_DMASYNC_POSTWRITE | 840 BUS_DMASYNC_POSTREAD); 841 } 842 break; 843 } 844 845 icp_ccb_free(icp, ic); 846 } while (rv != 0 && --retries > 0); 847 848 return (icp->icp_status == ICP_S_OK); 849 } 850 851 int 852 icp_ucmd(struct icp_softc *icp, gdt_ucmd_t *ucmd) 853 { 854 struct icp_ccb *ic; 855 struct icp_ucmd_ctx iu; 856 u_int32_t cnt; 857 int error; 858 859 if (ucmd->service == ICP_CACHESERVICE) { 860 if (ucmd->command.cmd_opcode == ICP_IOCTL) { 861 cnt = ucmd->command.cmd_packet.ic.ic_bufsize; 862 if (cnt > GDT_SCRATCH_SZ) { 863 aprint_error_dev(icp->icp_dv, "scratch buffer too small (%d/%d)\n", 864 GDT_SCRATCH_SZ, cnt); 865 return (EINVAL); 866 } 867 } else { 868 cnt = ucmd->command.cmd_packet.cc.cc_blockcnt * 869 ICP_SECTOR_SIZE; 870 if (cnt > GDT_SCRATCH_SZ) { 871 aprint_error_dev(icp->icp_dv, "scratch buffer too small (%d/%d)\n", 872 GDT_SCRATCH_SZ, cnt); 873 return (EINVAL); 874 } 875 } 876 } else { 877 cnt = ucmd->command.cmd_packet.rc.rc_sdlen + 878 ucmd->command.cmd_packet.rc.rc_sense_len; 879 if (cnt > GDT_SCRATCH_SZ) { 880 aprint_error_dev(icp->icp_dv, "scratch buffer too small (%d/%d)\n", 881 GDT_SCRATCH_SZ, cnt); 882 return (EINVAL); 883 } 884 } 885 886 iu.iu_ucmd = ucmd; 887 iu.iu_cnt = cnt; 888 889 ic = icp_ccb_alloc_wait(icp); 890 memset(&ic->ic_cmd, 0, sizeof(ic->ic_cmd)); 891 ic->ic_cmd.cmd_opcode = htole16(ucmd->command.cmd_opcode); 892 893 if (ucmd->service == ICP_CACHESERVICE) { 894 if (ucmd->command.cmd_opcode == ICP_IOCTL) { 895 struct icp_ioctlcmd *icmd, *uicmd; 896 897 icmd = &ic->ic_cmd.cmd_packet.ic; 898 uicmd = &ucmd->command.cmd_packet.ic; 899 900 icmd->ic_subfunc = htole16(uicmd->ic_subfunc); 901 icmd->ic_channel = htole32(uicmd->ic_channel); 902 icmd->ic_bufsize = htole32(uicmd->ic_bufsize); 903 icmd->ic_addr = 904 htole32(icp->icp_scr_seg[0].ds_addr + 905 ICP_SCRATCH_UCMD); 906 } else { 907 struct icp_cachecmd *cc, *ucc; 908 909 cc = &ic->ic_cmd.cmd_packet.cc; 910 ucc = &ucmd->command.cmd_packet.cc; 911 912 cc->cc_deviceno = htole16(ucc->cc_deviceno); 913 cc->cc_blockno = htole32(ucc->cc_blockno); 914 cc->cc_blockcnt = htole32(ucc->cc_blockcnt); 915 cc->cc_addr = htole32(0xffffffffU); 916 cc->cc_nsgent = htole32(1); 917 cc->cc_sg[0].sg_addr = 918 htole32(icp->icp_scr_seg[0].ds_addr + 919 ICP_SCRATCH_UCMD); 920 cc->cc_sg[0].sg_len = htole32(cnt); 921 } 922 } else { 923 struct icp_rawcmd *rc, *urc; 924 925 rc = &ic->ic_cmd.cmd_packet.rc; 926 urc = &ucmd->command.cmd_packet.rc; 927 928 rc->rc_direction = htole32(urc->rc_direction); 929 rc->rc_sdata = htole32(0xffffffffU); 930 rc->rc_sdlen = htole32(urc->rc_sdlen); 931 rc->rc_clen = htole32(urc->rc_clen); 932 memcpy(rc->rc_cdb, urc->rc_cdb, sizeof(rc->rc_cdb)); 933 rc->rc_target = urc->rc_target; 934 rc->rc_lun = urc->rc_lun; 935 rc->rc_bus = urc->rc_bus; 936 rc->rc_sense_len = htole32(urc->rc_sense_len); 937 rc->rc_sense_addr = 938 htole32(icp->icp_scr_seg[0].ds_addr + 939 ICP_SCRATCH_UCMD + urc->rc_sdlen); 940 rc->rc_nsgent = htole32(1); 941 rc->rc_sg[0].sg_addr = 942 htole32(icp->icp_scr_seg[0].ds_addr + ICP_SCRATCH_UCMD); 943 rc->rc_sg[0].sg_len = htole32(cnt - urc->rc_sense_len); 944 } 945 946 ic->ic_service = ucmd->service; 947 ic->ic_cmdlen = sizeof(ic->ic_cmd); 948 ic->ic_context = &iu; 949 950 /* 951 * XXX What units are ucmd->timeout in? Until we know, we 952 * XXX just pull a number out of thin air. 953 */ 954 if (__predict_false((error = icp_ccb_wait_user(icp, ic, 30000)) != 0)) 955 aprint_error_dev(icp->icp_dv, "error %d waiting for ucmd to complete\n", 956 error); 957 958 /* icp_ucmd_intr() has updated ucmd. */ 959 icp_ccb_free(icp, ic); 960 961 return (error); 962 } 963 964 struct icp_ccb * 965 icp_ccb_alloc(struct icp_softc *icp) 966 { 967 struct icp_ccb *ic; 968 int s; 969 970 s = splbio(); 971 if (__predict_false((ic = 972 SIMPLEQ_FIRST(&icp->icp_ccb_freelist)) == NULL)) { 973 splx(s); 974 return (NULL); 975 } 976 SIMPLEQ_REMOVE_HEAD(&icp->icp_ccb_freelist, ic_chain); 977 splx(s); 978 979 ic->ic_flags = IC_ALLOCED; 980 return (ic); 981 } 982 983 struct icp_ccb * 984 icp_ccb_alloc_wait(struct icp_softc *icp) 985 { 986 struct icp_ccb *ic; 987 int s; 988 989 s = splbio(); 990 while ((ic = SIMPLEQ_FIRST(&icp->icp_ccb_freelist)) == NULL) { 991 icp->icp_flags |= ICP_F_WAIT_CCB; 992 (void) tsleep(&icp->icp_ccb_freelist, PRIBIO, "icpccb", 0); 993 } 994 SIMPLEQ_REMOVE_HEAD(&icp->icp_ccb_freelist, ic_chain); 995 splx(s); 996 997 ic->ic_flags = IC_ALLOCED; 998 return (ic); 999 } 1000 1001 void 1002 icp_ccb_free(struct icp_softc *icp, struct icp_ccb *ic) 1003 { 1004 int s; 1005 1006 s = splbio(); 1007 ic->ic_flags = 0; 1008 ic->ic_intr = NULL; 1009 SIMPLEQ_INSERT_HEAD(&icp->icp_ccb_freelist, ic, ic_chain); 1010 if (__predict_false((icp->icp_flags & ICP_F_WAIT_CCB) != 0)) { 1011 icp->icp_flags &= ~ICP_F_WAIT_CCB; 1012 wakeup(&icp->icp_ccb_freelist); 1013 } 1014 splx(s); 1015 } 1016 1017 void 1018 icp_ccb_enqueue(struct icp_softc *icp, struct icp_ccb *ic) 1019 { 1020 int s; 1021 1022 s = splbio(); 1023 1024 if (ic != NULL) { 1025 if (__predict_false((ic->ic_flags & IC_UCMD) != 0)) 1026 SIMPLEQ_INSERT_TAIL(&icp->icp_ucmd_queue, ic, ic_chain); 1027 else 1028 SIMPLEQ_INSERT_TAIL(&icp->icp_ccb_queue, ic, ic_chain); 1029 } 1030 1031 for (; icp->icp_qfreeze == 0;) { 1032 if (__predict_false((ic = 1033 SIMPLEQ_FIRST(&icp->icp_ucmd_queue)) != NULL)) { 1034 struct icp_ucmd_ctx *iu = ic->ic_context; 1035 gdt_ucmd_t *ucmd = iu->iu_ucmd; 1036 1037 /* 1038 * All user-generated commands share the same 1039 * scratch space, so if one is already running, 1040 * we have to stall the command queue. 1041 */ 1042 if (icp->icp_ucmd_ccb != NULL) 1043 break; 1044 if ((*icp->icp_test_busy)(icp)) 1045 break; 1046 icp->icp_ucmd_ccb = ic; 1047 1048 if (iu->iu_cnt != 0) { 1049 memcpy((char *)icp->icp_scr + ICP_SCRATCH_UCMD, 1050 ucmd->data, iu->iu_cnt); 1051 bus_dmamap_sync(icp->icp_dmat, 1052 icp->icp_scr_dmamap, 1053 ICP_SCRATCH_UCMD, iu->iu_cnt, 1054 BUS_DMASYNC_PREREAD | 1055 BUS_DMASYNC_PREWRITE); 1056 } 1057 } else if (__predict_true((ic = 1058 SIMPLEQ_FIRST(&icp->icp_ccb_queue)) != NULL)) { 1059 if ((*icp->icp_test_busy)(icp)) 1060 break; 1061 } else { 1062 /* no command found */ 1063 break; 1064 } 1065 icp_ccb_submit(icp, ic); 1066 if (__predict_false((ic->ic_flags & IC_UCMD) != 0)) 1067 SIMPLEQ_REMOVE_HEAD(&icp->icp_ucmd_queue, ic_chain); 1068 else 1069 SIMPLEQ_REMOVE_HEAD(&icp->icp_ccb_queue, ic_chain); 1070 } 1071 1072 splx(s); 1073 } 1074 1075 int 1076 icp_ccb_map(struct icp_softc *icp, struct icp_ccb *ic, void *data, int size, 1077 int dir) 1078 { 1079 struct icp_sg *sg; 1080 int nsegs, i, rv; 1081 bus_dmamap_t xfer; 1082 1083 xfer = ic->ic_xfer_map; 1084 1085 rv = bus_dmamap_load(icp->icp_dmat, xfer, data, size, NULL, 1086 BUS_DMA_NOWAIT | BUS_DMA_STREAMING | 1087 ((dir & IC_XFER_IN) ? BUS_DMA_READ : BUS_DMA_WRITE)); 1088 if (rv != 0) 1089 return (rv); 1090 1091 nsegs = xfer->dm_nsegs; 1092 ic->ic_xfer_size = size; 1093 ic->ic_nsgent = nsegs; 1094 ic->ic_flags |= dir; 1095 sg = ic->ic_sg; 1096 1097 if (sg != NULL) { 1098 for (i = 0; i < nsegs; i++, sg++) { 1099 sg->sg_addr = htole32(xfer->dm_segs[i].ds_addr); 1100 sg->sg_len = htole32(xfer->dm_segs[i].ds_len); 1101 } 1102 } else if (nsegs > 1) 1103 panic("icp_ccb_map: no SG list specified, but nsegs > 1"); 1104 1105 if ((dir & IC_XFER_OUT) != 0) 1106 i = BUS_DMASYNC_PREWRITE; 1107 else /* if ((dir & IC_XFER_IN) != 0) */ 1108 i = BUS_DMASYNC_PREREAD; 1109 1110 bus_dmamap_sync(icp->icp_dmat, xfer, 0, ic->ic_xfer_size, i); 1111 return (0); 1112 } 1113 1114 void 1115 icp_ccb_unmap(struct icp_softc *icp, struct icp_ccb *ic) 1116 { 1117 int i; 1118 1119 if ((ic->ic_flags & IC_XFER_OUT) != 0) 1120 i = BUS_DMASYNC_POSTWRITE; 1121 else /* if ((ic->ic_flags & IC_XFER_IN) != 0) */ 1122 i = BUS_DMASYNC_POSTREAD; 1123 1124 bus_dmamap_sync(icp->icp_dmat, ic->ic_xfer_map, 0, ic->ic_xfer_size, i); 1125 bus_dmamap_unload(icp->icp_dmat, ic->ic_xfer_map); 1126 } 1127 1128 int 1129 icp_ccb_poll(struct icp_softc *icp, struct icp_ccb *ic, int timo) 1130 { 1131 int s, rv; 1132 1133 s = splbio(); 1134 1135 for (timo = ICP_BUSY_WAIT_MS * 100; timo != 0; timo--) { 1136 if (!(*icp->icp_test_busy)(icp)) 1137 break; 1138 DELAY(10); 1139 } 1140 if (timo == 0) { 1141 printf("%s: submit: busy\n", device_xname(icp->icp_dv)); 1142 return (EAGAIN); 1143 } 1144 1145 icp_ccb_submit(icp, ic); 1146 1147 if (cold) { 1148 for (timo *= 10; timo != 0; timo--) { 1149 DELAY(100); 1150 icp_intr(icp); 1151 if ((ic->ic_flags & IC_COMPLETE) != 0) 1152 break; 1153 } 1154 } else { 1155 ic->ic_flags |= IC_WAITING; 1156 while ((ic->ic_flags & IC_COMPLETE) == 0) { 1157 if ((rv = tsleep(ic, PRIBIO, "icpwccb", 1158 mstohz(timo))) != 0) { 1159 timo = 0; 1160 break; 1161 } 1162 } 1163 } 1164 1165 if (timo != 0) { 1166 if (ic->ic_status != ICP_S_OK) { 1167 #ifdef ICP_DEBUG 1168 printf("%s: request failed; status=0x%04x\n", 1169 device_xname(icp->icp_dv), ic->ic_status); 1170 #endif 1171 rv = EIO; 1172 } else 1173 rv = 0; 1174 } else { 1175 aprint_error_dev(icp->icp_dv, "command timed out\n"); 1176 rv = EIO; 1177 } 1178 1179 while ((*icp->icp_test_busy)(icp) != 0) 1180 DELAY(10); 1181 1182 splx(s); 1183 1184 return (rv); 1185 } 1186 1187 int 1188 icp_ccb_wait(struct icp_softc *icp, struct icp_ccb *ic, int timo) 1189 { 1190 int s, rv; 1191 1192 ic->ic_flags |= IC_WAITING; 1193 1194 s = splbio(); 1195 icp_ccb_enqueue(icp, ic); 1196 while ((ic->ic_flags & IC_COMPLETE) == 0) { 1197 if ((rv = tsleep(ic, PRIBIO, "icpwccb", mstohz(timo))) != 0) { 1198 splx(s); 1199 return (rv); 1200 } 1201 } 1202 splx(s); 1203 1204 if (ic->ic_status != ICP_S_OK) { 1205 aprint_error_dev(icp->icp_dv, "command failed; status=%x\n", 1206 ic->ic_status); 1207 return (EIO); 1208 } 1209 1210 return (0); 1211 } 1212 1213 int 1214 icp_ccb_wait_user(struct icp_softc *icp, struct icp_ccb *ic, int timo) 1215 { 1216 int s, rv; 1217 1218 ic->ic_dv = icp->icp_dv; 1219 ic->ic_intr = icp_ucmd_intr; 1220 ic->ic_flags |= IC_UCMD; 1221 1222 s = splbio(); 1223 icp_ccb_enqueue(icp, ic); 1224 while ((ic->ic_flags & IC_COMPLETE) == 0) { 1225 if ((rv = tsleep(ic, PRIBIO, "icpwuccb", mstohz(timo))) != 0) { 1226 splx(s); 1227 return (rv); 1228 } 1229 } 1230 splx(s); 1231 1232 return (0); 1233 } 1234 1235 void 1236 icp_ccb_submit(struct icp_softc *icp, struct icp_ccb *ic) 1237 { 1238 1239 ic->ic_cmdlen = (ic->ic_cmdlen + 3) & ~3; 1240 1241 (*icp->icp_set_sema0)(icp); 1242 DELAY(10); 1243 1244 ic->ic_cmd.cmd_boardnode = htole32(ICP_LOCALBOARD); 1245 ic->ic_cmd.cmd_cmdindex = htole32(ic->ic_ident); 1246 1247 icp->icp_running++; 1248 1249 (*icp->icp_copy_cmd)(icp, ic); 1250 (*icp->icp_release_event)(icp, ic); 1251 } 1252 1253 int 1254 icp_freeze(struct icp_softc *icp) 1255 { 1256 int s, error = 0; 1257 1258 s = splbio(); 1259 if (icp->icp_qfreeze++ == 0) { 1260 while (icp->icp_running != 0) { 1261 icp->icp_flags |= ICP_F_WAIT_FREEZE; 1262 error = tsleep(&icp->icp_qfreeze, PRIBIO|PCATCH, 1263 "icpqfrz", 0); 1264 if (error != 0 && --icp->icp_qfreeze == 0 && 1265 ICP_HAS_WORK(icp)) { 1266 icp_ccb_enqueue(icp, NULL); 1267 break; 1268 } 1269 } 1270 } 1271 splx(s); 1272 1273 return (error); 1274 } 1275 1276 void 1277 icp_unfreeze(struct icp_softc *icp) 1278 { 1279 int s; 1280 1281 s = splbio(); 1282 KDASSERT(icp->icp_qfreeze != 0); 1283 if (--icp->icp_qfreeze == 0 && ICP_HAS_WORK(icp)) 1284 icp_ccb_enqueue(icp, NULL); 1285 splx(s); 1286 } 1287 1288 /* XXX Global - should be per-controller? XXX */ 1289 static gdt_evt_str icp_event_buffer[ICP_MAX_EVENTS]; 1290 static int icp_event_oldidx; 1291 static int icp_event_lastidx; 1292 1293 gdt_evt_str * 1294 icp_store_event(struct icp_softc *icp, u_int16_t source, u_int16_t idx, 1295 gdt_evt_data *evt) 1296 { 1297 gdt_evt_str *e; 1298 1299 /* no source == no event */ 1300 if (source == 0) 1301 return (NULL); 1302 1303 e = &icp_event_buffer[icp_event_lastidx]; 1304 if (e->event_source == source && e->event_idx == idx && 1305 ((evt->size != 0 && e->event_data.size != 0 && 1306 memcmp(&e->event_data.eu, &evt->eu, evt->size) == 0) || 1307 (evt->size == 0 && e->event_data.size == 0 && 1308 strcmp((char *) e->event_data.event_string, 1309 (char *) evt->event_string) == 0))) { 1310 e->last_stamp = time_second; 1311 e->same_count++; 1312 } else { 1313 if (icp_event_buffer[icp_event_lastidx].event_source != 0) { 1314 icp_event_lastidx++; 1315 if (icp_event_lastidx == ICP_MAX_EVENTS) 1316 icp_event_lastidx = 0; 1317 if (icp_event_lastidx == icp_event_oldidx) { 1318 icp_event_oldidx++; 1319 if (icp_event_oldidx == ICP_MAX_EVENTS) 1320 icp_event_oldidx = 0; 1321 } 1322 } 1323 e = &icp_event_buffer[icp_event_lastidx]; 1324 e->event_source = source; 1325 e->event_idx = idx; 1326 e->first_stamp = e->last_stamp = time_second; 1327 e->same_count = 1; 1328 e->event_data = *evt; 1329 e->application = 0; 1330 } 1331 return (e); 1332 } 1333 1334 int 1335 icp_read_event(struct icp_softc *icp, int handle, gdt_evt_str *estr) 1336 { 1337 gdt_evt_str *e; 1338 int eindex, s; 1339 1340 s = splbio(); 1341 1342 if (handle == -1) 1343 eindex = icp_event_oldidx; 1344 else 1345 eindex = handle; 1346 1347 estr->event_source = 0; 1348 1349 if (eindex < 0 || eindex >= ICP_MAX_EVENTS) { 1350 splx(s); 1351 return (eindex); 1352 } 1353 1354 e = &icp_event_buffer[eindex]; 1355 if (e->event_source != 0) { 1356 if (eindex != icp_event_lastidx) { 1357 eindex++; 1358 if (eindex == ICP_MAX_EVENTS) 1359 eindex = 0; 1360 } else 1361 eindex = -1; 1362 memcpy(estr, e, sizeof(gdt_evt_str)); 1363 } 1364 1365 splx(s); 1366 1367 return (eindex); 1368 } 1369 1370 void 1371 icp_readapp_event(struct icp_softc *icp, u_int8_t application, 1372 gdt_evt_str *estr) 1373 { 1374 gdt_evt_str *e; 1375 int found = 0, eindex, s; 1376 1377 s = splbio(); 1378 1379 eindex = icp_event_oldidx; 1380 for (;;) { 1381 e = &icp_event_buffer[eindex]; 1382 if (e->event_source == 0) 1383 break; 1384 if ((e->application & application) == 0) { 1385 e->application |= application; 1386 found = 1; 1387 break; 1388 } 1389 if (eindex == icp_event_lastidx) 1390 break; 1391 eindex++; 1392 if (eindex == ICP_MAX_EVENTS) 1393 eindex = 0; 1394 } 1395 if (found) 1396 memcpy(estr, e, sizeof(gdt_evt_str)); 1397 else 1398 estr->event_source = 0; 1399 1400 splx(s); 1401 } 1402 1403 void 1404 icp_clear_events(struct icp_softc *icp) 1405 { 1406 int s; 1407 1408 s = splbio(); 1409 icp_event_oldidx = icp_event_lastidx = 0; 1410 memset(icp_event_buffer, 0, sizeof(icp_event_buffer)); 1411 splx(s); 1412 } 1413