1 /* 2 * Copyright (c) 1997 by Simon Shapiro 3 * All Rights Reserved 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright 9 * notice, this list of conditions, and the following disclaimer, 10 * without modification, immediately at the beginning of the file. 11 * 2. Redistributions in binary form must reproduce the above copyright 12 * notice, this list of conditions and the following disclaimer in the 13 * documentation and/or other materials provided with the distribution. 14 * 3. The name of the author may not be used to endorse or promote products 15 * derived from this software without specific prior written permission. 16 * 17 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 18 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 19 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 20 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE FOR 21 * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 22 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 23 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 24 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 25 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 26 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 27 * SUCH DAMAGE. 28 */ 29 30 /* 31 * dpt_scsi.c: SCSI dependant code for the DPT driver 32 * 33 * credits: Assisted by Mike Neuffer in the early low level DPT code 34 * Thanx to Mark Salyzyn of DPT for his assistance. 35 * Special thanx to Justin Gibbs for invaluable help in 36 * making this driver look and work like a FreeBSD component. 37 * Last but not least, many thanx to UCB and the FreeBSD 38 * team for creating and maintaining such a wonderful O/S. 39 * 40 * TODO: * Add ISA probe code. 41 * * Add driver-level RAID-0. This will allow interoperability with 42 * NiceTry, M$-Doze, Win-Dog, Slowlaris, etc., in recognizing RAID 43 * arrays that span controllers (Wow!). 44 */ 45 46 #ident "$FreeBSD: src/sys/dev/dpt/dpt_scsi.c,v 1.28.2.3 2003/01/31 02:47:10 grog Exp $" 47 #ident "$DragonFly: src/sys/dev/raid/dpt/dpt_scsi.c,v 1.17 2008/02/10 00:01:03 pavalos Exp $" 48 49 #define _DPT_C_ 50 51 #include "opt_dpt.h" 52 #include <sys/param.h> 53 #include <sys/systm.h> 54 #include <sys/eventhandler.h> 55 #include <sys/malloc.h> 56 #include <sys/kernel.h> 57 #include <sys/bus.h> 58 #include <sys/thread2.h> 59 60 #include <machine/clock.h> 61 62 #include <bus/cam/cam.h> 63 #include <bus/cam/cam_ccb.h> 64 #include <bus/cam/cam_sim.h> 65 #include <bus/cam/cam_xpt_sim.h> 66 #include <bus/cam/cam_debug.h> 67 #include <bus/cam/scsi/scsi_all.h> 68 #include <bus/cam/scsi/scsi_message.h> 69 70 #include <vm/vm.h> 71 #include <vm/pmap.h> 72 73 #include "dpt.h" 74 75 /* dpt_isa.c, dpt_eisa.c, and dpt_pci.c need this in a central place */ 76 int dpt_controllers_present; 77 78 u_long dpt_unit; /* Next unit number to use */ 79 80 /* The linked list of softc structures */ 81 struct dpt_softc_list dpt_softcs = TAILQ_HEAD_INITIALIZER(dpt_softcs); 82 83 #define microtime_now dpt_time_now() 84 85 #define dpt_inl(dpt, port) \ 86 bus_space_read_4((dpt)->tag, (dpt)->bsh, port) 87 #define dpt_inb(dpt, port) \ 88 bus_space_read_1((dpt)->tag, (dpt)->bsh, port) 89 #define dpt_outl(dpt, port, value) \ 90 bus_space_write_4((dpt)->tag, (dpt)->bsh, port, value) 91 #define dpt_outb(dpt, port, value) \ 92 bus_space_write_1((dpt)->tag, (dpt)->bsh, port, value) 93 94 /* 95 * These will have to be setup by parameters passed at boot/load time. For 96 * perfromance reasons, we make them constants for the time being. 97 */ 98 #define dpt_min_segs DPT_MAX_SEGS 99 #define dpt_max_segs DPT_MAX_SEGS 100 101 /* Definitions for our use of the SIM private CCB area */ 102 #define ccb_dccb_ptr spriv_ptr0 103 #define ccb_dpt_ptr spriv_ptr1 104 105 /* ================= Private Inline Function declarations ===================*/ 106 static __inline int dpt_just_reset(dpt_softc_t * dpt); 107 static __inline int dpt_raid_busy(dpt_softc_t * dpt); 108 static __inline int dpt_pio_wait (u_int32_t, u_int, u_int, u_int); 109 static __inline int dpt_wait(dpt_softc_t *dpt, u_int bits, 110 u_int state); 111 static __inline struct dpt_ccb* dptgetccb(struct dpt_softc *dpt); 112 static __inline void dptfreeccb(struct dpt_softc *dpt, 113 struct dpt_ccb *dccb); 114 static __inline u_int32_t dptccbvtop(struct dpt_softc *dpt, 115 struct dpt_ccb *dccb); 116 117 static __inline int dpt_send_immediate(dpt_softc_t *dpt, 118 eata_ccb_t *cmd_block, 119 u_int32_t cmd_busaddr, 120 u_int retries, 121 u_int ifc, u_int code, 122 u_int code2); 123 124 /* ==================== Private Function declarations =======================*/ 125 static void dptmapmem(void *arg, bus_dma_segment_t *segs, 126 int nseg, int error); 127 128 static struct sg_map_node* 129 dptallocsgmap(struct dpt_softc *dpt); 130 131 static int dptallocccbs(dpt_softc_t *dpt); 132 133 static int dpt_get_conf(dpt_softc_t *dpt, dpt_ccb_t *dccb, 134 u_int32_t dccb_busaddr, u_int size, 135 u_int page, u_int target, int extent); 136 static void dpt_detect_cache(dpt_softc_t *dpt, dpt_ccb_t *dccb, 137 u_int32_t dccb_busaddr, 138 u_int8_t *buff); 139 140 static void dpt_poll(struct cam_sim *sim); 141 142 static void dptexecuteccb(void *arg, bus_dma_segment_t *dm_segs, 143 int nseg, int error); 144 145 static void dpt_action(struct cam_sim *sim, union ccb *ccb); 146 147 static int dpt_send_eata_command(dpt_softc_t *dpt, eata_ccb_t *cmd, 148 u_int32_t cmd_busaddr, 149 u_int command, u_int retries, 150 u_int ifc, u_int code, 151 u_int code2); 152 static void dptprocesserror(dpt_softc_t *dpt, dpt_ccb_t *dccb, 153 union ccb *ccb, u_int hba_stat, 154 u_int scsi_stat, u_int32_t resid); 155 156 static void dpttimeout(void *arg); 157 static void dptshutdown(void *arg, int howto); 158 159 /* ================= Private Inline Function definitions ====================*/ 160 static __inline int 161 dpt_just_reset(dpt_softc_t * dpt) 162 { 163 if ((dpt_inb(dpt, 2) == 'D') 164 && (dpt_inb(dpt, 3) == 'P') 165 && (dpt_inb(dpt, 4) == 'T') 166 && (dpt_inb(dpt, 5) == 'H')) 167 return (1); 168 else 169 return (0); 170 } 171 172 static __inline int 173 dpt_raid_busy(dpt_softc_t * dpt) 174 { 175 if ((dpt_inb(dpt, 0) == 'D') 176 && (dpt_inb(dpt, 1) == 'P') 177 && (dpt_inb(dpt, 2) == 'T')) 178 return (1); 179 else 180 return (0); 181 } 182 183 static __inline int 184 dpt_pio_wait (u_int32_t base, u_int reg, u_int bits, u_int state) 185 { 186 int i; 187 u_int c; 188 189 for (i = 0; i < 20000; i++) { /* wait 20ms for not busy */ 190 c = inb(base + reg) & bits; 191 if (!(c == state)) 192 return (0); 193 else 194 DELAY(50); 195 } 196 return (-1); 197 } 198 199 static __inline int 200 dpt_wait(dpt_softc_t *dpt, u_int bits, u_int state) 201 { 202 int i; 203 u_int c; 204 205 for (i = 0; i < 20000; i++) { /* wait 20ms for not busy */ 206 c = dpt_inb(dpt, HA_RSTATUS) & bits; 207 if (c == state) 208 return (0); 209 else 210 DELAY(50); 211 } 212 return (-1); 213 } 214 215 static __inline struct dpt_ccb* 216 dptgetccb(struct dpt_softc *dpt) 217 { 218 struct dpt_ccb* dccb; 219 220 crit_enter(); 221 if ((dccb = SLIST_FIRST(&dpt->free_dccb_list)) != NULL) { 222 SLIST_REMOVE_HEAD(&dpt->free_dccb_list, links); 223 dpt->free_dccbs--; 224 } else if (dpt->total_dccbs < dpt->max_dccbs) { 225 dptallocccbs(dpt); 226 dccb = SLIST_FIRST(&dpt->free_dccb_list); 227 if (dccb == NULL) 228 kprintf("dpt%d: Can't malloc DCCB\n", dpt->unit); 229 else { 230 SLIST_REMOVE_HEAD(&dpt->free_dccb_list, links); 231 dpt->free_dccbs--; 232 } 233 } 234 crit_exit(); 235 236 return (dccb); 237 } 238 239 static __inline void 240 dptfreeccb(struct dpt_softc *dpt, struct dpt_ccb *dccb) 241 { 242 crit_enter(); 243 if ((dccb->state & DCCB_ACTIVE) != 0) 244 LIST_REMOVE(&dccb->ccb->ccb_h, sim_links.le); 245 if ((dccb->state & DCCB_RELEASE_SIMQ) != 0) 246 dccb->ccb->ccb_h.status |= CAM_RELEASE_SIMQ; 247 else if (dpt->resource_shortage != 0 248 && (dccb->ccb->ccb_h.status & CAM_RELEASE_SIMQ) == 0) { 249 dccb->ccb->ccb_h.status |= CAM_RELEASE_SIMQ; 250 dpt->resource_shortage = FALSE; 251 } 252 dccb->state = DCCB_FREE; 253 SLIST_INSERT_HEAD(&dpt->free_dccb_list, dccb, links); 254 ++dpt->free_dccbs; 255 crit_exit(); 256 } 257 258 static __inline u_int32_t 259 dptccbvtop(struct dpt_softc *dpt, struct dpt_ccb *dccb) 260 { 261 return (dpt->dpt_ccb_busbase 262 + (u_int32_t)((caddr_t)dccb - (caddr_t)dpt->dpt_dccbs)); 263 } 264 265 static __inline struct dpt_ccb * 266 dptccbptov(struct dpt_softc *dpt, u_int32_t busaddr) 267 { 268 return (dpt->dpt_dccbs 269 + ((struct dpt_ccb *)busaddr 270 - (struct dpt_ccb *)dpt->dpt_ccb_busbase)); 271 } 272 273 /* 274 * Send a command for immediate execution by the DPT 275 * See above function for IMPORTANT notes. 276 */ 277 static __inline int 278 dpt_send_immediate(dpt_softc_t *dpt, eata_ccb_t *cmd_block, 279 u_int32_t cmd_busaddr, u_int retries, 280 u_int ifc, u_int code, u_int code2) 281 { 282 return (dpt_send_eata_command(dpt, cmd_block, cmd_busaddr, 283 EATA_CMD_IMMEDIATE, retries, ifc, 284 code, code2)); 285 } 286 287 288 /* ===================== Private Function definitions =======================*/ 289 static void 290 dptmapmem(void *arg, bus_dma_segment_t *segs, int nseg, int error) 291 { 292 bus_addr_t *busaddrp; 293 294 busaddrp = (bus_addr_t *)arg; 295 *busaddrp = segs->ds_addr; 296 } 297 298 static struct sg_map_node * 299 dptallocsgmap(struct dpt_softc *dpt) 300 { 301 struct sg_map_node *sg_map; 302 303 sg_map = kmalloc(sizeof(*sg_map), M_DEVBUF, M_INTWAIT); 304 305 /* Allocate S/G space for the next batch of CCBS */ 306 if (bus_dmamem_alloc(dpt->sg_dmat, (void **)&sg_map->sg_vaddr, 307 BUS_DMA_NOWAIT, &sg_map->sg_dmamap) != 0) { 308 kfree(sg_map, M_DEVBUF); 309 return (NULL); 310 } 311 312 (void)bus_dmamap_load(dpt->sg_dmat, sg_map->sg_dmamap, sg_map->sg_vaddr, 313 PAGE_SIZE, dptmapmem, &sg_map->sg_physaddr, 314 /*flags*/0); 315 316 SLIST_INSERT_HEAD(&dpt->sg_maps, sg_map, links); 317 318 return (sg_map); 319 } 320 321 /* 322 * Allocate another chunk of CCB's. Return count of entries added. 323 * Assumed to be called at splcam(). 324 */ 325 static int 326 dptallocccbs(dpt_softc_t *dpt) 327 { 328 struct dpt_ccb *next_ccb; 329 struct sg_map_node *sg_map; 330 bus_addr_t physaddr; 331 dpt_sg_t *segs; 332 int newcount; 333 int i; 334 335 next_ccb = &dpt->dpt_dccbs[dpt->total_dccbs]; 336 337 if (next_ccb == dpt->dpt_dccbs) { 338 /* 339 * First time through. Re-use the S/G 340 * space we allocated for initialization 341 * CCBS. 342 */ 343 sg_map = SLIST_FIRST(&dpt->sg_maps); 344 } else { 345 sg_map = dptallocsgmap(dpt); 346 } 347 348 if (sg_map == NULL) 349 return (0); 350 351 segs = sg_map->sg_vaddr; 352 physaddr = sg_map->sg_physaddr; 353 354 newcount = (PAGE_SIZE / (dpt->sgsize * sizeof(dpt_sg_t))); 355 for (i = 0; dpt->total_dccbs < dpt->max_dccbs && i < newcount; i++) { 356 int error; 357 358 error = bus_dmamap_create(dpt->buffer_dmat, /*flags*/0, 359 &next_ccb->dmamap); 360 if (error != 0) 361 break; 362 next_ccb->sg_list = segs; 363 next_ccb->sg_busaddr = htonl(physaddr); 364 next_ccb->eata_ccb.cp_dataDMA = htonl(physaddr); 365 next_ccb->eata_ccb.cp_statDMA = htonl(dpt->sp_physaddr); 366 next_ccb->eata_ccb.cp_reqDMA = 367 htonl(dptccbvtop(dpt, next_ccb) 368 + offsetof(struct dpt_ccb, sense_data)); 369 next_ccb->eata_ccb.cp_busaddr = dpt->dpt_ccb_busend; 370 next_ccb->state = DCCB_FREE; 371 next_ccb->tag = dpt->total_dccbs; 372 SLIST_INSERT_HEAD(&dpt->free_dccb_list, next_ccb, links); 373 segs += dpt->sgsize; 374 physaddr += (dpt->sgsize * sizeof(dpt_sg_t)); 375 dpt->dpt_ccb_busend += sizeof(*next_ccb); 376 next_ccb++; 377 dpt->total_dccbs++; 378 } 379 return (i); 380 } 381 382 dpt_conf_t * 383 dpt_pio_get_conf (u_int32_t base) 384 { 385 static dpt_conf_t * conf; 386 u_int16_t * p; 387 int i; 388 389 /* 390 * Allocate a dpt_conf_t 391 */ 392 if (conf == NULL) 393 conf = kmalloc(sizeof(dpt_conf_t), M_DEVBUF, M_INTWAIT); 394 395 /* 396 * If we have one, clean it up. 397 */ 398 bzero(conf, sizeof(dpt_conf_t)); 399 400 /* 401 * Reset the controller. 402 */ 403 outb((base + HA_WCOMMAND), EATA_CMD_RESET); 404 405 /* 406 * Wait for the controller to become ready. 407 * For some reason there can be -no- delays after calling reset 408 * before we wait on ready status. 409 */ 410 if (dpt_pio_wait(base, HA_RSTATUS, HA_SBUSY, 0)) { 411 kprintf("dpt: timeout waiting for controller to become ready\n"); 412 return (NULL); 413 } 414 415 if (dpt_pio_wait(base, HA_RAUXSTAT, HA_ABUSY, 0)) { 416 kprintf("dpt: timetout waiting for adapter ready.\n"); 417 return (NULL); 418 } 419 420 /* 421 * Send the PIO_READ_CONFIG command. 422 */ 423 outb((base + HA_WCOMMAND), EATA_CMD_PIO_READ_CONFIG); 424 425 /* 426 * Read the data into the struct. 427 */ 428 p = (u_int16_t *)conf; 429 for (i = 0; i < (sizeof(dpt_conf_t) / 2); i++) { 430 431 if (dpt_pio_wait(base, HA_RSTATUS, HA_SDRQ, 0)) { 432 kprintf("dpt: timeout in data read.\n"); 433 return (NULL); 434 } 435 436 (*p) = inw(base + HA_RDATA); 437 p++; 438 } 439 440 if (inb(base + HA_RSTATUS) & HA_SERROR) { 441 kprintf("dpt: error reading configuration data.\n"); 442 return (NULL); 443 } 444 445 #define BE_EATA_SIGNATURE 0x45415441 446 #define LE_EATA_SIGNATURE 0x41544145 447 448 /* 449 * Test to see if we have a valid card. 450 */ 451 if ((conf->signature == BE_EATA_SIGNATURE) || 452 (conf->signature == LE_EATA_SIGNATURE)) { 453 454 while (inb(base + HA_RSTATUS) & HA_SDRQ) { 455 inw(base + HA_RDATA); 456 } 457 458 return (conf); 459 } 460 return (NULL); 461 } 462 463 /* 464 * Read a configuration page into the supplied dpt_cont_t buffer. 465 */ 466 static int 467 dpt_get_conf(dpt_softc_t *dpt, dpt_ccb_t *dccb, u_int32_t dccb_busaddr, 468 u_int size, u_int page, u_int target, int extent) 469 { 470 eata_ccb_t *cp; 471 472 u_int8_t status; 473 474 int ndx; 475 int result; 476 477 cp = &dccb->eata_ccb; 478 bzero((void *)(uintptr_t)(volatile void *)dpt->sp, sizeof(*dpt->sp)); 479 480 cp->Interpret = 1; 481 cp->DataIn = 1; 482 cp->Auto_Req_Sen = 1; 483 cp->reqlen = sizeof(struct scsi_sense_data); 484 485 cp->cp_id = target; 486 cp->cp_LUN = 0; /* In the EATA packet */ 487 cp->cp_lun = 0; /* In the SCSI command */ 488 489 cp->cp_scsi_cmd = INQUIRY; 490 cp->cp_len = size; 491 492 cp->cp_extent = extent; 493 494 cp->cp_page = page; 495 cp->cp_channel = 0; /* DNC, Interpret mode is set */ 496 cp->cp_identify = 1; 497 cp->cp_datalen = htonl(size); 498 499 crit_enter(); 500 501 /* 502 * This could be a simple for loop, but we suspected the compiler To 503 * have optimized it a bit too much. Wait for the controller to 504 * become ready 505 */ 506 while (((status = dpt_inb(dpt, HA_RSTATUS)) != (HA_SREADY | HA_SSC) 507 && (status != (HA_SREADY | HA_SSC | HA_SERROR)) 508 && (status != (HA_SDRDY | HA_SERROR | HA_SDRQ))) 509 || (dpt_wait(dpt, HA_SBUSY, 0))) { 510 511 /* 512 * RAID Drives still Spinning up? (This should only occur if 513 * the DPT controller is in a NON PC (PCI?) platform). 514 */ 515 if (dpt_raid_busy(dpt)) { 516 kprintf("dpt%d WARNING: Get_conf() RSUS failed.\n", 517 dpt->unit); 518 crit_exit(); 519 return (0); 520 } 521 } 522 523 DptStat_Reset_BUSY(dpt->sp); 524 525 /* 526 * XXXX We might want to do something more clever than aborting at 527 * this point, like resetting (rebooting) the controller and trying 528 * again. 529 */ 530 if ((result = dpt_send_eata_command(dpt, cp, dccb_busaddr, 531 EATA_CMD_DMA_SEND_CP, 532 10000, 0, 0, 0)) != 0) { 533 kprintf("dpt%d WARNING: Get_conf() failed (%d) to send " 534 "EATA_CMD_DMA_READ_CONFIG\n", 535 dpt->unit, result); 536 crit_exit(); 537 return (0); 538 } 539 /* Wait for two seconds for a response. This can be slow */ 540 for (ndx = 0; 541 (ndx < 20000) 542 && !((status = dpt_inb(dpt, HA_RAUXSTAT)) & HA_AIRQ); 543 ndx++) { 544 DELAY(50); 545 } 546 547 /* Grab the status and clear interrupts */ 548 status = dpt_inb(dpt, HA_RSTATUS); 549 550 crit_exit(); 551 552 /* 553 * Check the status carefully. Return only if the 554 * command was successful. 555 */ 556 if (((status & HA_SERROR) == 0) 557 && (dpt->sp->hba_stat == 0) 558 && (dpt->sp->scsi_stat == 0) 559 && (dpt->sp->residue_len == 0)) 560 return (0); 561 562 if (dpt->sp->scsi_stat == SCSI_STATUS_CHECK_COND) 563 return (0); 564 565 return (1); 566 } 567 568 /* Detect Cache parameters and size */ 569 static void 570 dpt_detect_cache(dpt_softc_t *dpt, dpt_ccb_t *dccb, u_int32_t dccb_busaddr, 571 u_int8_t *buff) 572 { 573 eata_ccb_t *cp; 574 u_int8_t *param; 575 int bytes; 576 int result; 577 int ndx; 578 u_int8_t status; 579 580 /* 581 * Default setting, for best perfromance.. 582 * This is what virtually all cards default to.. 583 */ 584 dpt->cache_type = DPT_CACHE_WRITEBACK; 585 dpt->cache_size = 0; 586 587 cp = &dccb->eata_ccb; 588 bzero((void *)(uintptr_t)(volatile void *)dpt->sp, sizeof(dpt->sp)); 589 bzero(buff, 512); 590 591 /* Setup the command structure */ 592 cp->Interpret = 1; 593 cp->DataIn = 1; 594 cp->Auto_Req_Sen = 1; 595 cp->reqlen = sizeof(struct scsi_sense_data); 596 597 cp->cp_id = 0; /* who cares? The HBA will interpret.. */ 598 cp->cp_LUN = 0; /* In the EATA packet */ 599 cp->cp_lun = 0; /* In the SCSI command */ 600 cp->cp_channel = 0; 601 602 cp->cp_scsi_cmd = EATA_CMD_DMA_SEND_CP; 603 cp->cp_len = 56; 604 605 cp->cp_extent = 0; 606 cp->cp_page = 0; 607 cp->cp_identify = 1; 608 cp->cp_dispri = 1; 609 610 /* 611 * Build the EATA Command Packet structure 612 * for a Log Sense Command. 613 */ 614 cp->cp_cdb[0] = 0x4d; 615 cp->cp_cdb[1] = 0x0; 616 cp->cp_cdb[2] = 0x40 | 0x33; 617 cp->cp_cdb[7] = 1; 618 619 cp->cp_datalen = htonl(512); 620 621 crit_enter(); 622 result = dpt_send_eata_command(dpt, cp, dccb_busaddr, 623 EATA_CMD_DMA_SEND_CP, 624 10000, 0, 0, 0); 625 if (result != 0) { 626 kprintf("dpt%d WARNING: detect_cache() failed (%d) to send " 627 "EATA_CMD_DMA_SEND_CP\n", dpt->unit, result); 628 crit_exit(); 629 return; 630 } 631 /* Wait for two seconds for a response. This can be slow... */ 632 for (ndx = 0; 633 (ndx < 20000) && 634 !((status = dpt_inb(dpt, HA_RAUXSTAT)) & HA_AIRQ); 635 ndx++) { 636 DELAY(50); 637 } 638 639 /* Grab the status and clear interrupts */ 640 status = dpt_inb(dpt, HA_RSTATUS); 641 crit_exit(); 642 643 /* 644 * Sanity check 645 */ 646 if (buff[0] != 0x33) { 647 return; 648 } 649 bytes = DPT_HCP_LENGTH(buff); 650 param = DPT_HCP_FIRST(buff); 651 652 if (DPT_HCP_CODE(param) != 1) { 653 /* 654 * DPT Log Page layout error 655 */ 656 kprintf("dpt%d: NOTICE: Log Page (1) layout error\n", 657 dpt->unit); 658 return; 659 } 660 if (!(param[4] & 0x4)) { 661 dpt->cache_type = DPT_NO_CACHE; 662 return; 663 } 664 while (DPT_HCP_CODE(param) != 6) { 665 param = DPT_HCP_NEXT(param); 666 if ((param < buff) 667 || (param >= &buff[bytes])) { 668 return; 669 } 670 } 671 672 if (param[4] & 0x2) { 673 /* 674 * Cache disabled 675 */ 676 dpt->cache_type = DPT_NO_CACHE; 677 return; 678 } 679 680 if (param[4] & 0x4) { 681 dpt->cache_type = DPT_CACHE_WRITETHROUGH; 682 } 683 684 /* XXX This isn't correct. This log parameter only has two bytes.... */ 685 #if 0 686 dpt->cache_size = param[5] 687 | (param[6] << 8) 688 | (param[7] << 16) 689 | (param[8] << 24); 690 #endif 691 } 692 693 static void 694 dpt_poll(struct cam_sim *sim) 695 { 696 dpt_intr(cam_sim_softc(sim)); 697 } 698 699 static void 700 dptexecuteccb(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error) 701 { 702 struct dpt_ccb *dccb; 703 union ccb *ccb; 704 struct dpt_softc *dpt; 705 706 dccb = (struct dpt_ccb *)arg; 707 ccb = dccb->ccb; 708 dpt = (struct dpt_softc *)ccb->ccb_h.ccb_dpt_ptr; 709 710 if (error != 0) { 711 if (error != EFBIG) 712 kprintf("dpt%d: Unexpected error 0x%x returned from " 713 "bus_dmamap_load\n", dpt->unit, error); 714 if (ccb->ccb_h.status == CAM_REQ_INPROG) { 715 xpt_freeze_devq(ccb->ccb_h.path, /*count*/1); 716 ccb->ccb_h.status = CAM_REQ_TOO_BIG|CAM_DEV_QFRZN; 717 } 718 dptfreeccb(dpt, dccb); 719 xpt_done(ccb); 720 return; 721 } 722 723 if (nseg != 0) { 724 dpt_sg_t *sg; 725 bus_dma_segment_t *end_seg; 726 bus_dmasync_op_t op; 727 728 end_seg = dm_segs + nseg; 729 730 /* Copy the segments into our SG list */ 731 sg = dccb->sg_list; 732 while (dm_segs < end_seg) { 733 sg->seg_len = htonl(dm_segs->ds_len); 734 sg->seg_addr = htonl(dm_segs->ds_addr); 735 sg++; 736 dm_segs++; 737 } 738 739 if (nseg > 1) { 740 dccb->eata_ccb.scatter = 1; 741 dccb->eata_ccb.cp_dataDMA = dccb->sg_busaddr; 742 dccb->eata_ccb.cp_datalen = 743 htonl(nseg * sizeof(dpt_sg_t)); 744 } else { 745 dccb->eata_ccb.cp_dataDMA = dccb->sg_list[0].seg_addr; 746 dccb->eata_ccb.cp_datalen = dccb->sg_list[0].seg_len; 747 } 748 749 if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) 750 op = BUS_DMASYNC_PREREAD; 751 else 752 op = BUS_DMASYNC_PREWRITE; 753 754 bus_dmamap_sync(dpt->buffer_dmat, dccb->dmamap, op); 755 756 } else { 757 dccb->eata_ccb.cp_dataDMA = 0; 758 dccb->eata_ccb.cp_datalen = 0; 759 } 760 761 crit_enter(); 762 763 /* 764 * Last time we need to check if this CCB needs to 765 * be aborted. 766 */ 767 if (ccb->ccb_h.status != CAM_REQ_INPROG) { 768 if (nseg != 0) 769 bus_dmamap_unload(dpt->buffer_dmat, dccb->dmamap); 770 dptfreeccb(dpt, dccb); 771 xpt_done(ccb); 772 crit_exit(); 773 return; 774 } 775 776 dccb->state |= DCCB_ACTIVE; 777 ccb->ccb_h.status |= CAM_SIM_QUEUED; 778 LIST_INSERT_HEAD(&dpt->pending_ccb_list, &ccb->ccb_h, sim_links.le); 779 callout_reset(&ccb->ccb_h.timeout_ch, (ccb->ccb_h.timeout * hz) / 1000, 780 dpttimeout, dccb); 781 if (dpt_send_eata_command(dpt, &dccb->eata_ccb, 782 dccb->eata_ccb.cp_busaddr, 783 EATA_CMD_DMA_SEND_CP, 0, 0, 0, 0) != 0) { 784 ccb->ccb_h.status = CAM_NO_HBA; /* HBA dead or just busy?? */ 785 if (nseg != 0) 786 bus_dmamap_unload(dpt->buffer_dmat, dccb->dmamap); 787 dptfreeccb(dpt, dccb); 788 xpt_done(ccb); 789 } 790 791 crit_exit(); 792 } 793 794 static void 795 dpt_action(struct cam_sim *sim, union ccb *ccb) 796 { 797 struct dpt_softc *dpt; 798 799 CAM_DEBUG(ccb->ccb_h.path, CAM_DEBUG_TRACE, ("dpt_action\n")); 800 801 dpt = (struct dpt_softc *)cam_sim_softc(sim); 802 803 if ((dpt->state & DPT_HA_SHUTDOWN_ACTIVE) != 0) { 804 xpt_print_path(ccb->ccb_h.path); 805 kprintf("controller is shutdown. Aborting CCB.\n"); 806 ccb->ccb_h.status = CAM_NO_HBA; 807 xpt_done(ccb); 808 return; 809 } 810 811 switch (ccb->ccb_h.func_code) { 812 /* Common cases first */ 813 case XPT_SCSI_IO: /* Execute the requested I/O operation */ 814 { 815 struct ccb_scsiio *csio; 816 struct ccb_hdr *ccbh; 817 struct dpt_ccb *dccb; 818 struct eata_ccb *eccb; 819 820 csio = &ccb->csio; 821 ccbh = &ccb->ccb_h; 822 /* Max CDB length is 12 bytes */ 823 if (csio->cdb_len > 12) { 824 ccb->ccb_h.status = CAM_REQ_INVALID; 825 xpt_done(ccb); 826 return; 827 } 828 if ((dccb = dptgetccb(dpt)) == NULL) { 829 crit_enter(); 830 dpt->resource_shortage = 1; 831 crit_exit(); 832 xpt_freeze_simq(sim, /*count*/1); 833 ccb->ccb_h.status = CAM_REQUEUE_REQ; 834 xpt_done(ccb); 835 return; 836 } 837 eccb = &dccb->eata_ccb; 838 839 /* Link dccb and ccb so we can find one from the other */ 840 dccb->ccb = ccb; 841 ccb->ccb_h.ccb_dccb_ptr = dccb; 842 ccb->ccb_h.ccb_dpt_ptr = dpt; 843 844 /* 845 * Explicitly set all flags so that the compiler can 846 * be smart about setting them. 847 */ 848 eccb->SCSI_Reset = 0; 849 eccb->HBA_Init = 0; 850 eccb->Auto_Req_Sen = (ccb->ccb_h.flags & CAM_DIS_AUTOSENSE) 851 ? 0 : 1; 852 eccb->scatter = 0; 853 eccb->Quick = 0; 854 eccb->Interpret = 855 ccb->ccb_h.target_id == dpt->hostid[cam_sim_bus(sim)] 856 ? 1 : 0; 857 eccb->DataOut = (ccb->ccb_h.flags & CAM_DIR_OUT) ? 1 : 0; 858 eccb->DataIn = (ccb->ccb_h.flags & CAM_DIR_IN) ? 1 : 0; 859 eccb->reqlen = csio->sense_len; 860 eccb->cp_id = ccb->ccb_h.target_id; 861 eccb->cp_channel = cam_sim_bus(sim); 862 eccb->cp_LUN = ccb->ccb_h.target_lun; 863 eccb->cp_luntar = 0; 864 eccb->cp_dispri = (ccb->ccb_h.flags & CAM_DIS_DISCONNECT) 865 ? 0 : 1; 866 eccb->cp_identify = 1; 867 868 if ((ccb->ccb_h.flags & CAM_TAG_ACTION_VALID) != 0 869 && csio->tag_action != CAM_TAG_ACTION_NONE) { 870 eccb->cp_msg[0] = csio->tag_action; 871 eccb->cp_msg[1] = dccb->tag; 872 } else { 873 eccb->cp_msg[0] = 0; 874 eccb->cp_msg[1] = 0; 875 } 876 eccb->cp_msg[2] = 0; 877 878 if ((ccb->ccb_h.flags & CAM_CDB_POINTER) != 0) { 879 if ((ccb->ccb_h.flags & CAM_CDB_PHYS) == 0) { 880 bcopy(csio->cdb_io.cdb_ptr, 881 eccb->cp_cdb, csio->cdb_len); 882 } else { 883 /* I guess I could map it in... */ 884 ccb->ccb_h.status = CAM_REQ_INVALID; 885 dptfreeccb(dpt, dccb); 886 xpt_done(ccb); 887 return; 888 } 889 } else { 890 bcopy(csio->cdb_io.cdb_bytes, 891 eccb->cp_cdb, csio->cdb_len); 892 } 893 /* 894 * If we have any data to send with this command, 895 * map it into bus space. 896 */ 897 /* Only use S/G if there is a transfer */ 898 if ((ccbh->flags & CAM_DIR_MASK) != CAM_DIR_NONE) { 899 if ((ccbh->flags & CAM_SCATTER_VALID) == 0) { 900 /* 901 * We've been given a pointer 902 * to a single buffer. 903 */ 904 if ((ccbh->flags & CAM_DATA_PHYS) == 0) { 905 int error; 906 907 crit_enter(); 908 error = 909 bus_dmamap_load(dpt->buffer_dmat, 910 dccb->dmamap, 911 csio->data_ptr, 912 csio->dxfer_len, 913 dptexecuteccb, 914 dccb, /*flags*/0); 915 if (error == EINPROGRESS) { 916 /* 917 * So as to maintain ordering, 918 * freeze the controller queue 919 * until our mapping is 920 * returned. 921 */ 922 xpt_freeze_simq(sim, 1); 923 dccb->state |= CAM_RELEASE_SIMQ; 924 } 925 crit_exit(); 926 } else { 927 struct bus_dma_segment seg; 928 929 /* Pointer to physical buffer */ 930 seg.ds_addr = 931 (bus_addr_t)csio->data_ptr; 932 seg.ds_len = csio->dxfer_len; 933 dptexecuteccb(dccb, &seg, 1, 0); 934 } 935 } else { 936 struct bus_dma_segment *segs; 937 938 if ((ccbh->flags & CAM_DATA_PHYS) != 0) 939 panic("dpt_action - Physical " 940 "segment pointers " 941 "unsupported"); 942 943 if ((ccbh->flags&CAM_SG_LIST_PHYS)==0) 944 panic("dpt_action - Virtual " 945 "segment addresses " 946 "unsupported"); 947 948 /* Just use the segments provided */ 949 segs = (struct bus_dma_segment *)csio->data_ptr; 950 dptexecuteccb(dccb, segs, csio->sglist_cnt, 0); 951 } 952 } else { 953 /* 954 * XXX JGibbs. 955 * Does it want them both on or both off? 956 * CAM_DIR_NONE is both on, so this code can 957 * be removed if this is also what the DPT 958 * exptects. 959 */ 960 eccb->DataOut = 0; 961 eccb->DataIn = 0; 962 dptexecuteccb(dccb, NULL, 0, 0); 963 } 964 break; 965 } 966 case XPT_RESET_DEV: /* Bus Device Reset the specified SCSI device */ 967 case XPT_ABORT: /* Abort the specified CCB */ 968 /* XXX Implement */ 969 ccb->ccb_h.status = CAM_REQ_INVALID; 970 xpt_done(ccb); 971 break; 972 case XPT_SET_TRAN_SETTINGS: 973 { 974 ccb->ccb_h.status = CAM_FUNC_NOTAVAIL; 975 xpt_done(ccb); 976 break; 977 } 978 case XPT_GET_TRAN_SETTINGS: 979 /* Get default/user set transfer settings for the target */ 980 { 981 struct ccb_trans_settings *cts = &ccb->cts; 982 struct ccb_trans_settings_scsi *scsi = 983 &cts->proto_specific.scsi; 984 struct ccb_trans_settings_spi *spi = 985 &cts->xport_specific.spi; 986 987 cts->protocol = PROTO_SCSI; 988 cts->protocol_version = SCSI_REV_2; 989 cts->transport = XPORT_SPI; 990 cts->transport_version = 2; 991 992 if (cts->type == CTS_TYPE_USER_SETTINGS) { 993 spi->flags = CTS_SPI_FLAGS_DISC_ENB; 994 spi->bus_width = (dpt->max_id > 7) 995 ? MSG_EXT_WDTR_BUS_8_BIT 996 : MSG_EXT_WDTR_BUS_16_BIT; 997 spi->sync_period = 25; /* 10MHz */ 998 if (spi->sync_period != 0) 999 spi->sync_offset = 15; 1000 scsi->flags = CTS_SCSI_FLAGS_TAG_ENB; 1001 1002 spi->valid = CTS_SPI_VALID_SYNC_RATE 1003 | CTS_SPI_VALID_SYNC_OFFSET 1004 | CTS_SPI_VALID_SYNC_RATE 1005 | CTS_SPI_VALID_BUS_WIDTH 1006 | CTS_SPI_VALID_DISC; 1007 scsi->valid = CTS_SCSI_VALID_TQ; 1008 ccb->ccb_h.status = CAM_REQ_CMP; 1009 } else { 1010 ccb->ccb_h.status = CAM_FUNC_NOTAVAIL; 1011 } 1012 xpt_done(ccb); 1013 break; 1014 } 1015 case XPT_CALC_GEOMETRY: 1016 { 1017 struct ccb_calc_geometry *ccg; 1018 u_int32_t size_mb; 1019 u_int32_t secs_per_cylinder; 1020 int extended; 1021 1022 /* 1023 * XXX Use Adaptec translation until I find out how to 1024 * get this information from the card. 1025 */ 1026 ccg = &ccb->ccg; 1027 size_mb = ccg->volume_size 1028 / ((1024L * 1024L) / ccg->block_size); 1029 extended = 1; 1030 1031 if (size_mb > 1024 && extended) { 1032 ccg->heads = 255; 1033 ccg->secs_per_track = 63; 1034 } else { 1035 ccg->heads = 64; 1036 ccg->secs_per_track = 32; 1037 } 1038 secs_per_cylinder = ccg->heads * ccg->secs_per_track; 1039 ccg->cylinders = ccg->volume_size / secs_per_cylinder; 1040 ccb->ccb_h.status = CAM_REQ_CMP; 1041 xpt_done(ccb); 1042 break; 1043 } 1044 case XPT_RESET_BUS: /* Reset the specified SCSI bus */ 1045 { 1046 /* XXX Implement */ 1047 ccb->ccb_h.status = CAM_REQ_CMP; 1048 xpt_done(ccb); 1049 break; 1050 } 1051 case XPT_TERM_IO: /* Terminate the I/O process */ 1052 /* XXX Implement */ 1053 ccb->ccb_h.status = CAM_REQ_INVALID; 1054 xpt_done(ccb); 1055 break; 1056 case XPT_PATH_INQ: /* Path routing inquiry */ 1057 { 1058 struct ccb_pathinq *cpi = &ccb->cpi; 1059 1060 cpi->version_num = 1; 1061 cpi->hba_inquiry = PI_SDTR_ABLE|PI_TAG_ABLE; 1062 if (dpt->max_id > 7) 1063 cpi->hba_inquiry |= PI_WIDE_16; 1064 cpi->target_sprt = 0; 1065 cpi->hba_misc = 0; 1066 cpi->hba_eng_cnt = 0; 1067 cpi->max_target = dpt->max_id; 1068 cpi->max_lun = dpt->max_lun; 1069 cpi->initiator_id = dpt->hostid[cam_sim_bus(sim)]; 1070 cpi->bus_id = cam_sim_bus(sim); 1071 cpi->base_transfer_speed = 3300; 1072 strncpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN); 1073 strncpy(cpi->hba_vid, "DPT", HBA_IDLEN); 1074 strncpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN); 1075 cpi->unit_number = cam_sim_unit(sim); 1076 cpi->transport = XPORT_SPI; 1077 cpi->transport_version = 2; 1078 cpi->protocol = PROTO_SCSI; 1079 cpi->protocol_version = SCSI_REV_2; 1080 cpi->ccb_h.status = CAM_REQ_CMP; 1081 xpt_done(ccb); 1082 break; 1083 } 1084 default: 1085 ccb->ccb_h.status = CAM_REQ_INVALID; 1086 xpt_done(ccb); 1087 break; 1088 } 1089 } 1090 1091 /* 1092 * This routine will try to send an EATA command to the DPT HBA. 1093 * It will, by default, try 20,000 times, waiting 50us between tries. 1094 * It returns 0 on success and 1 on failure. 1095 * It is assumed to be called at splcam(). 1096 */ 1097 static int 1098 dpt_send_eata_command(dpt_softc_t *dpt, eata_ccb_t *cmd_block, 1099 u_int32_t cmd_busaddr, u_int command, u_int retries, 1100 u_int ifc, u_int code, u_int code2) 1101 { 1102 u_int loop; 1103 1104 if (!retries) 1105 retries = 20000; 1106 1107 /* 1108 * I hate this polling nonsense. Wish there was a way to tell the DPT 1109 * to go get commands at its own pace, or to interrupt when ready. 1110 * In the mean time we will measure how many itterations it really 1111 * takes. 1112 */ 1113 for (loop = 0; loop < retries; loop++) { 1114 if ((dpt_inb(dpt, HA_RAUXSTAT) & HA_ABUSY) == 0) 1115 break; 1116 else 1117 DELAY(50); 1118 } 1119 1120 if (loop < retries) { 1121 #ifdef DPT_MEASURE_PERFORMANCE 1122 if (loop > dpt->performance.max_eata_tries) 1123 dpt->performance.max_eata_tries = loop; 1124 1125 if (loop < dpt->performance.min_eata_tries) 1126 dpt->performance.min_eata_tries = loop; 1127 #endif 1128 } else { 1129 #ifdef DPT_MEASURE_PERFORMANCE 1130 ++dpt->performance.command_too_busy; 1131 #endif 1132 return (1); 1133 } 1134 1135 /* The controller is alive, advance the wedge timer */ 1136 #ifdef DPT_RESET_HBA 1137 dpt->last_contact = microtime_now; 1138 #endif 1139 1140 if (cmd_block == NULL) 1141 cmd_busaddr = 0; 1142 #if (BYTE_ORDER == BIG_ENDIAN) 1143 else { 1144 cmd_busaddr = ((cmd_busaddr >> 24) & 0xFF) 1145 | ((cmd_busaddr >> 16) & 0xFF) 1146 | ((cmd_busaddr >> 8) & 0xFF) 1147 | (cmd_busaddr & 0xFF); 1148 } 1149 #endif 1150 /* And now the address */ 1151 dpt_outl(dpt, HA_WDMAADDR, cmd_busaddr); 1152 1153 if (command == EATA_CMD_IMMEDIATE) { 1154 if (cmd_block == NULL) { 1155 dpt_outb(dpt, HA_WCODE2, code2); 1156 dpt_outb(dpt, HA_WCODE, code); 1157 } 1158 dpt_outb(dpt, HA_WIFC, ifc); 1159 } 1160 dpt_outb(dpt, HA_WCOMMAND, command); 1161 1162 return (0); 1163 } 1164 1165 1166 /* ==================== Exported Function definitions =======================*/ 1167 dpt_softc_t * 1168 dpt_alloc(device_t dev, bus_space_tag_t tag, bus_space_handle_t bsh) 1169 { 1170 dpt_softc_t *dpt = device_get_softc(dev); 1171 int i; 1172 1173 bzero(dpt, sizeof(dpt_softc_t)); 1174 dpt->tag = tag; 1175 dpt->bsh = bsh; 1176 dpt->unit = device_get_unit(dev); 1177 SLIST_INIT(&dpt->free_dccb_list); 1178 LIST_INIT(&dpt->pending_ccb_list); 1179 TAILQ_INSERT_TAIL(&dpt_softcs, dpt, links); 1180 for (i = 0; i < MAX_CHANNELS; i++) 1181 dpt->resetlevel[i] = DPT_HA_OK; 1182 1183 #ifdef DPT_MEASURE_PERFORMANCE 1184 dpt_reset_performance(dpt); 1185 #endif /* DPT_MEASURE_PERFORMANCE */ 1186 return (dpt); 1187 } 1188 1189 void 1190 dpt_free(struct dpt_softc *dpt) 1191 { 1192 switch (dpt->init_level) { 1193 default: 1194 case 5: 1195 bus_dmamap_unload(dpt->dccb_dmat, dpt->dccb_dmamap); 1196 case 4: 1197 bus_dmamem_free(dpt->dccb_dmat, dpt->dpt_dccbs, 1198 dpt->dccb_dmamap); 1199 bus_dmamap_destroy(dpt->dccb_dmat, dpt->dccb_dmamap); 1200 case 3: 1201 bus_dma_tag_destroy(dpt->dccb_dmat); 1202 case 2: 1203 bus_dma_tag_destroy(dpt->buffer_dmat); 1204 case 1: 1205 { 1206 struct sg_map_node *sg_map; 1207 1208 while ((sg_map = SLIST_FIRST(&dpt->sg_maps)) != NULL) { 1209 SLIST_REMOVE_HEAD(&dpt->sg_maps, links); 1210 bus_dmamap_unload(dpt->sg_dmat, 1211 sg_map->sg_dmamap); 1212 bus_dmamem_free(dpt->sg_dmat, sg_map->sg_vaddr, 1213 sg_map->sg_dmamap); 1214 kfree(sg_map, M_DEVBUF); 1215 } 1216 bus_dma_tag_destroy(dpt->sg_dmat); 1217 } 1218 case 0: 1219 break; 1220 } 1221 TAILQ_REMOVE(&dpt_softcs, dpt, links); 1222 } 1223 1224 static u_int8_t string_sizes[] = 1225 { 1226 sizeof(((dpt_inq_t*)NULL)->vendor), 1227 sizeof(((dpt_inq_t*)NULL)->modelNum), 1228 sizeof(((dpt_inq_t*)NULL)->firmware), 1229 sizeof(((dpt_inq_t*)NULL)->protocol), 1230 }; 1231 1232 int 1233 dpt_init(struct dpt_softc *dpt) 1234 { 1235 dpt_conf_t conf; 1236 struct sg_map_node *sg_map; 1237 dpt_ccb_t *dccb; 1238 u_int8_t *strp; 1239 int index; 1240 int i; 1241 int retval; 1242 1243 dpt->init_level = 0; 1244 SLIST_INIT(&dpt->sg_maps); 1245 1246 #ifdef DPT_RESET_BOARD 1247 kprintf("dpt%d: resetting HBA\n", dpt->unit); 1248 dpt_outb(dpt, HA_WCOMMAND, EATA_CMD_RESET); 1249 DELAY(750000); 1250 /* XXX Shouldn't we poll a status register or something??? */ 1251 #endif 1252 /* DMA tag for our S/G structures. We allocate in page sized chunks */ 1253 if (bus_dma_tag_create(dpt->parent_dmat, /*alignment*/1, /*boundary*/0, 1254 /*lowaddr*/BUS_SPACE_MAXADDR, 1255 /*highaddr*/BUS_SPACE_MAXADDR, 1256 /*filter*/NULL, /*filterarg*/NULL, 1257 PAGE_SIZE, /*nsegments*/1, 1258 /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT, 1259 /*flags*/0, &dpt->sg_dmat) != 0) { 1260 goto error_exit; 1261 } 1262 1263 dpt->init_level++; 1264 1265 /* 1266 * We allocate our DPT ccbs as a contiguous array of bus dma'able 1267 * memory. To get the allocation size, we need to know how many 1268 * ccbs the card supports. This requires a ccb. We solve this 1269 * chicken and egg problem by allocating some re-usable S/G space 1270 * up front, and treating it as our status packet, CCB, and target 1271 * memory space for these commands. 1272 */ 1273 sg_map = dptallocsgmap(dpt); 1274 if (sg_map == NULL) 1275 goto error_exit; 1276 1277 dpt->sp = (volatile dpt_sp_t *)sg_map->sg_vaddr; 1278 dccb = (struct dpt_ccb *)(uintptr_t)(volatile void *)&dpt->sp[1]; 1279 bzero(dccb, sizeof(*dccb)); 1280 dpt->sp_physaddr = sg_map->sg_physaddr; 1281 dccb->eata_ccb.cp_dataDMA = 1282 htonl(sg_map->sg_physaddr + sizeof(dpt_sp_t) + sizeof(*dccb)); 1283 dccb->eata_ccb.cp_busaddr = ~0; 1284 dccb->eata_ccb.cp_statDMA = htonl(dpt->sp_physaddr); 1285 dccb->eata_ccb.cp_reqDMA = htonl(dpt->sp_physaddr + sizeof(*dccb) 1286 + offsetof(struct dpt_ccb, sense_data)); 1287 1288 /* Okay. Fetch our config */ 1289 bzero(&dccb[1], sizeof(conf)); /* data area */ 1290 retval = dpt_get_conf(dpt, dccb, sg_map->sg_physaddr + sizeof(dpt_sp_t), 1291 sizeof(conf), 0xc1, 7, 1); 1292 1293 if (retval != 0) { 1294 kprintf("dpt%d: Failed to get board configuration\n", dpt->unit); 1295 return (retval); 1296 } 1297 bcopy(&dccb[1], &conf, sizeof(conf)); 1298 1299 bzero(&dccb[1], sizeof(dpt->board_data)); 1300 retval = dpt_get_conf(dpt, dccb, sg_map->sg_physaddr + sizeof(dpt_sp_t), 1301 sizeof(dpt->board_data), 0, conf.scsi_id0, 0); 1302 if (retval != 0) { 1303 kprintf("dpt%d: Failed to get inquiry information\n", dpt->unit); 1304 return (retval); 1305 } 1306 bcopy(&dccb[1], &dpt->board_data, sizeof(dpt->board_data)); 1307 1308 dpt_detect_cache(dpt, dccb, sg_map->sg_physaddr + sizeof(dpt_sp_t), 1309 (u_int8_t *)&dccb[1]); 1310 1311 switch (ntohl(conf.splen)) { 1312 case DPT_EATA_REVA: 1313 dpt->EATA_revision = 'a'; 1314 break; 1315 case DPT_EATA_REVB: 1316 dpt->EATA_revision = 'b'; 1317 break; 1318 case DPT_EATA_REVC: 1319 dpt->EATA_revision = 'c'; 1320 break; 1321 case DPT_EATA_REVZ: 1322 dpt->EATA_revision = 'z'; 1323 break; 1324 default: 1325 dpt->EATA_revision = '?'; 1326 } 1327 1328 dpt->max_id = conf.MAX_ID; 1329 dpt->max_lun = conf.MAX_LUN; 1330 dpt->irq = conf.IRQ; 1331 dpt->dma_channel = (8 - conf.DMA_channel) & 7; 1332 dpt->channels = conf.MAX_CHAN + 1; 1333 dpt->state |= DPT_HA_OK; 1334 if (conf.SECOND) 1335 dpt->primary = FALSE; 1336 else 1337 dpt->primary = TRUE; 1338 1339 dpt->more_support = conf.MORE_support; 1340 1341 if (strncmp(dpt->board_data.firmware, "07G0", 4) >= 0) 1342 dpt->immediate_support = 1; 1343 else 1344 dpt->immediate_support = 0; 1345 1346 dpt->broken_INQUIRY = FALSE; 1347 1348 dpt->cplen = ntohl(conf.cplen); 1349 dpt->cppadlen = ntohs(conf.cppadlen); 1350 dpt->max_dccbs = ntohs(conf.queuesiz); 1351 1352 if (dpt->max_dccbs > 256) { 1353 kprintf("dpt%d: Max CCBs reduced from %d to " 1354 "256 due to tag algorithm\n", dpt->unit, dpt->max_dccbs); 1355 dpt->max_dccbs = 256; 1356 } 1357 1358 dpt->hostid[0] = conf.scsi_id0; 1359 dpt->hostid[1] = conf.scsi_id1; 1360 dpt->hostid[2] = conf.scsi_id2; 1361 1362 if (conf.SG_64K) 1363 dpt->sgsize = 8192; 1364 else 1365 dpt->sgsize = ntohs(conf.SGsiz); 1366 1367 /* We can only get 64k buffers, so don't bother to waste space. */ 1368 if (dpt->sgsize < 17 || dpt->sgsize > 32) 1369 dpt->sgsize = 32; 1370 1371 if (dpt->sgsize > dpt_max_segs) 1372 dpt->sgsize = dpt_max_segs; 1373 1374 /* DMA tag for mapping buffers into device visible space. */ 1375 if (bus_dma_tag_create(dpt->parent_dmat, /*alignment*/1, /*boundary*/0, 1376 /*lowaddr*/BUS_SPACE_MAXADDR, 1377 /*highaddr*/BUS_SPACE_MAXADDR, 1378 /*filter*/NULL, /*filterarg*/NULL, 1379 /*maxsize*/MAXBSIZE, /*nsegments*/dpt->sgsize, 1380 /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT, 1381 /*flags*/BUS_DMA_ALLOCNOW, 1382 &dpt->buffer_dmat) != 0) { 1383 kprintf("dpt: bus_dma_tag_create(...,dpt->buffer_dmat) failed\n"); 1384 goto error_exit; 1385 } 1386 1387 dpt->init_level++; 1388 1389 /* DMA tag for our ccb structures and interrupt status packet */ 1390 if (bus_dma_tag_create(dpt->parent_dmat, /*alignment*/1, /*boundary*/0, 1391 /*lowaddr*/BUS_SPACE_MAXADDR, 1392 /*highaddr*/BUS_SPACE_MAXADDR, 1393 /*filter*/NULL, /*filterarg*/NULL, 1394 (dpt->max_dccbs * sizeof(struct dpt_ccb)) 1395 + sizeof(dpt_sp_t), 1396 /*nsegments*/1, 1397 /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT, 1398 /*flags*/0, &dpt->dccb_dmat) != 0) { 1399 kprintf("dpt: bus_dma_tag_create(...,dpt->dccb_dmat) failed\n"); 1400 goto error_exit; 1401 } 1402 1403 dpt->init_level++; 1404 1405 /* Allocation for our ccbs and interrupt status packet */ 1406 if (bus_dmamem_alloc(dpt->dccb_dmat, (void **)&dpt->dpt_dccbs, 1407 BUS_DMA_NOWAIT, &dpt->dccb_dmamap) != 0) { 1408 kprintf("dpt: bus_dmamem_alloc(dpt->dccb_dmat,...) failed\n"); 1409 goto error_exit; 1410 } 1411 1412 dpt->init_level++; 1413 1414 /* And permanently map them */ 1415 bus_dmamap_load(dpt->dccb_dmat, dpt->dccb_dmamap, 1416 dpt->dpt_dccbs, 1417 (dpt->max_dccbs * sizeof(struct dpt_ccb)) 1418 + sizeof(dpt_sp_t), 1419 dptmapmem, &dpt->dpt_ccb_busbase, /*flags*/0); 1420 1421 /* Clear them out. */ 1422 bzero(dpt->dpt_dccbs, 1423 (dpt->max_dccbs * sizeof(struct dpt_ccb)) + sizeof(dpt_sp_t)); 1424 1425 dpt->dpt_ccb_busend = dpt->dpt_ccb_busbase; 1426 1427 dpt->sp = (dpt_sp_t*)&dpt->dpt_dccbs[dpt->max_dccbs]; 1428 dpt->sp_physaddr = dpt->dpt_ccb_busbase 1429 + (dpt->max_dccbs * sizeof(dpt_ccb_t)); 1430 dpt->init_level++; 1431 1432 /* Allocate our first batch of ccbs */ 1433 if (dptallocccbs(dpt) == 0) { 1434 kprintf("dpt: dptallocccbs(dpt) == 0\n"); 1435 return (2); 1436 } 1437 1438 /* Prepare for Target Mode */ 1439 dpt->target_mode_enabled = 1; 1440 1441 /* Nuke excess spaces from inquiry information */ 1442 strp = dpt->board_data.vendor; 1443 for (i = 0; i < sizeof(string_sizes); i++) { 1444 index = string_sizes[i] - 1; 1445 while (index && (strp[index] == ' ')) 1446 strp[index--] = '\0'; 1447 strp += string_sizes[i]; 1448 } 1449 1450 kprintf("dpt%d: %.8s %.16s FW Rev. %.4s, ", 1451 dpt->unit, dpt->board_data.vendor, 1452 dpt->board_data.modelNum, dpt->board_data.firmware); 1453 1454 kprintf("%d channel%s, ", dpt->channels, dpt->channels > 1 ? "s" : ""); 1455 1456 if (dpt->cache_type != DPT_NO_CACHE 1457 && dpt->cache_size != 0) { 1458 kprintf("%s Cache, ", 1459 dpt->cache_type == DPT_CACHE_WRITETHROUGH 1460 ? "Write-Through" : "Write-Back"); 1461 } 1462 1463 kprintf("%d CCBs\n", dpt->max_dccbs); 1464 return (0); 1465 1466 error_exit: 1467 return (1); 1468 } 1469 1470 int 1471 dpt_attach(dpt_softc_t *dpt) 1472 { 1473 struct cam_devq *devq; 1474 int i; 1475 1476 /* 1477 * Create the device queue for our SIM. 1478 */ 1479 devq = cam_simq_alloc(dpt->max_dccbs); 1480 if (devq == NULL) 1481 return (0); 1482 1483 for (i = 0; i < dpt->channels; i++) { 1484 /* 1485 * Construct our SIM entry 1486 */ 1487 dpt->sims[i] = cam_sim_alloc(dpt_action, dpt_poll, "dpt", 1488 dpt, dpt->unit, /*untagged*/2, 1489 /*tagged*/dpt->max_dccbs, devq); 1490 if (xpt_bus_register(dpt->sims[i], i) != CAM_SUCCESS) { 1491 cam_sim_free(dpt->sims[i]); 1492 break; 1493 } 1494 1495 if (xpt_create_path(&dpt->paths[i], /*periph*/NULL, 1496 cam_sim_path(dpt->sims[i]), 1497 CAM_TARGET_WILDCARD, 1498 CAM_LUN_WILDCARD) != CAM_REQ_CMP) { 1499 xpt_bus_deregister(cam_sim_path(dpt->sims[i])); 1500 cam_sim_free(dpt->sims[i]); 1501 break; 1502 } 1503 1504 } 1505 cam_simq_release(devq); 1506 if (i > 0) 1507 EVENTHANDLER_REGISTER(shutdown_post_sync, dptshutdown, 1508 dpt, SHUTDOWN_PRI_DRIVER); 1509 return (i); 1510 } 1511 1512 1513 /* 1514 * This is the interrupt handler for the DPT driver. 1515 */ 1516 void 1517 dpt_intr(void *arg) 1518 { 1519 dpt_softc_t *dpt; 1520 dpt_ccb_t *dccb; 1521 union ccb *ccb; 1522 u_int status; 1523 u_int aux_status; 1524 u_int hba_stat; 1525 u_int scsi_stat; 1526 u_int32_t residue_len; /* Number of bytes not transferred */ 1527 1528 dpt = (dpt_softc_t *)arg; 1529 1530 /* First order of business is to check if this interrupt is for us */ 1531 while (((aux_status = dpt_inb(dpt, HA_RAUXSTAT)) & HA_AIRQ) != 0) { 1532 1533 /* 1534 * What we want to do now, is to capture the status, all of it, 1535 * move it where it belongs, wake up whoever sleeps waiting to 1536 * process this result, and get out of here. 1537 */ 1538 if (dpt->sp->ccb_busaddr < dpt->dpt_ccb_busbase 1539 || dpt->sp->ccb_busaddr >= dpt->dpt_ccb_busend) { 1540 kprintf("Encountered bogus status packet\n"); 1541 status = dpt_inb(dpt, HA_RSTATUS); 1542 return; 1543 } 1544 1545 dccb = dptccbptov(dpt, dpt->sp->ccb_busaddr); 1546 1547 dpt->sp->ccb_busaddr = ~0; 1548 1549 /* Ignore status packets with EOC not set */ 1550 if (dpt->sp->EOC == 0) { 1551 kprintf("dpt%d ERROR: Request %d received with " 1552 "clear EOC.\n Marking as LOST.\n", 1553 dpt->unit, dccb->transaction_id); 1554 1555 #ifdef DPT_HANDLE_TIMEOUTS 1556 dccb->state |= DPT_CCB_STATE_MARKED_LOST; 1557 #endif 1558 /* This CLEARS the interrupt! */ 1559 status = dpt_inb(dpt, HA_RSTATUS); 1560 continue; 1561 } 1562 dpt->sp->EOC = 0; 1563 1564 /* 1565 * Double buffer the status information so the hardware can 1566 * work on updating the status packet while we decifer the 1567 * one we were just interrupted for. 1568 * According to Mark Salyzyn, we only need few pieces of it. 1569 */ 1570 hba_stat = dpt->sp->hba_stat; 1571 scsi_stat = dpt->sp->scsi_stat; 1572 residue_len = dpt->sp->residue_len; 1573 1574 /* Clear interrupts, check for error */ 1575 if ((status = dpt_inb(dpt, HA_RSTATUS)) & HA_SERROR) { 1576 /* 1577 * Error Condition. Check for magic cookie. Exit 1578 * this test on earliest sign of non-reset condition 1579 */ 1580 1581 /* Check that this is not a board reset interrupt */ 1582 if (dpt_just_reset(dpt)) { 1583 kprintf("dpt%d: HBA rebooted.\n" 1584 " All transactions should be " 1585 "resubmitted\n", 1586 dpt->unit); 1587 1588 kprintf("dpt%d: >>---->> This is incomplete, " 1589 "fix me.... <<----<<", dpt->unit); 1590 panic("DPT Rebooted"); 1591 1592 } 1593 } 1594 /* Process CCB */ 1595 ccb = dccb->ccb; 1596 callout_stop(&ccb->ccb_h.timeout_ch); 1597 if ((ccb->ccb_h.flags & CAM_DIR_MASK) != CAM_DIR_NONE) { 1598 bus_dmasync_op_t op; 1599 1600 if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) 1601 op = BUS_DMASYNC_POSTREAD; 1602 else 1603 op = BUS_DMASYNC_POSTWRITE; 1604 bus_dmamap_sync(dpt->buffer_dmat, dccb->dmamap, op); 1605 bus_dmamap_unload(dpt->buffer_dmat, dccb->dmamap); 1606 } 1607 1608 /* Common Case inline... */ 1609 if (hba_stat == HA_NO_ERROR) { 1610 ccb->csio.scsi_status = scsi_stat; 1611 ccb->ccb_h.status = 0; 1612 switch (scsi_stat) { 1613 case SCSI_STATUS_OK: 1614 ccb->ccb_h.status |= CAM_REQ_CMP; 1615 break; 1616 case SCSI_STATUS_CHECK_COND: 1617 case SCSI_STATUS_CMD_TERMINATED: 1618 bcopy(&dccb->sense_data, &ccb->csio.sense_data, 1619 ccb->csio.sense_len); 1620 ccb->ccb_h.status |= CAM_AUTOSNS_VALID; 1621 /* FALLTHROUGH */ 1622 default: 1623 ccb->ccb_h.status |= CAM_SCSI_STATUS_ERROR; 1624 /* XXX Freeze DevQ */ 1625 break; 1626 } 1627 ccb->csio.resid = residue_len; 1628 dptfreeccb(dpt, dccb); 1629 xpt_done(ccb); 1630 } else { 1631 dptprocesserror(dpt, dccb, ccb, hba_stat, scsi_stat, 1632 residue_len); 1633 } 1634 } 1635 } 1636 1637 static void 1638 dptprocesserror(dpt_softc_t *dpt, dpt_ccb_t *dccb, union ccb *ccb, 1639 u_int hba_stat, u_int scsi_stat, u_int32_t resid) 1640 { 1641 ccb->csio.resid = resid; 1642 switch (hba_stat) { 1643 case HA_ERR_SEL_TO: 1644 ccb->ccb_h.status = CAM_SEL_TIMEOUT; 1645 break; 1646 case HA_ERR_CMD_TO: 1647 ccb->ccb_h.status = CAM_CMD_TIMEOUT; 1648 break; 1649 case HA_SCSIBUS_RESET: 1650 case HA_HBA_POWER_UP: /* Similar effect to a bus reset??? */ 1651 ccb->ccb_h.status = CAM_SCSI_BUS_RESET; 1652 break; 1653 case HA_CP_ABORTED: 1654 case HA_CP_RESET: /* XXX ??? */ 1655 case HA_CP_ABORT_NA: /* XXX ??? */ 1656 case HA_CP_RESET_NA: /* XXX ??? */ 1657 if ((ccb->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_INPROG) 1658 ccb->ccb_h.status = CAM_REQ_ABORTED; 1659 break; 1660 case HA_PCI_PARITY: 1661 case HA_PCI_MABORT: 1662 case HA_PCI_TABORT: 1663 case HA_PCI_STABORT: 1664 case HA_BUS_PARITY: 1665 case HA_PARITY_ERR: 1666 case HA_ECC_ERR: 1667 ccb->ccb_h.status = CAM_UNCOR_PARITY; 1668 break; 1669 case HA_UNX_MSGRJCT: 1670 ccb->ccb_h.status = CAM_MSG_REJECT_REC; 1671 break; 1672 case HA_UNX_BUSPHASE: 1673 ccb->ccb_h.status = CAM_SEQUENCE_FAIL; 1674 break; 1675 case HA_UNX_BUS_FREE: 1676 ccb->ccb_h.status = CAM_UNEXP_BUSFREE; 1677 break; 1678 case HA_SCSI_HUNG: 1679 case HA_RESET_STUCK: 1680 /* 1681 * Dead??? Can the controller get unstuck 1682 * from these conditions 1683 */ 1684 ccb->ccb_h.status = CAM_NO_HBA; 1685 break; 1686 case HA_RSENSE_FAIL: 1687 ccb->ccb_h.status = CAM_AUTOSENSE_FAIL; 1688 break; 1689 default: 1690 kprintf("dpt%d: Undocumented Error %x\n", dpt->unit, hba_stat); 1691 kprintf("Please mail this message to shimon@simon-shapiro.org\n"); 1692 ccb->ccb_h.status = CAM_REQ_CMP_ERR; 1693 break; 1694 } 1695 dptfreeccb(dpt, dccb); 1696 xpt_done(ccb); 1697 } 1698 1699 static void 1700 dpttimeout(void *arg) 1701 { 1702 struct dpt_ccb *dccb; 1703 union ccb *ccb; 1704 struct dpt_softc *dpt; 1705 1706 dccb = (struct dpt_ccb *)arg; 1707 ccb = dccb->ccb; 1708 dpt = (struct dpt_softc *)ccb->ccb_h.ccb_dpt_ptr; 1709 xpt_print_path(ccb->ccb_h.path); 1710 kprintf("CCB %p - timed out\n", (void *)dccb); 1711 1712 crit_enter(); 1713 1714 /* 1715 * Try to clear any pending jobs. FreeBSD will loose interrupts, 1716 * leaving the controller suspended, and commands timed-out. 1717 * By calling the interrupt handler, any command thus stuck will be 1718 * completed. 1719 */ 1720 dpt_intr(dpt); 1721 1722 if ((dccb->state & DCCB_ACTIVE) == 0) { 1723 xpt_print_path(ccb->ccb_h.path); 1724 kprintf("CCB %p - timed out CCB already completed\n", 1725 (void *)dccb); 1726 crit_exit(); 1727 return; 1728 } 1729 1730 /* Abort this particular command. Leave all others running */ 1731 dpt_send_immediate(dpt, &dccb->eata_ccb, dccb->eata_ccb.cp_busaddr, 1732 /*retries*/20000, EATA_SPECIFIC_ABORT, 0, 0); 1733 ccb->ccb_h.status = CAM_CMD_TIMEOUT; 1734 crit_exit(); 1735 } 1736 1737 /* 1738 * Shutdown the controller and ensure that the cache is completely flushed. 1739 * Called from the shutdown_final event after all disk access has completed. 1740 */ 1741 static void 1742 dptshutdown(void *arg, int howto) 1743 { 1744 dpt_softc_t *dpt; 1745 1746 dpt = (dpt_softc_t *)arg; 1747 1748 kprintf("dpt%d: Shutting down (mode %x) HBA. Please wait...\n", 1749 dpt->unit, howto); 1750 1751 /* 1752 * What we do for a shutdown, is give the DPT early power loss warning 1753 */ 1754 dpt_send_immediate(dpt, NULL, 0, EATA_POWER_OFF_WARN, 0, 0, 0); 1755 DELAY(1000 * 1000 * 5); 1756 kprintf("dpt%d: Controller was warned of shutdown and is now " 1757 "disabled\n", dpt->unit); 1758 } 1759 1760 /*============================================================================*/ 1761 1762 #if 0 1763 #ifdef DPT_RESET_HBA 1764 1765 /* 1766 ** Function name : dpt_reset_hba 1767 ** 1768 ** Description : Reset the HBA and properly discard all pending work 1769 ** Input : Softc 1770 ** Output : Nothing 1771 */ 1772 static void 1773 dpt_reset_hba(dpt_softc_t *dpt) 1774 { 1775 eata_ccb_t *ccb; 1776 dpt_ccb_t dccb, *dccbp; 1777 int result; 1778 struct scsi_xfer *xs; 1779 1780 /* Prepare a control block. The SCSI command part is immaterial */ 1781 dccb.xs = NULL; 1782 dccb.flags = 0; 1783 dccb.state = DPT_CCB_STATE_NEW; 1784 dccb.std_callback = NULL; 1785 dccb.wrbuff_callback = NULL; 1786 1787 ccb = &dccb.eata_ccb; 1788 ccb->CP_OpCode = EATA_CMD_RESET; 1789 ccb->SCSI_Reset = 0; 1790 ccb->HBA_Init = 1; 1791 ccb->Auto_Req_Sen = 1; 1792 ccb->cp_id = 0; /* Should be ignored */ 1793 ccb->DataIn = 1; 1794 ccb->DataOut = 0; 1795 ccb->Interpret = 1; 1796 ccb->reqlen = htonl(sizeof(struct scsi_sense_data)); 1797 ccb->cp_statDMA = htonl(vtophys(&ccb->cp_statDMA)); 1798 ccb->cp_reqDMA = htonl(vtophys(&ccb->cp_reqDMA)); 1799 ccb->cp_viraddr = (u_int32_t) & ccb; 1800 1801 ccb->cp_msg[0] = HA_IDENTIFY_MSG | HA_DISCO_RECO; 1802 ccb->cp_scsi_cmd = 0; /* Should be ignored */ 1803 1804 /* Lock up the submitted queue. We are very persistant here */ 1805 crit_enter(); 1806 while (dpt->queue_status & DPT_SUBMITTED_QUEUE_ACTIVE) { 1807 DELAY(100); 1808 } 1809 1810 dpt->queue_status |= DPT_SUBMITTED_QUEUE_ACTIVE; 1811 crit_exit(); 1812 1813 /* Send the RESET message */ 1814 if ((result = dpt_send_eata_command(dpt, &dccb.eata_ccb, 1815 EATA_CMD_RESET, 0, 0, 0, 0)) != 0) { 1816 kprintf("dpt%d: Failed to send the RESET message.\n" 1817 " Trying cold boot (ouch!)\n", dpt->unit); 1818 1819 1820 if ((result = dpt_send_eata_command(dpt, &dccb.eata_ccb, 1821 EATA_COLD_BOOT, 0, 0, 1822 0, 0)) != 0) { 1823 panic("dpt%d: Faild to cold boot the HBA\n", 1824 dpt->unit); 1825 } 1826 #ifdef DPT_MEASURE_PERFORMANCE 1827 dpt->performance.cold_boots++; 1828 #endif /* DPT_MEASURE_PERFORMANCE */ 1829 } 1830 1831 #ifdef DPT_MEASURE_PERFORMANCE 1832 dpt->performance.warm_starts++; 1833 #endif /* DPT_MEASURE_PERFORMANCE */ 1834 1835 kprintf("dpt%d: Aborting pending requests. O/S should re-submit\n", 1836 dpt->unit); 1837 1838 while ((dccbp = TAILQ_FIRST(&dpt->completed_ccbs)) != NULL) { 1839 struct scsi_xfer *xs = dccbp->xs; 1840 1841 /* Not all transactions have xs structs */ 1842 if (xs != NULL) { 1843 /* Tell the kernel proper this did not complete well */ 1844 xs->error |= XS_SELTIMEOUT; 1845 xs->flags |= SCSI_ITSDONE; 1846 scsi_done(xs); 1847 } 1848 1849 dpt_Qremove_submitted(dpt, dccbp); 1850 1851 /* Remember, Callbacks are NOT in the standard queue */ 1852 if (dccbp->std_callback != NULL) { 1853 (dccbp->std_callback)(dpt, dccbp->eata_ccb.cp_channel, 1854 dccbp); 1855 } else { 1856 crit_enter(); 1857 dpt_Qpush_free(dpt, dccbp); 1858 crit_exit(); 1859 } 1860 } 1861 1862 kprintf("dpt%d: reset done aborting all pending commands\n", dpt->unit); 1863 dpt->queue_status &= ~DPT_SUBMITTED_QUEUE_ACTIVE; 1864 } 1865 1866 #endif /* DPT_RESET_HBA */ 1867 1868 /* 1869 * Build a Command Block for target mode READ/WRITE BUFFER, 1870 * with the ``sync'' bit ON. 1871 * 1872 * Although the length and offset are 24 bit fields in the command, they cannot 1873 * exceed 8192 bytes, so we take them as short integers andcheck their range. 1874 * If they are sensless, we round them to zero offset, maximum length and 1875 * complain. 1876 */ 1877 1878 static void 1879 dpt_target_ccb(dpt_softc_t * dpt, int bus, u_int8_t target, u_int8_t lun, 1880 dpt_ccb_t * ccb, int mode, u_int8_t command, 1881 u_int16_t length, u_int16_t offset) 1882 { 1883 eata_ccb_t *cp; 1884 1885 if ((length + offset) > DPT_MAX_TARGET_MODE_BUFFER_SIZE) { 1886 kprintf("dpt%d: Length of %d, and offset of %d are wrong\n", 1887 dpt->unit, length, offset); 1888 length = DPT_MAX_TARGET_MODE_BUFFER_SIZE; 1889 offset = 0; 1890 } 1891 ccb->xs = NULL; 1892 ccb->flags = 0; 1893 ccb->state = DPT_CCB_STATE_NEW; 1894 ccb->std_callback = (ccb_callback) dpt_target_done; 1895 ccb->wrbuff_callback = NULL; 1896 1897 cp = &ccb->eata_ccb; 1898 cp->CP_OpCode = EATA_CMD_DMA_SEND_CP; 1899 cp->SCSI_Reset = 0; 1900 cp->HBA_Init = 0; 1901 cp->Auto_Req_Sen = 1; 1902 cp->cp_id = target; 1903 cp->DataIn = 1; 1904 cp->DataOut = 0; 1905 cp->Interpret = 0; 1906 cp->reqlen = htonl(sizeof(struct scsi_sense_data)); 1907 cp->cp_statDMA = htonl(vtophys(&cp->cp_statDMA)); 1908 cp->cp_reqDMA = htonl(vtophys(&cp->cp_reqDMA)); 1909 cp->cp_viraddr = (u_int32_t) & ccb; 1910 1911 cp->cp_msg[0] = HA_IDENTIFY_MSG | HA_DISCO_RECO; 1912 1913 cp->cp_scsi_cmd = command; 1914 cp->cp_cdb[1] = (u_int8_t) (mode & SCSI_TM_MODE_MASK); 1915 cp->cp_lun = lun; /* Order is important here! */ 1916 cp->cp_cdb[2] = 0x00; /* Buffer Id, only 1 :-( */ 1917 cp->cp_cdb[3] = (length >> 16) & 0xFF; /* Buffer offset MSB */ 1918 cp->cp_cdb[4] = (length >> 8) & 0xFF; 1919 cp->cp_cdb[5] = length & 0xFF; 1920 cp->cp_cdb[6] = (length >> 16) & 0xFF; /* Length MSB */ 1921 cp->cp_cdb[7] = (length >> 8) & 0xFF; 1922 cp->cp_cdb[8] = length & 0xFF; /* Length LSB */ 1923 cp->cp_cdb[9] = 0; /* No sync, no match bits */ 1924 1925 /* 1926 * This could be optimized to live in dpt_register_buffer. 1927 * We keep it here, just in case the kernel decides to reallocate pages 1928 */ 1929 if (dpt_scatter_gather(dpt, ccb, DPT_RW_BUFFER_SIZE, 1930 dpt->rw_buffer[bus][target][lun])) { 1931 kprintf("dpt%d: Failed to setup Scatter/Gather for " 1932 "Target-Mode buffer\n", dpt->unit); 1933 } 1934 } 1935 1936 /* Setup a target mode READ command */ 1937 1938 static void 1939 dpt_set_target(int redo, dpt_softc_t * dpt, 1940 u_int8_t bus, u_int8_t target, u_int8_t lun, int mode, 1941 u_int16_t length, u_int16_t offset, dpt_ccb_t * ccb) 1942 { 1943 if (dpt->target_mode_enabled) { 1944 crit_enter(); 1945 1946 if (!redo) 1947 dpt_target_ccb(dpt, bus, target, lun, ccb, mode, 1948 SCSI_TM_READ_BUFFER, length, offset); 1949 1950 ccb->transaction_id = ++dpt->commands_processed; 1951 1952 #ifdef DPT_MEASURE_PERFORMANCE 1953 dpt->performance.command_count[ccb->eata_ccb.cp_scsi_cmd]++; 1954 ccb->command_started = microtime_now; 1955 #endif 1956 dpt_Qadd_waiting(dpt, ccb); 1957 dpt_sched_queue(dpt); 1958 1959 crit_exit(); 1960 } else { 1961 kprintf("dpt%d: Target Mode Request, but Target Mode is OFF\n", 1962 dpt->unit); 1963 } 1964 } 1965 1966 /* 1967 * Schedule a buffer to be sent to another target. 1968 * The work will be scheduled and the callback provided will be called when 1969 * the work is actually done. 1970 * 1971 * Please NOTE: ``Anyone'' can send a buffer, but only registered clients 1972 * get notified of receipt of buffers. 1973 */ 1974 1975 int 1976 dpt_send_buffer(int unit, u_int8_t channel, u_int8_t target, u_int8_t lun, 1977 u_int8_t mode, u_int16_t length, u_int16_t offset, void *data, 1978 buff_wr_done callback) 1979 { 1980 dpt_softc_t *dpt; 1981 dpt_ccb_t *ccb = NULL; 1982 1983 /* This is an external call. Be a bit paranoid */ 1984 for (dpt = TAILQ_FIRST(&dpt_softc_list); 1985 dpt != NULL; 1986 dpt = TAILQ_NEXT(dpt, links)) { 1987 if (dpt->unit == unit) 1988 goto valid_unit; 1989 } 1990 1991 return (INVALID_UNIT); 1992 1993 valid_unit: 1994 1995 if (dpt->target_mode_enabled) { 1996 if ((channel >= dpt->channels) || (target > dpt->max_id) || 1997 (lun > dpt->max_lun)) { 1998 return (INVALID_SENDER); 1999 } 2000 if ((dpt->rw_buffer[channel][target][lun] == NULL) || 2001 (dpt->buffer_receiver[channel][target][lun] == NULL)) 2002 return (NOT_REGISTERED); 2003 2004 crit_enter(); 2005 /* Process the free list */ 2006 if ((TAILQ_EMPTY(&dpt->free_ccbs)) && dpt_alloc_freelist(dpt)) { 2007 kprintf("dpt%d ERROR: Cannot allocate any more free CCB's.\n" 2008 " Please try later\n", 2009 dpt->unit); 2010 crit_exit(); 2011 return (NO_RESOURCES); 2012 } 2013 /* Now grab the newest CCB */ 2014 if ((ccb = dpt_Qpop_free(dpt)) == NULL) { 2015 crit_exit(); 2016 panic("dpt%d: Got a NULL CCB from pop_free()\n", dpt->unit); 2017 } 2018 crit_exit(); 2019 2020 bcopy(dpt->rw_buffer[channel][target][lun] + offset, data, length); 2021 dpt_target_ccb(dpt, channel, target, lun, ccb, mode, 2022 SCSI_TM_WRITE_BUFFER, 2023 length, offset); 2024 ccb->std_callback = (ccb_callback) callback; /* Potential trouble */ 2025 2026 crit_enter(); 2027 ccb->transaction_id = ++dpt->commands_processed; 2028 2029 #ifdef DPT_MEASURE_PERFORMANCE 2030 dpt->performance.command_count[ccb->eata_ccb.cp_scsi_cmd]++; 2031 ccb->command_started = microtime_now; 2032 #endif 2033 dpt_Qadd_waiting(dpt, ccb); 2034 dpt_sched_queue(dpt); 2035 2036 crit_exit(); 2037 return (0); 2038 } 2039 return (DRIVER_DOWN); 2040 } 2041 2042 static void 2043 dpt_target_done(dpt_softc_t * dpt, int bus, dpt_ccb_t * ccb) 2044 { 2045 eata_ccb_t *cp; 2046 2047 cp = &ccb->eata_ccb; 2048 2049 /* 2050 * Remove the CCB from the waiting queue. 2051 * We do NOT put it back on the free, etc., queues as it is a special 2052 * ccb, owned by the dpt_softc of this unit. 2053 */ 2054 crit_enter(); 2055 dpt_Qremove_completed(dpt, ccb); 2056 crit_exit(); 2057 2058 #define br_channel (ccb->eata_ccb.cp_channel) 2059 #define br_target (ccb->eata_ccb.cp_id) 2060 #define br_lun (ccb->eata_ccb.cp_LUN) 2061 #define br_index [br_channel][br_target][br_lun] 2062 #define read_buffer_callback (dpt->buffer_receiver br_index ) 2063 #define read_buffer (dpt->rw_buffer[br_channel][br_target][br_lun]) 2064 #define cb(offset) (ccb->eata_ccb.cp_cdb[offset]) 2065 #define br_offset ((cb(3) << 16) | (cb(4) << 8) | cb(5)) 2066 #define br_length ((cb(6) << 16) | (cb(7) << 8) | cb(8)) 2067 2068 /* Different reasons for being here, you know... */ 2069 switch (ccb->eata_ccb.cp_scsi_cmd) { 2070 case SCSI_TM_READ_BUFFER: 2071 if (read_buffer_callback != NULL) { 2072 /* This is a buffer generated by a kernel process */ 2073 read_buffer_callback(dpt->unit, br_channel, 2074 br_target, br_lun, 2075 read_buffer, 2076 br_offset, br_length); 2077 } else { 2078 /* 2079 * This is a buffer waited for by a user (sleeping) 2080 * command 2081 */ 2082 wakeup(ccb); 2083 } 2084 2085 /* We ALWAYS re-issue the same command; args are don't-care */ 2086 dpt_set_target(1, 0, 0, 0, 0, 0, 0, 0, 0); 2087 break; 2088 2089 case SCSI_TM_WRITE_BUFFER: 2090 (ccb->wrbuff_callback) (dpt->unit, br_channel, br_target, 2091 br_offset, br_length, 2092 br_lun, ccb->status_packet.hba_stat); 2093 break; 2094 default: 2095 kprintf("dpt%d: %s is an unsupported command for target mode\n", 2096 dpt->unit, scsi_cmd_name(ccb->eata_ccb.cp_scsi_cmd)); 2097 } 2098 crit_enter(); 2099 dpt->target_ccb[br_channel][br_target][br_lun] = NULL; 2100 dpt_Qpush_free(dpt, ccb); 2101 crit_exit(); 2102 } 2103 2104 2105 /* 2106 * Use this function to register a client for a buffer read target operation. 2107 * The function you register will be called every time a buffer is received 2108 * by the target mode code. 2109 */ 2110 dpt_rb_t 2111 dpt_register_buffer(int unit, u_int8_t channel, u_int8_t target, u_int8_t lun, 2112 u_int8_t mode, u_int16_t length, u_int16_t offset, 2113 dpt_rec_buff callback, dpt_rb_op_t op) 2114 { 2115 dpt_softc_t *dpt; 2116 dpt_ccb_t *ccb = NULL; 2117 2118 for (dpt = TAILQ_FIRST(&dpt_softc_list); 2119 dpt != NULL; 2120 dpt = TAILQ_NEXT(dpt, links)) { 2121 if (dpt->unit == unit) 2122 goto valid_unit; 2123 } 2124 2125 return (INVALID_UNIT); 2126 2127 valid_unit: 2128 2129 if (dpt->state & DPT_HA_SHUTDOWN_ACTIVE) 2130 return (DRIVER_DOWN); 2131 2132 if ((channel > (dpt->channels - 1)) || (target > (dpt->max_id - 1)) || 2133 (lun > (dpt->max_lun - 1))) 2134 return (INVALID_SENDER); 2135 2136 if (dpt->buffer_receiver[channel][target][lun] == NULL) { 2137 if (op == REGISTER_BUFFER) { 2138 /* Assign the requested callback */ 2139 dpt->buffer_receiver[channel][target][lun] = callback; 2140 /* Get a CCB */ 2141 crit_enter(); 2142 2143 /* Process the free list */ 2144 if ((TAILQ_EMPTY(&dpt->free_ccbs)) && dpt_alloc_freelist(dpt)) { 2145 kprintf("dpt%d ERROR: Cannot allocate any more free CCB's.\n" 2146 " Please try later\n", 2147 dpt->unit); 2148 crit_exit(); 2149 return (NO_RESOURCES); 2150 } 2151 /* Now grab the newest CCB */ 2152 if ((ccb = dpt_Qpop_free(dpt)) == NULL) { 2153 crit_exit(); 2154 panic("dpt%d: Got a NULL CCB from pop_free()\n", 2155 dpt->unit); 2156 } 2157 crit_exit(); 2158 2159 /* Clean up the leftover of the previous tenant */ 2160 ccb->status = DPT_CCB_STATE_NEW; 2161 dpt->target_ccb[channel][target][lun] = ccb; 2162 2163 dpt->rw_buffer[channel][target][lun] = 2164 kmalloc(DPT_RW_BUFFER_SIZE, M_DEVBUF, M_INTWAIT); 2165 dpt_set_target(0, dpt, channel, target, lun, mode, 2166 length, offset, ccb); 2167 return (SUCCESSFULLY_REGISTERED); 2168 } else 2169 return (NOT_REGISTERED); 2170 } else { 2171 if (op == REGISTER_BUFFER) { 2172 if (dpt->buffer_receiver[channel][target][lun] == callback) 2173 return (ALREADY_REGISTERED); 2174 else 2175 return (REGISTERED_TO_ANOTHER); 2176 } else { 2177 if (dpt->buffer_receiver[channel][target][lun] == callback) { 2178 dpt->buffer_receiver[channel][target][lun] = NULL; 2179 crit_enter(); 2180 dpt_Qpush_free(dpt, ccb); 2181 crit_exit(); 2182 kfree(dpt->rw_buffer[channel][target][lun], M_DEVBUF); 2183 return (SUCCESSFULLY_REGISTERED); 2184 } else 2185 return (INVALID_CALLBACK); 2186 } 2187 2188 } 2189 } 2190 2191 /* Return the state of the blinking DPT LED's */ 2192 u_int8_t 2193 dpt_blinking_led(dpt_softc_t * dpt) 2194 { 2195 int ndx; 2196 u_int32_t state; 2197 u_int32_t previous; 2198 u_int8_t result; 2199 2200 crit_enter(); 2201 2202 result = 0; 2203 2204 for (ndx = 0, state = 0, previous = 0; 2205 (ndx < 10) && (state != previous); 2206 ndx++) { 2207 previous = state; 2208 state = dpt_inl(dpt, 1); 2209 } 2210 2211 if ((state == previous) && (state == DPT_BLINK_INDICATOR)) 2212 result = dpt_inb(dpt, 5); 2213 2214 crit_exit(); 2215 return (result); 2216 } 2217 2218 /* 2219 * Execute a command which did not come from the kernel's SCSI layer. 2220 * The only way to map user commands to bus and target is to comply with the 2221 * standard DPT wire-down scheme: 2222 */ 2223 int 2224 dpt_user_cmd(dpt_softc_t * dpt, eata_pt_t * user_cmd, 2225 caddr_t cmdarg, int minor_no) 2226 { 2227 dpt_ccb_t *ccb; 2228 void *data; 2229 int channel, target, lun; 2230 int huh; 2231 int result; 2232 int submitted; 2233 2234 data = NULL; 2235 channel = minor2hba(minor_no); 2236 target = minor2target(minor_no); 2237 lun = minor2lun(minor_no); 2238 2239 if ((channel > (dpt->channels - 1)) 2240 || (target > dpt->max_id) 2241 || (lun > dpt->max_lun)) 2242 return (ENXIO); 2243 2244 if (target == dpt->sc_scsi_link[channel].adapter_targ) { 2245 /* This one is for the controller itself */ 2246 if ((user_cmd->eataID[0] != 'E') 2247 || (user_cmd->eataID[1] != 'A') 2248 || (user_cmd->eataID[2] != 'T') 2249 || (user_cmd->eataID[3] != 'A')) { 2250 return (ENXIO); 2251 } 2252 } 2253 /* Get a DPT CCB, so we can prepare a command */ 2254 crit_enter(); 2255 2256 /* Process the free list */ 2257 if ((TAILQ_EMPTY(&dpt->free_ccbs)) && dpt_alloc_freelist(dpt)) { 2258 kprintf("dpt%d ERROR: Cannot allocate any more free CCB's.\n" 2259 " Please try later\n", 2260 dpt->unit); 2261 crit_exit(); 2262 return (EFAULT); 2263 } 2264 /* Now grab the newest CCB */ 2265 if ((ccb = dpt_Qpop_free(dpt)) == NULL) { 2266 crit_exit(); 2267 panic("dpt%d: Got a NULL CCB from pop_free()\n", dpt->unit); 2268 } else { 2269 crit_exit(); 2270 /* Clean up the leftover of the previous tenant */ 2271 ccb->status = DPT_CCB_STATE_NEW; 2272 } 2273 2274 bcopy((caddr_t) & user_cmd->command_packet, (caddr_t) & ccb->eata_ccb, 2275 sizeof(eata_ccb_t)); 2276 2277 /* We do not want to do user specified scatter/gather. Why?? */ 2278 if (ccb->eata_ccb.scatter == 1) 2279 return (EINVAL); 2280 2281 ccb->eata_ccb.Auto_Req_Sen = 1; 2282 ccb->eata_ccb.reqlen = htonl(sizeof(struct scsi_sense_data)); 2283 ccb->eata_ccb.cp_datalen = htonl(sizeof(ccb->eata_ccb.cp_datalen)); 2284 ccb->eata_ccb.cp_dataDMA = htonl(vtophys(ccb->eata_ccb.cp_dataDMA)); 2285 ccb->eata_ccb.cp_statDMA = htonl(vtophys(&ccb->eata_ccb.cp_statDMA)); 2286 ccb->eata_ccb.cp_reqDMA = htonl(vtophys(&ccb->eata_ccb.cp_reqDMA)); 2287 ccb->eata_ccb.cp_viraddr = (u_int32_t) & ccb; 2288 2289 if (ccb->eata_ccb.DataIn || ccb->eata_ccb.DataOut) { 2290 /* Data I/O is involved in this command. Alocate buffer */ 2291 if (ccb->eata_ccb.cp_datalen > PAGE_SIZE) { 2292 data = contigmalloc(ccb->eata_ccb.cp_datalen, 2293 M_TEMP, M_WAITOK, 0, ~0, 2294 ccb->eata_ccb.cp_datalen, 2295 0x10000); 2296 } else { 2297 data = kmalloc(ccb->eata_ccb.cp_datalen, M_TEMP, 2298 M_WAITOK); 2299 } 2300 2301 if (data == NULL) { 2302 kprintf("dpt%d: Cannot allocate %d bytes " 2303 "for EATA command\n", dpt->unit, 2304 ccb->eata_ccb.cp_datalen); 2305 return (EFAULT); 2306 } 2307 #define usr_cmd_DMA (caddr_t)user_cmd->command_packet.cp_dataDMA 2308 if (ccb->eata_ccb.DataIn == 1) { 2309 if (copyin(usr_cmd_DMA, 2310 data, ccb->eata_ccb.cp_datalen) == -1) 2311 return (EFAULT); 2312 } 2313 } else { 2314 /* No data I/O involved here. Make sure the DPT knows that */ 2315 ccb->eata_ccb.cp_datalen = 0; 2316 data = NULL; 2317 } 2318 2319 if (ccb->eata_ccb.FWNEST == 1) 2320 ccb->eata_ccb.FWNEST = 0; 2321 2322 if (ccb->eata_ccb.cp_datalen != 0) { 2323 if (dpt_scatter_gather(dpt, ccb, ccb->eata_ccb.cp_datalen, 2324 data) != 0) { 2325 if (data != NULL) 2326 kfree(data, M_TEMP); 2327 return (EFAULT); 2328 } 2329 } 2330 /** 2331 * We are required to quiet a SCSI bus. 2332 * since we do not queue comands on a bus basis, 2333 * we wait for ALL commands on a controller to complete. 2334 * In the mean time, sched_queue() will not schedule new commands. 2335 */ 2336 if ((ccb->eata_ccb.cp_cdb[0] == MULTIFUNCTION_CMD) 2337 && (ccb->eata_ccb.cp_cdb[2] == BUS_QUIET)) { 2338 /* We wait for ALL traffic for this HBa to subside */ 2339 crit_enter(); 2340 dpt->state |= DPT_HA_QUIET; 2341 crit_exit(); 2342 2343 while ((submitted = dpt->submitted_ccbs_count) != 0) { 2344 huh = tsleep((void *) dpt, PCATCH, "dptqt", 100 * hz); 2345 switch (huh) { 2346 case 0: 2347 /* Wakeup call received */ 2348 break; 2349 case EWOULDBLOCK: 2350 /* Timer Expired */ 2351 break; 2352 default: 2353 /* anything else */ 2354 break; 2355 } 2356 } 2357 } 2358 /* Resume normal operation */ 2359 if ((ccb->eata_ccb.cp_cdb[0] == MULTIFUNCTION_CMD) 2360 && (ccb->eata_ccb.cp_cdb[2] == BUS_UNQUIET)) { 2361 crit_enter(); 2362 dpt->state &= ~DPT_HA_QUIET; 2363 crit_exit(); 2364 } 2365 /** 2366 * Schedule the command and submit it. 2367 * We bypass dpt_sched_queue, as it will block on DPT_HA_QUIET 2368 */ 2369 ccb->xs = NULL; 2370 ccb->flags = 0; 2371 ccb->eata_ccb.Auto_Req_Sen = 1; /* We always want this feature */ 2372 2373 ccb->transaction_id = ++dpt->commands_processed; 2374 ccb->std_callback = (ccb_callback) dpt_user_cmd_done; 2375 ccb->result = (u_int32_t) & cmdarg; 2376 ccb->data = data; 2377 2378 #ifdef DPT_MEASURE_PERFORMANCE 2379 ++dpt->performance.command_count[ccb->eata_ccb.cp_scsi_cmd]; 2380 ccb->command_started = microtime_now; 2381 #endif 2382 crit_enter(); 2383 dpt_Qadd_waiting(dpt, ccb); 2384 crit_exit(); 2385 2386 dpt_sched_queue(dpt); 2387 2388 /* Wait for the command to complete */ 2389 (void) tsleep((void *) ccb, PCATCH, "dptucw", 100 * hz); 2390 2391 /* Free allocated memory */ 2392 if (data != NULL) 2393 kfree(data, M_TEMP); 2394 2395 return (0); 2396 } 2397 2398 static void 2399 dpt_user_cmd_done(dpt_softc_t * dpt, int bus, dpt_ccb_t * ccb) 2400 { 2401 u_int32_t result; 2402 caddr_t cmd_arg; 2403 2404 crit_enter(); 2405 2406 /** 2407 * If Auto Request Sense is on, copyout the sense struct 2408 */ 2409 #define usr_pckt_DMA (caddr_t)(intptr_t)ntohl(ccb->eata_ccb.cp_reqDMA) 2410 #define usr_pckt_len ntohl(ccb->eata_ccb.cp_datalen) 2411 if (ccb->eata_ccb.Auto_Req_Sen == 1) { 2412 if (copyout((caddr_t) & ccb->sense_data, usr_pckt_DMA, 2413 sizeof(struct scsi_sense_data))) { 2414 ccb->result = EFAULT; 2415 dpt_Qpush_free(dpt, ccb); 2416 crit_exit(); 2417 wakeup(ccb); 2418 return; 2419 } 2420 } 2421 /* If DataIn is on, copyout the data */ 2422 if ((ccb->eata_ccb.DataIn == 1) 2423 && (ccb->status_packet.hba_stat == HA_NO_ERROR)) { 2424 if (copyout(ccb->data, usr_pckt_DMA, usr_pckt_len)) { 2425 dpt_Qpush_free(dpt, ccb); 2426 ccb->result = EFAULT; 2427 2428 crit_exit(); 2429 wakeup(ccb); 2430 return; 2431 } 2432 } 2433 /* Copyout the status */ 2434 result = ccb->status_packet.hba_stat; 2435 cmd_arg = (caddr_t) ccb->result; 2436 2437 if (copyout((caddr_t) & result, cmd_arg, sizeof(result))) { 2438 dpt_Qpush_free(dpt, ccb); 2439 ccb->result = EFAULT; 2440 crit_exit(); 2441 wakeup(ccb); 2442 return; 2443 } 2444 /* Put the CCB back in the freelist */ 2445 ccb->state |= DPT_CCB_STATE_COMPLETED; 2446 dpt_Qpush_free(dpt, ccb); 2447 2448 /* Free allocated memory */ 2449 crit_exit(); 2450 return; 2451 } 2452 2453 #ifdef DPT_HANDLE_TIMEOUTS 2454 /** 2455 * This function walks down the SUBMITTED queue. 2456 * Every request that is too old gets aborted and marked. 2457 * Since the DPT will complete (interrupt) immediately (what does that mean?), 2458 * We just walk the list, aborting old commands and marking them as such. 2459 * The dpt_complete function will get rid of the that were interrupted in the 2460 * normal manner. 2461 * 2462 * This function needs to run at splcam(), as it interacts with the submitted 2463 * queue, as well as the completed and free queues. Just like dpt_intr() does. 2464 * To run it at any ISPL other than that of dpt_intr(), will mean that dpt_intr 2465 * willbe able to pre-empt it, grab a transaction in progress (towards 2466 * destruction) and operate on it. The state of this transaction will be not 2467 * very clear. 2468 * The only other option, is to lock it only as long as necessary but have 2469 * dpt_intr() spin-wait on it. In a UP environment this makes no sense and in 2470 * a SMP environment, the advantage is dubvious for a function that runs once 2471 * every ten seconds for few microseconds and, on systems with healthy 2472 * hardware, does not do anything anyway. 2473 */ 2474 2475 static void 2476 dpt_handle_timeouts(dpt_softc_t * dpt) 2477 { 2478 dpt_ccb_t *ccb; 2479 2480 crit_enter(); 2481 2482 if (dpt->state & DPT_HA_TIMEOUTS_ACTIVE) { 2483 kprintf("dpt%d WARNING: Timeout Handling Collision\n", 2484 dpt->unit); 2485 crit_exit(); 2486 return; 2487 } 2488 dpt->state |= DPT_HA_TIMEOUTS_ACTIVE; 2489 2490 /* Loop through the entire submitted queue, looking for lost souls */ 2491 for (ccb = TAILQ_FIRST(&dpt->submitted_ccbs); 2492 ccb != NULL; 2493 ccb = TAILQ_NEXT(ccb, links)) { 2494 struct scsi_xfer *xs; 2495 u_int32_t age, max_age; 2496 2497 xs = ccb->xs; 2498 age = dpt_time_delta(ccb->command_started, microtime_now); 2499 2500 #define TenSec 10000000 2501 2502 if (xs == NULL) { /* Local, non-kernel call */ 2503 max_age = TenSec; 2504 } else { 2505 max_age = (((xs->timeout * (dpt->submitted_ccbs_count 2506 + DPT_TIMEOUT_FACTOR)) 2507 > TenSec) 2508 ? (xs->timeout * (dpt->submitted_ccbs_count 2509 + DPT_TIMEOUT_FACTOR)) 2510 : TenSec); 2511 } 2512 2513 /* 2514 * If a transaction is marked lost and is TWICE as old as we 2515 * care, then, and only then do we destroy it! 2516 */ 2517 if (ccb->state & DPT_CCB_STATE_MARKED_LOST) { 2518 /* Remember who is next */ 2519 if (age > (max_age * 2)) { 2520 dpt_Qremove_submitted(dpt, ccb); 2521 ccb->state &= ~DPT_CCB_STATE_MARKED_LOST; 2522 ccb->state |= DPT_CCB_STATE_ABORTED; 2523 #define cmd_name scsi_cmd_name(ccb->eata_ccb.cp_scsi_cmd) 2524 if (ccb->retries++ > DPT_RETRIES) { 2525 kprintf("dpt%d ERROR: Destroying stale " 2526 "%d (%s)\n" 2527 " on " 2528 "c%db%dt%du%d (%d/%d)\n", 2529 dpt->unit, ccb->transaction_id, 2530 cmd_name, 2531 dpt->unit, 2532 ccb->eata_ccb.cp_channel, 2533 ccb->eata_ccb.cp_id, 2534 ccb->eata_ccb.cp_LUN, age, 2535 ccb->retries); 2536 #define send_ccb &ccb->eata_ccb 2537 #define ESA EATA_SPECIFIC_ABORT 2538 (void) dpt_send_immediate(dpt, 2539 send_ccb, 2540 ESA, 2541 0, 0); 2542 dpt_Qpush_free(dpt, ccb); 2543 2544 /* The SCSI layer should re-try */ 2545 xs->error |= XS_TIMEOUT; 2546 xs->flags |= SCSI_ITSDONE; 2547 scsi_done(xs); 2548 } else { 2549 kprintf("dpt%d ERROR: Stale %d (%s) on " 2550 "c%db%dt%du%d (%d)\n" 2551 " gets another " 2552 "chance(%d/%d)\n", 2553 dpt->unit, ccb->transaction_id, 2554 cmd_name, 2555 dpt->unit, 2556 ccb->eata_ccb.cp_channel, 2557 ccb->eata_ccb.cp_id, 2558 ccb->eata_ccb.cp_LUN, 2559 age, ccb->retries, DPT_RETRIES); 2560 2561 dpt_Qpush_waiting(dpt, ccb); 2562 dpt_sched_queue(dpt); 2563 } 2564 } 2565 } else { 2566 /* 2567 * This is a transaction that is not to be destroyed 2568 * (yet) But it is too old for our liking. We wait as 2569 * long as the upper layer thinks. Not really, we 2570 * multiply that by the number of commands in the 2571 * submitted queue + 1. 2572 */ 2573 if (!(ccb->state & DPT_CCB_STATE_MARKED_LOST) && 2574 (age != ~0) && (age > max_age)) { 2575 kprintf("dpt%d ERROR: Marking %d (%s) on " 2576 "c%db%dt%du%d \n" 2577 " as late after %dusec\n", 2578 dpt->unit, ccb->transaction_id, 2579 cmd_name, 2580 dpt->unit, ccb->eata_ccb.cp_channel, 2581 ccb->eata_ccb.cp_id, 2582 ccb->eata_ccb.cp_LUN, age); 2583 ccb->state |= DPT_CCB_STATE_MARKED_LOST; 2584 } 2585 } 2586 } 2587 2588 dpt->state &= ~DPT_HA_TIMEOUTS_ACTIVE; 2589 crit_exit(); 2590 } 2591 2592 #endif /* DPT_HANDLE_TIMEOUTS */ 2593 2594 #endif 2595