xref: /netbsd-src/sys/dev/ic/icp.c (revision b1c86f5f087524e68db12794ee9c3e3da1ab17a0)
1 /*	$NetBSD: icp.c,v 1.29 2008/04/30 14:07:13 ad Exp $	*/
2 
3 /*-
4  * Copyright (c) 2002, 2003 The NetBSD Foundation, Inc.
5  * All rights reserved.
6  *
7  * This code is derived from software contributed to The NetBSD Foundation
8  * by Andrew Doran, and by Jason R. Thorpe of Wasabi Systems, Inc.
9  *
10  * Redistribution and use in source and binary forms, with or without
11  * modification, are permitted provided that the following conditions
12  * are met:
13  * 1. Redistributions of source code must retain the above copyright
14  *    notice, this list of conditions and the following disclaimer.
15  * 2. Redistributions in binary form must reproduce the above copyright
16  *    notice, this list of conditions and the following disclaimer in the
17  *    documentation and/or other materials provided with the distribution.
18  *
19  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
20  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
21  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
22  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
23  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29  * POSSIBILITY OF SUCH DAMAGE.
30  */
31 
32 /*
33  * Copyright (c) 1999, 2000 Niklas Hallqvist.  All rights reserved.
34  *
35  * Redistribution and use in source and binary forms, with or without
36  * modification, are permitted provided that the following conditions
37  * are met:
38  * 1. Redistributions of source code must retain the above copyright
39  *    notice, this list of conditions and the following disclaimer.
40  * 2. Redistributions in binary form must reproduce the above copyright
41  *    notice, this list of conditions and the following disclaimer in the
42  *    documentation and/or other materials provided with the distribution.
43  * 3. All advertising materials mentioning features or use of this software
44  *    must display the following acknowledgement:
45  *	This product includes software developed by Niklas Hallqvist.
46  * 4. The name of the author may not be used to endorse or promote products
47  *    derived from this software without specific prior written permission.
48  *
49  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
50  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
51  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
52  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
53  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
54   * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
55  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
56  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
57  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
58  * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
59  *
60  * from OpenBSD: gdt_common.c,v 1.12 2001/07/04 06:43:18 niklas Exp
61  */
62 
63 /*
64  * This driver would not have written if it was not for the hardware donations
65  * from both ICP-Vortex and �ko.neT.  I want to thank them for their support.
66  *
67  * Re-worked for NetBSD by Andrew Doran.  Test hardware kindly supplied by
68  * Intel.
69  *
70  * Support for the ICP-Vortex management tools added by
71  * Jason R. Thorpe of Wasabi Systems, Inc., based on code
72  * provided by Achim Leubner <achim.leubner@intel.com>.
73  *
74  * Additional support for dynamic rescan of cacheservice drives by
75  * Jason R. Thorpe of Wasabi Systems, Inc.
76  */
77 
78 #include <sys/cdefs.h>
79 __KERNEL_RCSID(0, "$NetBSD: icp.c,v 1.29 2008/04/30 14:07:13 ad Exp $");
80 
81 #include <sys/param.h>
82 #include <sys/systm.h>
83 #include <sys/kernel.h>
84 #include <sys/device.h>
85 #include <sys/queue.h>
86 #include <sys/proc.h>
87 #include <sys/buf.h>
88 #include <sys/endian.h>
89 #include <sys/malloc.h>
90 #include <sys/disk.h>
91 
92 #include <uvm/uvm_extern.h>
93 
94 #include <sys/bswap.h>
95 #include <sys/bus.h>
96 
97 #include <dev/pci/pcireg.h>
98 #include <dev/pci/pcivar.h>
99 #include <dev/pci/pcidevs.h>
100 
101 #include <dev/ic/icpreg.h>
102 #include <dev/ic/icpvar.h>
103 
104 #include <dev/scsipi/scsipi_all.h>
105 #include <dev/scsipi/scsiconf.h>
106 
107 #include "locators.h"
108 
109 int	icp_async_event(struct icp_softc *, int);
110 void	icp_ccb_submit(struct icp_softc *icp, struct icp_ccb *ic);
111 void	icp_chain(struct icp_softc *);
112 int	icp_print(void *, const char *);
113 void	icp_watchdog(void *);
114 void	icp_ucmd_intr(struct icp_ccb *);
115 void	icp_recompute_openings(struct icp_softc *);
116 
117 int	icp_count;	/* total # of controllers, for ioctl interface */
118 
119 /*
120  * Statistics for the ioctl interface to query.
121  *
122  * XXX Global.  They should probably be made per-controller
123  * XXX at some point.
124  */
125 gdt_statist_t icp_stats;
126 
127 int
128 icp_init(struct icp_softc *icp, const char *intrstr)
129 {
130 	struct icp_attach_args icpa;
131 	struct icp_binfo binfo;
132 	struct icp_ccb *ic;
133 	u_int16_t cdev_cnt;
134 	int i, j, state, feat, nsegs, rv;
135 	int locs[ICPCF_NLOCS];
136 
137 	state = 0;
138 
139 	if (intrstr != NULL)
140 		aprint_normal_dev(&icp->icp_dv, "interrupting at %s\n",
141 		    intrstr);
142 
143 	SIMPLEQ_INIT(&icp->icp_ccb_queue);
144 	SIMPLEQ_INIT(&icp->icp_ccb_freelist);
145 	SIMPLEQ_INIT(&icp->icp_ucmd_queue);
146 	callout_init(&icp->icp_wdog_callout, 0);
147 
148 	/*
149 	 * Allocate a scratch area.
150 	 */
151 	if (bus_dmamap_create(icp->icp_dmat, ICP_SCRATCH_SIZE, 1,
152 	    ICP_SCRATCH_SIZE, 0, BUS_DMA_NOWAIT | BUS_DMA_ALLOCNOW,
153 	    &icp->icp_scr_dmamap) != 0) {
154 		aprint_error_dev(&icp->icp_dv, "cannot create scratch dmamap\n");
155 		return (1);
156 	}
157 	state++;
158 
159 	if (bus_dmamem_alloc(icp->icp_dmat, ICP_SCRATCH_SIZE, PAGE_SIZE, 0,
160 	    icp->icp_scr_seg, 1, &nsegs, BUS_DMA_NOWAIT) != 0) {
161 		aprint_error_dev(&icp->icp_dv, "cannot alloc scratch dmamem\n");
162 		goto bail_out;
163 	}
164 	state++;
165 
166 	if (bus_dmamem_map(icp->icp_dmat, icp->icp_scr_seg, nsegs,
167 	    ICP_SCRATCH_SIZE, &icp->icp_scr, 0)) {
168 		aprint_error_dev(&icp->icp_dv, "cannot map scratch dmamem\n");
169 		goto bail_out;
170 	}
171 	state++;
172 
173 	if (bus_dmamap_load(icp->icp_dmat, icp->icp_scr_dmamap, icp->icp_scr,
174 	    ICP_SCRATCH_SIZE, NULL, BUS_DMA_NOWAIT)) {
175 		aprint_error_dev(&icp->icp_dv, "cannot load scratch dmamap\n");
176 		goto bail_out;
177 	}
178 	state++;
179 
180 	/*
181 	 * Allocate and initialize the command control blocks.
182 	 */
183 	ic = malloc(sizeof(*ic) * ICP_NCCBS, M_DEVBUF, M_NOWAIT | M_ZERO);
184 	if ((icp->icp_ccbs = ic) == NULL) {
185 		aprint_error_dev(&icp->icp_dv, "malloc() failed\n");
186 		goto bail_out;
187 	}
188 	state++;
189 
190 	for (i = 0; i < ICP_NCCBS; i++, ic++) {
191 		/*
192 		 * The first two command indexes have special meanings, so
193 		 * we can't use them.
194 		 */
195 		ic->ic_ident = i + 2;
196 		rv = bus_dmamap_create(icp->icp_dmat, ICP_MAX_XFER,
197 		    ICP_MAXSG, ICP_MAX_XFER, 0,
198 		    BUS_DMA_NOWAIT | BUS_DMA_ALLOCNOW,
199 		    &ic->ic_xfer_map);
200 		if (rv != 0)
201 			break;
202 		icp->icp_nccbs++;
203 		icp_ccb_free(icp, ic);
204 	}
205 #ifdef DIAGNOSTIC
206 	if (icp->icp_nccbs != ICP_NCCBS)
207 		aprint_error_dev(&icp->icp_dv, "%d/%d CCBs usable\n",
208 		    icp->icp_nccbs, ICP_NCCBS);
209 #endif
210 
211 	/*
212 	 * Initalize the controller.
213 	 */
214 	if (!icp_cmd(icp, ICP_SCREENSERVICE, ICP_INIT, 0, 0, 0)) {
215 		aprint_error_dev(&icp->icp_dv, "screen service init error %d\n",
216 		    icp->icp_status);
217 		goto bail_out;
218 	}
219 
220 	if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INIT, ICP_LINUX_OS, 0, 0)) {
221 		aprint_error_dev(&icp->icp_dv, "cache service init error %d\n",
222 		    icp->icp_status);
223 		goto bail_out;
224 	}
225 
226 	icp_cmd(icp, ICP_CACHESERVICE, ICP_UNFREEZE_IO, 0, 0, 0);
227 
228 	if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_MOUNT, 0xffff, 1, 0)) {
229 		aprint_error_dev(&icp->icp_dv, "cache service mount error %d\n",
230 		    icp->icp_status);
231 		goto bail_out;
232 	}
233 
234 	if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INIT, ICP_LINUX_OS, 0, 0)) {
235 		aprint_error_dev(&icp->icp_dv, "cache service post-mount init error %d\n",
236 		    icp->icp_status);
237 		goto bail_out;
238 	}
239 	cdev_cnt = (u_int16_t)icp->icp_info;
240 	icp->icp_fw_vers = icp->icp_service;
241 
242 	if (!icp_cmd(icp, ICP_SCSIRAWSERVICE, ICP_INIT, 0, 0, 0)) {
243 		aprint_error_dev(&icp->icp_dv, "raw service init error %d\n",
244 		    icp->icp_status);
245 		goto bail_out;
246 	}
247 
248 	/*
249 	 * Set/get raw service features (scatter/gather).
250 	 */
251 	feat = 0;
252 	if (icp_cmd(icp, ICP_SCSIRAWSERVICE, ICP_SET_FEAT, ICP_SCATTER_GATHER,
253 	    0, 0))
254 		if (icp_cmd(icp, ICP_SCSIRAWSERVICE, ICP_GET_FEAT, 0, 0, 0))
255 			feat = icp->icp_info;
256 
257 	if ((feat & ICP_SCATTER_GATHER) == 0) {
258 #ifdef DIAGNOSTIC
259 		aprint_normal_dev(&icp->icp_dv,
260 		    "scatter/gather not supported (raw service)\n");
261 #endif
262 	} else
263 		icp->icp_features |= ICP_FEAT_RAWSERVICE;
264 
265 	/*
266 	 * Set/get cache service features (scatter/gather).
267 	 */
268 	feat = 0;
269 	if (icp_cmd(icp, ICP_CACHESERVICE, ICP_SET_FEAT, 0,
270 	    ICP_SCATTER_GATHER, 0))
271 		if (icp_cmd(icp, ICP_CACHESERVICE, ICP_GET_FEAT, 0, 0, 0))
272 			feat = icp->icp_info;
273 
274 	if ((feat & ICP_SCATTER_GATHER) == 0) {
275 #ifdef DIAGNOSTIC
276 		aprint_normal_dev(&icp->icp_dv,
277 		    "scatter/gather not supported (cache service)\n");
278 #endif
279 	} else
280 		icp->icp_features |= ICP_FEAT_CACHESERVICE;
281 
282 	/*
283 	 * Pull some information from the board and dump.
284 	 */
285 	if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_IOCTL, ICP_BOARD_INFO,
286 	    ICP_INVALID_CHANNEL, sizeof(struct icp_binfo))) {
287 		aprint_error_dev(&icp->icp_dv, "unable to retrive board info\n");
288 		goto bail_out;
289 	}
290 	memcpy(&binfo, icp->icp_scr, sizeof(binfo));
291 
292 	aprint_normal_dev(&icp->icp_dv,
293 	    "model <%s>, firmware <%s>, %d channel(s), %dMB memory\n",
294 	    binfo.bi_type_string, binfo.bi_raid_string,
295 	    binfo.bi_chan_count, le32toh(binfo.bi_memsize) >> 20);
296 
297 	/*
298 	 * Determine the number of devices, and number of openings per
299 	 * device.
300 	 */
301 	if (icp->icp_features & ICP_FEAT_CACHESERVICE) {
302 		for (j = 0; j < cdev_cnt && j < ICP_MAX_HDRIVES; j++) {
303 			if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INFO, j, 0,
304 			    0))
305 				continue;
306 
307 			icp->icp_cdr[j].cd_size = icp->icp_info;
308 			if (icp->icp_cdr[j].cd_size != 0)
309 				icp->icp_ndevs++;
310 
311 			if (icp_cmd(icp, ICP_CACHESERVICE, ICP_DEVTYPE, j, 0,
312 			    0))
313 				icp->icp_cdr[j].cd_type = icp->icp_info;
314 		}
315 	}
316 
317 	if (icp->icp_features & ICP_FEAT_RAWSERVICE) {
318 		icp->icp_nchan = binfo.bi_chan_count;
319 		icp->icp_ndevs += icp->icp_nchan;
320 	}
321 
322 	icp_recompute_openings(icp);
323 
324 	/*
325 	 * Attach SCSI channels.
326 	 */
327 	if (icp->icp_features & ICP_FEAT_RAWSERVICE) {
328 		struct icp_ioc_version *iv;
329 		struct icp_rawioc *ri;
330 		struct icp_getch *gc;
331 
332 		iv = (struct icp_ioc_version *)icp->icp_scr;
333 		iv->iv_version = htole32(ICP_IOC_NEWEST);
334 		iv->iv_listents = ICP_MAXBUS;
335 		iv->iv_firstchan = 0;
336 		iv->iv_lastchan = ICP_MAXBUS - 1;
337 		iv->iv_listoffset = htole32(sizeof(*iv));
338 
339 		if (icp_cmd(icp, ICP_CACHESERVICE, ICP_IOCTL,
340 		    ICP_IOCHAN_RAW_DESC, ICP_INVALID_CHANNEL,
341 		    sizeof(*iv) + ICP_MAXBUS * sizeof(*ri))) {
342 			ri = (struct icp_rawioc *)(iv + 1);
343 			for (j = 0; j < binfo.bi_chan_count; j++, ri++)
344 				icp->icp_bus_id[j] = ri->ri_procid;
345 		} else {
346 			/*
347 			 * Fall back to the old method.
348 			 */
349 			gc = (struct icp_getch *)icp->icp_scr;
350 
351 			for (j = 0; j < binfo.bi_chan_count; j++) {
352 				if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_IOCTL,
353 				    ICP_SCSI_CHAN_CNT | ICP_L_CTRL_PATTERN,
354 				    ICP_IO_CHANNEL | ICP_INVALID_CHANNEL,
355 				    sizeof(*gc))) {
356 				    	aprint_error_dev(&icp->icp_dv,
357 					    "unable to get chan info");
358 					goto bail_out;
359 				}
360 				icp->icp_bus_id[j] = gc->gc_scsiid;
361 			}
362 		}
363 
364 		for (j = 0; j < binfo.bi_chan_count; j++) {
365 			if (icp->icp_bus_id[j] > ICP_MAXID_FC)
366 				icp->icp_bus_id[j] = ICP_MAXID_FC;
367 
368 			icpa.icpa_unit = j + ICPA_UNIT_SCSI;
369 
370 			locs[ICPCF_UNIT] = j + ICPA_UNIT_SCSI;
371 
372 			icp->icp_children[icpa.icpa_unit] =
373 				config_found_sm_loc(&icp->icp_dv, "icp", locs,
374 					&icpa, icp_print, config_stdsubmatch);
375 		}
376 	}
377 
378 	/*
379 	 * Attach cache devices.
380 	 */
381 	if (icp->icp_features & ICP_FEAT_CACHESERVICE) {
382 		for (j = 0; j < cdev_cnt && j < ICP_MAX_HDRIVES; j++) {
383 			if (icp->icp_cdr[j].cd_size == 0)
384 				continue;
385 
386 			icpa.icpa_unit = j;
387 
388 			locs[ICPCF_UNIT] = j;
389 
390 			icp->icp_children[icpa.icpa_unit] =
391 			    config_found_sm_loc(&icp->icp_dv, "icp", locs,
392 				&icpa, icp_print, config_stdsubmatch);
393 		}
394 	}
395 
396 	/*
397 	 * Start the watchdog.
398 	 */
399 	icp_watchdog(icp);
400 
401 	/*
402 	 * Count the controller, and we're done!
403 	 */
404 	if (icp_count++ == 0)
405 		mutex_init(&icp_ioctl_mutex, MUTEX_DEFAULT, IPL_NONE);
406 
407 	return (0);
408 
409  bail_out:
410 	if (state > 4)
411 		for (j = 0; j < i; j++)
412 			bus_dmamap_destroy(icp->icp_dmat,
413 			    icp->icp_ccbs[j].ic_xfer_map);
414  	if (state > 3)
415 		free(icp->icp_ccbs, M_DEVBUF);
416 	if (state > 2)
417 		bus_dmamap_unload(icp->icp_dmat, icp->icp_scr_dmamap);
418 	if (state > 1)
419 		bus_dmamem_unmap(icp->icp_dmat, icp->icp_scr,
420 		    ICP_SCRATCH_SIZE);
421 	if (state > 0)
422 		bus_dmamem_free(icp->icp_dmat, icp->icp_scr_seg, nsegs);
423 	bus_dmamap_destroy(icp->icp_dmat, icp->icp_scr_dmamap);
424 
425 	return (1);
426 }
427 
428 void
429 icp_register_servicecb(struct icp_softc *icp, int unit,
430     const struct icp_servicecb *cb)
431 {
432 
433 	icp->icp_servicecb[unit] = cb;
434 }
435 
436 void
437 icp_rescan(struct icp_softc *icp, int unit)
438 {
439 	struct icp_attach_args icpa;
440 	u_int newsize, newtype;
441 	int locs[ICPCF_NLOCS];
442 
443 	/*
444 	 * NOTE: It is very important that the queue be frozen and not
445 	 * commands running when this is called.  The ioctl mutex must
446 	 * also be held.
447 	 */
448 
449 	KASSERT(icp->icp_qfreeze != 0);
450 	KASSERT(icp->icp_running == 0);
451 	KASSERT(unit < ICP_MAX_HDRIVES);
452 
453 	if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INFO, unit, 0, 0)) {
454 #ifdef ICP_DEBUG
455 		printf("%s: rescan: unit %d ICP_INFO failed -> 0x%04x\n",
456 		    device_xname(&icp->icp_dv), unit, icp->icp_status);
457 #endif
458 		goto gone;
459 	}
460 	if ((newsize = icp->icp_info) == 0) {
461 #ifdef ICP_DEBUG
462 		printf("%s: rescan: unit %d has zero size\n",
463 		    device_xname(&icp->icp_dv), unit);
464 #endif
465  gone:
466 		/*
467 		 * Host drive is no longer present; detach if a child
468 		 * is currently there.
469 		 */
470 		if (icp->icp_cdr[unit].cd_size != 0)
471 			icp->icp_ndevs--;
472 		icp->icp_cdr[unit].cd_size = 0;
473 		if (icp->icp_children[unit] != NULL) {
474 			(void) config_detach(icp->icp_children[unit],
475 			    DETACH_FORCE);
476 			icp->icp_children[unit] = NULL;
477 		}
478 		return;
479 	}
480 
481 	if (icp_cmd(icp, ICP_CACHESERVICE, ICP_DEVTYPE, unit, 0, 0))
482 		newtype = icp->icp_info;
483 	else {
484 #ifdef ICP_DEBUG
485 		printf("%s: rescan: unit %d ICP_DEVTYPE failed\n",
486 		    device_xname(&icp->icp_dv), unit);
487 #endif
488 		newtype = 0;	/* XXX? */
489 	}
490 
491 #ifdef ICP_DEBUG
492 	printf("%s: rescan: unit %d old %u/%u, new %u/%u\n",
493 	    device_xname(&icp->icp_dv), unit, icp->icp_cdr[unit].cd_size,
494 	    icp->icp_cdr[unit].cd_type, newsize, newtype);
495 #endif
496 
497 	/*
498 	 * If the type or size changed, detach any old child (if it exists)
499 	 * and attach a new one.
500 	 */
501 	if (icp->icp_children[unit] == NULL ||
502 	    newsize != icp->icp_cdr[unit].cd_size ||
503 	    newtype != icp->icp_cdr[unit].cd_type) {
504 		if (icp->icp_cdr[unit].cd_size == 0)
505 			icp->icp_ndevs++;
506 		icp->icp_cdr[unit].cd_size = newsize;
507 		icp->icp_cdr[unit].cd_type = newtype;
508 		if (icp->icp_children[unit] != NULL)
509 			(void) config_detach(icp->icp_children[unit],
510 			    DETACH_FORCE);
511 
512 		icpa.icpa_unit = unit;
513 
514 		locs[ICPCF_UNIT] = unit;
515 
516 		icp->icp_children[unit] = config_found_sm_loc(&icp->icp_dv,
517 			"icp", locs, &icpa, icp_print, config_stdsubmatch);
518 	}
519 
520 	icp_recompute_openings(icp);
521 }
522 
523 void
524 icp_rescan_all(struct icp_softc *icp)
525 {
526 	int unit;
527 	u_int16_t cdev_cnt;
528 
529 	/*
530 	 * This is the old method of rescanning the host drives.  We
531 	 * start by reinitializing the cache service.
532 	 */
533 	if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INIT, ICP_LINUX_OS, 0, 0)) {
534 		printf("%s: unable to re-initialize cache service for rescan\n",
535 		    device_xname(&icp->icp_dv));
536 		return;
537 	}
538 	cdev_cnt = (u_int16_t) icp->icp_info;
539 
540 	/* For each host drive, do the new-style rescan. */
541 	for (unit = 0; unit < cdev_cnt && unit < ICP_MAX_HDRIVES; unit++)
542 		icp_rescan(icp, unit);
543 
544 	/* Now detach anything in the slots after cdev_cnt. */
545 	for (; unit < ICP_MAX_HDRIVES; unit++) {
546 		if (icp->icp_cdr[unit].cd_size != 0) {
547 #ifdef ICP_DEBUG
548 			printf("%s: rescan all: unit %d < new cdev_cnt (%d)\n",
549 			    device_xname(&icp->icp_dv), unit, cdev_cnt);
550 #endif
551 			icp->icp_ndevs--;
552 			icp->icp_cdr[unit].cd_size = 0;
553 			if (icp->icp_children[unit] != NULL) {
554 				(void) config_detach(icp->icp_children[unit],
555 				    DETACH_FORCE);
556 				icp->icp_children[unit] = NULL;
557 			}
558 		}
559 	}
560 
561 	icp_recompute_openings(icp);
562 }
563 
564 void
565 icp_recompute_openings(struct icp_softc *icp)
566 {
567 	int unit, openings;
568 
569 	if (icp->icp_ndevs != 0)
570 		openings =
571 		    (icp->icp_nccbs - ICP_NCCB_RESERVE) / icp->icp_ndevs;
572 	else
573 		openings = 0;
574 	if (openings == icp->icp_openings)
575 		return;
576 	icp->icp_openings = openings;
577 
578 #ifdef ICP_DEBUG
579 	printf("%s: %d device%s, %d openings per device\n",
580 	    device_xname(&icp->icp_dv), icp->icp_ndevs,
581 	    icp->icp_ndevs == 1 ? "" : "s", icp->icp_openings);
582 #endif
583 
584 	for (unit = 0; unit < ICP_MAX_HDRIVES + ICP_MAXBUS; unit++) {
585 		if (icp->icp_children[unit] != NULL)
586 			(*icp->icp_servicecb[unit]->iscb_openings)(
587 			    icp->icp_children[unit], icp->icp_openings);
588 	}
589 }
590 
591 void
592 icp_watchdog(void *cookie)
593 {
594 	struct icp_softc *icp;
595 	int s;
596 
597 	icp = cookie;
598 
599 	s = splbio();
600 	icp_intr(icp);
601 	if (ICP_HAS_WORK(icp))
602 		icp_ccb_enqueue(icp, NULL);
603 	splx(s);
604 
605 	callout_reset(&icp->icp_wdog_callout, hz * ICP_WATCHDOG_FREQ,
606 	    icp_watchdog, icp);
607 }
608 
609 int
610 icp_print(void *aux, const char *pnp)
611 {
612 	struct icp_attach_args *icpa;
613 	const char *str;
614 
615 	icpa = (struct icp_attach_args *)aux;
616 
617 	if (pnp != NULL) {
618 		if (icpa->icpa_unit < ICPA_UNIT_SCSI)
619 			str = "block device";
620 		else
621 			str = "SCSI channel";
622 		aprint_normal("%s at %s", str, pnp);
623 	}
624 	aprint_normal(" unit %d", icpa->icpa_unit);
625 
626 	return (UNCONF);
627 }
628 
629 int
630 icp_async_event(struct icp_softc *icp, int service)
631 {
632 
633 	if (service == ICP_SCREENSERVICE) {
634 		if (icp->icp_status == ICP_S_MSG_REQUEST) {
635 			/* XXX */
636 		}
637 	} else {
638 		if ((icp->icp_fw_vers & 0xff) >= 0x1a) {
639 			icp->icp_evt.size = 0;
640 			icp->icp_evt.eu.async.ionode =
641 			    device_unit(&icp->icp_dv);
642 			icp->icp_evt.eu.async.status = icp->icp_status;
643 			/*
644 			 * Severity and event string are filled in by the
645 			 * hardware interface interrupt handler.
646 			 */
647 			printf("%s: %s\n", device_xname(&icp->icp_dv),
648 			    icp->icp_evt.event_string);
649 		} else {
650 			icp->icp_evt.size = sizeof(icp->icp_evt.eu.async);
651 			icp->icp_evt.eu.async.ionode =
652 			    device_unit(&icp->icp_dv);
653 			icp->icp_evt.eu.async.service = service;
654 			icp->icp_evt.eu.async.status = icp->icp_status;
655 			icp->icp_evt.eu.async.info = icp->icp_info;
656 			/* XXXJRT FIX THIS */
657 			*(u_int32_t *) icp->icp_evt.eu.async.scsi_coord =
658 			    icp->icp_info2;
659 		}
660 		icp_store_event(icp, GDT_ES_ASYNC, service, &icp->icp_evt);
661 	}
662 
663 	return (0);
664 }
665 
666 int
667 icp_intr(void *cookie)
668 {
669 	struct icp_softc *icp;
670 	struct icp_intr_ctx ctx;
671 	struct icp_ccb *ic;
672 
673 	icp = cookie;
674 
675 	ctx.istatus = (*icp->icp_get_status)(icp);
676 	if (!ctx.istatus) {
677 		icp->icp_status = ICP_S_NO_STATUS;
678 		return (0);
679 	}
680 
681 	(*icp->icp_intr)(icp, &ctx);
682 
683 	icp->icp_status = ctx.cmd_status;
684 	icp->icp_service = ctx.service;
685 	icp->icp_info = ctx.info;
686 	icp->icp_info2 = ctx.info2;
687 
688 	switch (ctx.istatus) {
689 	case ICP_ASYNCINDEX:
690 		icp_async_event(icp, ctx.service);
691 		return (1);
692 
693 	case ICP_SPEZINDEX:
694 		aprint_error_dev(&icp->icp_dv, "uninitialized or unknown service (%d/%d)\n",
695 		    ctx.info, ctx.info2);
696 		icp->icp_evt.size = sizeof(icp->icp_evt.eu.driver);
697 		icp->icp_evt.eu.driver.ionode = device_unit(&icp->icp_dv);
698 		icp_store_event(icp, GDT_ES_DRIVER, 4, &icp->icp_evt);
699 		return (1);
700 	}
701 
702 	if ((ctx.istatus - 2) > icp->icp_nccbs)
703 		panic("icp_intr: bad command index returned");
704 
705 	ic = &icp->icp_ccbs[ctx.istatus - 2];
706 	ic->ic_status = icp->icp_status;
707 
708 	if ((ic->ic_flags & IC_ALLOCED) == 0) {
709 		/* XXX ICP's "iir" driver just sends an event here. */
710 		panic("icp_intr: inactive CCB identified");
711 	}
712 
713 	/*
714 	 * Try to protect ourselves from the running command count already
715 	 * being 0 (e.g. if a polled command times out).
716 	 */
717 	KDASSERT(icp->icp_running != 0);
718 	if (--icp->icp_running == 0 &&
719 	    (icp->icp_flags & ICP_F_WAIT_FREEZE) != 0) {
720 		icp->icp_flags &= ~ICP_F_WAIT_FREEZE;
721 		wakeup(&icp->icp_qfreeze);
722 	}
723 
724 	switch (icp->icp_status) {
725 	case ICP_S_BSY:
726 #ifdef ICP_DEBUG
727 		printf("%s: ICP_S_BSY received\n", device_xname(&icp->icp_dv));
728 #endif
729 		if (__predict_false((ic->ic_flags & IC_UCMD) != 0))
730 			SIMPLEQ_INSERT_HEAD(&icp->icp_ucmd_queue, ic, ic_chain);
731 		else
732 			SIMPLEQ_INSERT_HEAD(&icp->icp_ccb_queue, ic, ic_chain);
733 		break;
734 
735 	default:
736 		ic->ic_flags |= IC_COMPLETE;
737 
738 		if ((ic->ic_flags & IC_WAITING) != 0)
739 			wakeup(ic);
740 		else if (ic->ic_intr != NULL)
741 			(*ic->ic_intr)(ic);
742 
743 		if (ICP_HAS_WORK(icp))
744 			icp_ccb_enqueue(icp, NULL);
745 
746 		break;
747 	}
748 
749 	return (1);
750 }
751 
752 struct icp_ucmd_ctx {
753 	gdt_ucmd_t *iu_ucmd;
754 	u_int32_t iu_cnt;
755 };
756 
757 void
758 icp_ucmd_intr(struct icp_ccb *ic)
759 {
760 	struct icp_softc *icp = (void *) ic->ic_dv;
761 	struct icp_ucmd_ctx *iu = ic->ic_context;
762 	gdt_ucmd_t *ucmd = iu->iu_ucmd;
763 
764 	ucmd->status = icp->icp_status;
765 	ucmd->info = icp->icp_info;
766 
767 	if (iu->iu_cnt != 0) {
768 		bus_dmamap_sync(icp->icp_dmat,
769 		    icp->icp_scr_dmamap,
770 		    ICP_SCRATCH_UCMD, iu->iu_cnt,
771 		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
772 		memcpy(ucmd->data,
773 		    (char *)icp->icp_scr + ICP_SCRATCH_UCMD, iu->iu_cnt);
774 	}
775 
776 	icp->icp_ucmd_ccb = NULL;
777 
778 	ic->ic_flags |= IC_COMPLETE;
779 	wakeup(ic);
780 }
781 
782 /*
783  * NOTE: We assume that it is safe to sleep here!
784  */
785 int
786 icp_cmd(struct icp_softc *icp, u_int8_t service, u_int16_t opcode,
787 	u_int32_t arg1, u_int32_t arg2, u_int32_t arg3)
788 {
789 	struct icp_ioctlcmd *icmd;
790 	struct icp_cachecmd *cc;
791 	struct icp_rawcmd *rc;
792 	int retries, rv;
793 	struct icp_ccb *ic;
794 
795 	retries = ICP_RETRIES;
796 
797 	do {
798 		ic = icp_ccb_alloc_wait(icp);
799 		memset(&ic->ic_cmd, 0, sizeof(ic->ic_cmd));
800 		ic->ic_cmd.cmd_opcode = htole16(opcode);
801 
802 		switch (service) {
803 		case ICP_CACHESERVICE:
804 			if (opcode == ICP_IOCTL) {
805 				icmd = &ic->ic_cmd.cmd_packet.ic;
806 				icmd->ic_subfunc = htole16(arg1);
807 				icmd->ic_channel = htole32(arg2);
808 				icmd->ic_bufsize = htole32(arg3);
809 				icmd->ic_addr =
810 				    htole32(icp->icp_scr_seg[0].ds_addr);
811 
812 				bus_dmamap_sync(icp->icp_dmat,
813 				    icp->icp_scr_dmamap, 0, arg3,
814 				    BUS_DMASYNC_PREWRITE |
815 				    BUS_DMASYNC_PREREAD);
816 			} else {
817 				cc = &ic->ic_cmd.cmd_packet.cc;
818 				cc->cc_deviceno = htole16(arg1);
819 				cc->cc_blockno = htole32(arg2);
820 			}
821 			break;
822 
823 		case ICP_SCSIRAWSERVICE:
824 			rc = &ic->ic_cmd.cmd_packet.rc;
825 			rc->rc_direction = htole32(arg1);
826 			rc->rc_bus = arg2;
827 			rc->rc_target = arg3;
828 			rc->rc_lun = arg3 >> 8;
829 			break;
830 		}
831 
832 		ic->ic_service = service;
833 		ic->ic_cmdlen = sizeof(ic->ic_cmd);
834 		rv = icp_ccb_poll(icp, ic, 10000);
835 
836 		switch (service) {
837 		case ICP_CACHESERVICE:
838 			if (opcode == ICP_IOCTL) {
839 				bus_dmamap_sync(icp->icp_dmat,
840 				    icp->icp_scr_dmamap, 0, arg3,
841 				    BUS_DMASYNC_POSTWRITE |
842 				    BUS_DMASYNC_POSTREAD);
843 			}
844 			break;
845 		}
846 
847 		icp_ccb_free(icp, ic);
848 	} while (rv != 0 && --retries > 0);
849 
850 	return (icp->icp_status == ICP_S_OK);
851 }
852 
853 int
854 icp_ucmd(struct icp_softc *icp, gdt_ucmd_t *ucmd)
855 {
856 	struct icp_ccb *ic;
857 	struct icp_ucmd_ctx iu;
858 	u_int32_t cnt;
859 	int error;
860 
861 	if (ucmd->service == ICP_CACHESERVICE) {
862 		if (ucmd->command.cmd_opcode == ICP_IOCTL) {
863 			cnt = ucmd->command.cmd_packet.ic.ic_bufsize;
864 			if (cnt > GDT_SCRATCH_SZ) {
865 				aprint_error_dev(&icp->icp_dv, "scratch buffer too small (%d/%d)\n",
866 				    GDT_SCRATCH_SZ, cnt);
867 				return (EINVAL);
868 			}
869 		} else {
870 			cnt = ucmd->command.cmd_packet.cc.cc_blockcnt *
871 			    ICP_SECTOR_SIZE;
872 			if (cnt > GDT_SCRATCH_SZ) {
873 				aprint_error_dev(&icp->icp_dv, "scratch buffer too small (%d/%d)\n",
874 				    GDT_SCRATCH_SZ, cnt);
875 				return (EINVAL);
876 			}
877 		}
878 	} else {
879 		cnt = ucmd->command.cmd_packet.rc.rc_sdlen +
880 		    ucmd->command.cmd_packet.rc.rc_sense_len;
881 		if (cnt > GDT_SCRATCH_SZ) {
882 			aprint_error_dev(&icp->icp_dv, "scratch buffer too small (%d/%d)\n",
883 			    GDT_SCRATCH_SZ, cnt);
884 			return (EINVAL);
885 		}
886 	}
887 
888 	iu.iu_ucmd = ucmd;
889 	iu.iu_cnt = cnt;
890 
891 	ic = icp_ccb_alloc_wait(icp);
892 	memset(&ic->ic_cmd, 0, sizeof(ic->ic_cmd));
893 	ic->ic_cmd.cmd_opcode = htole16(ucmd->command.cmd_opcode);
894 
895 	if (ucmd->service == ICP_CACHESERVICE) {
896 		if (ucmd->command.cmd_opcode == ICP_IOCTL) {
897 			struct icp_ioctlcmd *icmd, *uicmd;
898 
899 			icmd = &ic->ic_cmd.cmd_packet.ic;
900 			uicmd = &ucmd->command.cmd_packet.ic;
901 
902 			icmd->ic_subfunc = htole16(uicmd->ic_subfunc);
903 			icmd->ic_channel = htole32(uicmd->ic_channel);
904 			icmd->ic_bufsize = htole32(uicmd->ic_bufsize);
905 			icmd->ic_addr =
906 			    htole32(icp->icp_scr_seg[0].ds_addr +
907 				    ICP_SCRATCH_UCMD);
908 		} else {
909 			struct icp_cachecmd *cc, *ucc;
910 
911 			cc = &ic->ic_cmd.cmd_packet.cc;
912 			ucc = &ucmd->command.cmd_packet.cc;
913 
914 			cc->cc_deviceno = htole16(ucc->cc_deviceno);
915 			cc->cc_blockno = htole32(ucc->cc_blockno);
916 			cc->cc_blockcnt = htole32(ucc->cc_blockcnt);
917 			cc->cc_addr = htole32(0xffffffffU);
918 			cc->cc_nsgent = htole32(1);
919 			cc->cc_sg[0].sg_addr =
920 			    htole32(icp->icp_scr_seg[0].ds_addr +
921 				    ICP_SCRATCH_UCMD);
922 			cc->cc_sg[0].sg_len = htole32(cnt);
923 		}
924 	} else {
925 		struct icp_rawcmd *rc, *urc;
926 
927 		rc = &ic->ic_cmd.cmd_packet.rc;
928 		urc = &ucmd->command.cmd_packet.rc;
929 
930 		rc->rc_direction = htole32(urc->rc_direction);
931 		rc->rc_sdata = htole32(0xffffffffU);
932 		rc->rc_sdlen = htole32(urc->rc_sdlen);
933 		rc->rc_clen = htole32(urc->rc_clen);
934 		memcpy(rc->rc_cdb, urc->rc_cdb, sizeof(rc->rc_cdb));
935 		rc->rc_target = urc->rc_target;
936 		rc->rc_lun = urc->rc_lun;
937 		rc->rc_bus = urc->rc_bus;
938 		rc->rc_sense_len = htole32(urc->rc_sense_len);
939 		rc->rc_sense_addr =
940 		    htole32(icp->icp_scr_seg[0].ds_addr +
941 			    ICP_SCRATCH_UCMD + urc->rc_sdlen);
942 		rc->rc_nsgent = htole32(1);
943 		rc->rc_sg[0].sg_addr =
944 		    htole32(icp->icp_scr_seg[0].ds_addr + ICP_SCRATCH_UCMD);
945 		rc->rc_sg[0].sg_len = htole32(cnt - urc->rc_sense_len);
946 	}
947 
948 	ic->ic_service = ucmd->service;
949 	ic->ic_cmdlen = sizeof(ic->ic_cmd);
950 	ic->ic_context = &iu;
951 
952 	/*
953 	 * XXX What units are ucmd->timeout in?  Until we know, we
954 	 * XXX just pull a number out of thin air.
955 	 */
956 	if (__predict_false((error = icp_ccb_wait_user(icp, ic, 30000)) != 0))
957 		aprint_error_dev(&icp->icp_dv, "error %d waiting for ucmd to complete\n",
958 		    error);
959 
960 	/* icp_ucmd_intr() has updated ucmd. */
961 	icp_ccb_free(icp, ic);
962 
963 	return (error);
964 }
965 
966 struct icp_ccb *
967 icp_ccb_alloc(struct icp_softc *icp)
968 {
969 	struct icp_ccb *ic;
970 	int s;
971 
972 	s = splbio();
973 	if (__predict_false((ic =
974 			     SIMPLEQ_FIRST(&icp->icp_ccb_freelist)) == NULL)) {
975 		splx(s);
976 		return (NULL);
977 	}
978 	SIMPLEQ_REMOVE_HEAD(&icp->icp_ccb_freelist, ic_chain);
979 	splx(s);
980 
981 	ic->ic_flags = IC_ALLOCED;
982 	return (ic);
983 }
984 
985 struct icp_ccb *
986 icp_ccb_alloc_wait(struct icp_softc *icp)
987 {
988 	struct icp_ccb *ic;
989 	int s;
990 
991 	s = splbio();
992 	while ((ic = SIMPLEQ_FIRST(&icp->icp_ccb_freelist)) == NULL) {
993 		icp->icp_flags |= ICP_F_WAIT_CCB;
994 		(void) tsleep(&icp->icp_ccb_freelist, PRIBIO, "icpccb", 0);
995 	}
996 	SIMPLEQ_REMOVE_HEAD(&icp->icp_ccb_freelist, ic_chain);
997 	splx(s);
998 
999 	ic->ic_flags = IC_ALLOCED;
1000 	return (ic);
1001 }
1002 
1003 void
1004 icp_ccb_free(struct icp_softc *icp, struct icp_ccb *ic)
1005 {
1006 	int s;
1007 
1008 	s = splbio();
1009 	ic->ic_flags = 0;
1010 	ic->ic_intr = NULL;
1011 	SIMPLEQ_INSERT_HEAD(&icp->icp_ccb_freelist, ic, ic_chain);
1012 	if (__predict_false((icp->icp_flags & ICP_F_WAIT_CCB) != 0)) {
1013 		icp->icp_flags &= ~ICP_F_WAIT_CCB;
1014 		wakeup(&icp->icp_ccb_freelist);
1015 	}
1016 	splx(s);
1017 }
1018 
1019 void
1020 icp_ccb_enqueue(struct icp_softc *icp, struct icp_ccb *ic)
1021 {
1022 	int s;
1023 
1024 	s = splbio();
1025 
1026 	if (ic != NULL) {
1027 		if (__predict_false((ic->ic_flags & IC_UCMD) != 0))
1028 			SIMPLEQ_INSERT_TAIL(&icp->icp_ucmd_queue, ic, ic_chain);
1029 		else
1030 			SIMPLEQ_INSERT_TAIL(&icp->icp_ccb_queue, ic, ic_chain);
1031 	}
1032 
1033 	for (; icp->icp_qfreeze == 0;) {
1034 		if (__predict_false((ic =
1035 			    SIMPLEQ_FIRST(&icp->icp_ucmd_queue)) != NULL)) {
1036 			struct icp_ucmd_ctx *iu = ic->ic_context;
1037 			gdt_ucmd_t *ucmd = iu->iu_ucmd;
1038 
1039 			/*
1040 			 * All user-generated commands share the same
1041 			 * scratch space, so if one is already running,
1042 			 * we have to stall the command queue.
1043 			 */
1044 			if (icp->icp_ucmd_ccb != NULL)
1045 				break;
1046 			if ((*icp->icp_test_busy)(icp))
1047 				break;
1048 			icp->icp_ucmd_ccb = ic;
1049 
1050 			if (iu->iu_cnt != 0) {
1051 				memcpy((char *)icp->icp_scr + ICP_SCRATCH_UCMD,
1052 				    ucmd->data, iu->iu_cnt);
1053 				bus_dmamap_sync(icp->icp_dmat,
1054 				    icp->icp_scr_dmamap,
1055 				    ICP_SCRATCH_UCMD, iu->iu_cnt,
1056 				    BUS_DMASYNC_PREREAD |
1057 				    BUS_DMASYNC_PREWRITE);
1058 			}
1059 		} else if (__predict_true((ic =
1060 				SIMPLEQ_FIRST(&icp->icp_ccb_queue)) != NULL)) {
1061 			if ((*icp->icp_test_busy)(icp))
1062 				break;
1063 		} else {
1064 			/* no command found */
1065 			break;
1066 		}
1067 		icp_ccb_submit(icp, ic);
1068 		if (__predict_false((ic->ic_flags & IC_UCMD) != 0))
1069 			SIMPLEQ_REMOVE_HEAD(&icp->icp_ucmd_queue, ic_chain);
1070 		else
1071 			SIMPLEQ_REMOVE_HEAD(&icp->icp_ccb_queue, ic_chain);
1072 	}
1073 
1074 	splx(s);
1075 }
1076 
1077 int
1078 icp_ccb_map(struct icp_softc *icp, struct icp_ccb *ic, void *data, int size,
1079 	    int dir)
1080 {
1081 	struct icp_sg *sg;
1082 	int nsegs, i, rv;
1083 	bus_dmamap_t xfer;
1084 
1085 	xfer = ic->ic_xfer_map;
1086 
1087 	rv = bus_dmamap_load(icp->icp_dmat, xfer, data, size, NULL,
1088 	    BUS_DMA_NOWAIT | BUS_DMA_STREAMING |
1089 	    ((dir & IC_XFER_IN) ? BUS_DMA_READ : BUS_DMA_WRITE));
1090 	if (rv != 0)
1091 		return (rv);
1092 
1093 	nsegs = xfer->dm_nsegs;
1094 	ic->ic_xfer_size = size;
1095 	ic->ic_nsgent = nsegs;
1096 	ic->ic_flags |= dir;
1097 	sg = ic->ic_sg;
1098 
1099 	if (sg != NULL) {
1100 		for (i = 0; i < nsegs; i++, sg++) {
1101 			sg->sg_addr = htole32(xfer->dm_segs[i].ds_addr);
1102 			sg->sg_len = htole32(xfer->dm_segs[i].ds_len);
1103 		}
1104 	} else if (nsegs > 1)
1105 		panic("icp_ccb_map: no SG list specified, but nsegs > 1");
1106 
1107 	if ((dir & IC_XFER_OUT) != 0)
1108 		i = BUS_DMASYNC_PREWRITE;
1109 	else /* if ((dir & IC_XFER_IN) != 0) */
1110 		i = BUS_DMASYNC_PREREAD;
1111 
1112 	bus_dmamap_sync(icp->icp_dmat, xfer, 0, ic->ic_xfer_size, i);
1113 	return (0);
1114 }
1115 
1116 void
1117 icp_ccb_unmap(struct icp_softc *icp, struct icp_ccb *ic)
1118 {
1119 	int i;
1120 
1121 	if ((ic->ic_flags & IC_XFER_OUT) != 0)
1122 		i = BUS_DMASYNC_POSTWRITE;
1123 	else /* if ((ic->ic_flags & IC_XFER_IN) != 0) */
1124 		i = BUS_DMASYNC_POSTREAD;
1125 
1126 	bus_dmamap_sync(icp->icp_dmat, ic->ic_xfer_map, 0, ic->ic_xfer_size, i);
1127 	bus_dmamap_unload(icp->icp_dmat, ic->ic_xfer_map);
1128 }
1129 
1130 int
1131 icp_ccb_poll(struct icp_softc *icp, struct icp_ccb *ic, int timo)
1132 {
1133 	int s, rv;
1134 
1135 	s = splbio();
1136 
1137 	for (timo = ICP_BUSY_WAIT_MS * 100; timo != 0; timo--) {
1138 		if (!(*icp->icp_test_busy)(icp))
1139 			break;
1140 		DELAY(10);
1141 	}
1142 	if (timo == 0) {
1143 		printf("%s: submit: busy\n", device_xname(&icp->icp_dv));
1144 		return (EAGAIN);
1145 	}
1146 
1147 	icp_ccb_submit(icp, ic);
1148 
1149 	if (cold) {
1150 		for (timo *= 10; timo != 0; timo--) {
1151 			DELAY(100);
1152 			icp_intr(icp);
1153 			if ((ic->ic_flags & IC_COMPLETE) != 0)
1154 				break;
1155 		}
1156 	} else {
1157 		ic->ic_flags |= IC_WAITING;
1158 		while ((ic->ic_flags & IC_COMPLETE) == 0) {
1159 			if ((rv = tsleep(ic, PRIBIO, "icpwccb",
1160 					 mstohz(timo))) != 0) {
1161 				timo = 0;
1162 				break;
1163 			}
1164 		}
1165 	}
1166 
1167 	if (timo != 0) {
1168 		if (ic->ic_status != ICP_S_OK) {
1169 #ifdef ICP_DEBUG
1170 			printf("%s: request failed; status=0x%04x\n",
1171 			    device_xname(&icp->icp_dv), ic->ic_status);
1172 #endif
1173 			rv = EIO;
1174 		} else
1175 			rv = 0;
1176 	} else {
1177 		aprint_error_dev(&icp->icp_dv, "command timed out\n");
1178 		rv = EIO;
1179 	}
1180 
1181 	while ((*icp->icp_test_busy)(icp) != 0)
1182 		DELAY(10);
1183 
1184 	splx(s);
1185 
1186 	return (rv);
1187 }
1188 
1189 int
1190 icp_ccb_wait(struct icp_softc *icp, struct icp_ccb *ic, int timo)
1191 {
1192 	int s, rv;
1193 
1194 	ic->ic_flags |= IC_WAITING;
1195 
1196 	s = splbio();
1197 	icp_ccb_enqueue(icp, ic);
1198 	while ((ic->ic_flags & IC_COMPLETE) == 0) {
1199 		if ((rv = tsleep(ic, PRIBIO, "icpwccb", mstohz(timo))) != 0) {
1200 			splx(s);
1201 			return (rv);
1202 		}
1203 	}
1204 	splx(s);
1205 
1206 	if (ic->ic_status != ICP_S_OK) {
1207 		aprint_error_dev(&icp->icp_dv, "command failed; status=%x\n",
1208 		    ic->ic_status);
1209 		return (EIO);
1210 	}
1211 
1212 	return (0);
1213 }
1214 
1215 int
1216 icp_ccb_wait_user(struct icp_softc *icp, struct icp_ccb *ic, int timo)
1217 {
1218 	int s, rv;
1219 
1220 	ic->ic_dv = &icp->icp_dv;
1221 	ic->ic_intr = icp_ucmd_intr;
1222 	ic->ic_flags |= IC_UCMD;
1223 
1224 	s = splbio();
1225 	icp_ccb_enqueue(icp, ic);
1226 	while ((ic->ic_flags & IC_COMPLETE) == 0) {
1227 		if ((rv = tsleep(ic, PRIBIO, "icpwuccb", mstohz(timo))) != 0) {
1228 			splx(s);
1229 			return (rv);
1230 		}
1231 	}
1232 	splx(s);
1233 
1234 	return (0);
1235 }
1236 
1237 void
1238 icp_ccb_submit(struct icp_softc *icp, struct icp_ccb *ic)
1239 {
1240 
1241 	ic->ic_cmdlen = (ic->ic_cmdlen + 3) & ~3;
1242 
1243 	(*icp->icp_set_sema0)(icp);
1244 	DELAY(10);
1245 
1246 	ic->ic_cmd.cmd_boardnode = htole32(ICP_LOCALBOARD);
1247 	ic->ic_cmd.cmd_cmdindex = htole32(ic->ic_ident);
1248 
1249 	icp->icp_running++;
1250 
1251 	(*icp->icp_copy_cmd)(icp, ic);
1252 	(*icp->icp_release_event)(icp, ic);
1253 }
1254 
1255 int
1256 icp_freeze(struct icp_softc *icp)
1257 {
1258 	int s, error = 0;
1259 
1260 	s = splbio();
1261 	if (icp->icp_qfreeze++ == 0) {
1262 		while (icp->icp_running != 0) {
1263 			icp->icp_flags |= ICP_F_WAIT_FREEZE;
1264 			error = tsleep(&icp->icp_qfreeze, PRIBIO|PCATCH,
1265 			    "icpqfrz", 0);
1266 			if (error != 0 && --icp->icp_qfreeze == 0 &&
1267 			    ICP_HAS_WORK(icp)) {
1268 				icp_ccb_enqueue(icp, NULL);
1269 				break;
1270 			}
1271 		}
1272 	}
1273 	splx(s);
1274 
1275 	return (error);
1276 }
1277 
1278 void
1279 icp_unfreeze(struct icp_softc *icp)
1280 {
1281 	int s;
1282 
1283 	s = splbio();
1284 	KDASSERT(icp->icp_qfreeze != 0);
1285 	if (--icp->icp_qfreeze == 0 && ICP_HAS_WORK(icp))
1286 		icp_ccb_enqueue(icp, NULL);
1287 	splx(s);
1288 }
1289 
1290 /* XXX Global - should be per-controller? XXX */
1291 static gdt_evt_str icp_event_buffer[ICP_MAX_EVENTS];
1292 static int icp_event_oldidx;
1293 static int icp_event_lastidx;
1294 
1295 gdt_evt_str *
1296 icp_store_event(struct icp_softc *icp, u_int16_t source, u_int16_t idx,
1297     gdt_evt_data *evt)
1298 {
1299 	gdt_evt_str *e;
1300 
1301 	/* no source == no event */
1302 	if (source == 0)
1303 		return (NULL);
1304 
1305 	e = &icp_event_buffer[icp_event_lastidx];
1306 	if (e->event_source == source && e->event_idx == idx &&
1307 	    ((evt->size != 0 && e->event_data.size != 0 &&
1308 	      memcmp(&e->event_data.eu, &evt->eu, evt->size) == 0) ||
1309 	     (evt->size == 0 && e->event_data.size == 0 &&
1310 	      strcmp((char *) e->event_data.event_string,
1311 	      	     (char *) evt->event_string) == 0))) {
1312 		e->last_stamp = time_second;
1313 		e->same_count++;
1314 	} else {
1315 		if (icp_event_buffer[icp_event_lastidx].event_source != 0) {
1316 			icp_event_lastidx++;
1317 			if (icp_event_lastidx == ICP_MAX_EVENTS)
1318 				icp_event_lastidx = 0;
1319 			if (icp_event_lastidx == icp_event_oldidx) {
1320 				icp_event_oldidx++;
1321 				if (icp_event_oldidx == ICP_MAX_EVENTS)
1322 					icp_event_oldidx = 0;
1323 			}
1324 		}
1325 		e = &icp_event_buffer[icp_event_lastidx];
1326 		e->event_source = source;
1327 		e->event_idx = idx;
1328 		e->first_stamp = e->last_stamp = time_second;
1329 		e->same_count = 1;
1330 		e->event_data = *evt;
1331 		e->application = 0;
1332 	}
1333 	return (e);
1334 }
1335 
1336 int
1337 icp_read_event(struct icp_softc *icp, int handle, gdt_evt_str *estr)
1338 {
1339 	gdt_evt_str *e;
1340 	int eindex, s;
1341 
1342 	s = splbio();
1343 
1344 	if (handle == -1)
1345 		eindex = icp_event_oldidx;
1346 	else
1347 		eindex = handle;
1348 
1349 	estr->event_source = 0;
1350 
1351 	if (eindex < 0 || eindex >= ICP_MAX_EVENTS) {
1352 		splx(s);
1353 		return (eindex);
1354 	}
1355 
1356 	e = &icp_event_buffer[eindex];
1357 	if (e->event_source != 0) {
1358 		if (eindex != icp_event_lastidx) {
1359 			eindex++;
1360 			if (eindex == ICP_MAX_EVENTS)
1361 				eindex = 0;
1362 		} else
1363 			eindex = -1;
1364 		memcpy(estr, e, sizeof(gdt_evt_str));
1365 	}
1366 
1367 	splx(s);
1368 
1369 	return (eindex);
1370 }
1371 
1372 void
1373 icp_readapp_event(struct icp_softc *icp, u_int8_t application,
1374     gdt_evt_str *estr)
1375 {
1376 	gdt_evt_str *e;
1377 	int found = 0, eindex, s;
1378 
1379 	s = splbio();
1380 
1381 	eindex = icp_event_oldidx;
1382 	for (;;) {
1383 		e = &icp_event_buffer[eindex];
1384 		if (e->event_source == 0)
1385 			break;
1386 		if ((e->application & application) == 0) {
1387 			e->application |= application;
1388 			found = 1;
1389 			break;
1390 		}
1391 		if (eindex == icp_event_lastidx)
1392 			break;
1393 		eindex++;
1394 		if (eindex == ICP_MAX_EVENTS)
1395 			eindex = 0;
1396 	}
1397 	if (found)
1398 		memcpy(estr, e, sizeof(gdt_evt_str));
1399 	else
1400 		estr->event_source = 0;
1401 
1402 	splx(s);
1403 }
1404 
1405 void
1406 icp_clear_events(struct icp_softc *icp)
1407 {
1408 	int s;
1409 
1410 	s = splbio();
1411 	icp_event_oldidx = icp_event_lastidx = 0;
1412 	memset(icp_event_buffer, 0, sizeof(icp_event_buffer));
1413 	splx(s);
1414 }
1415