xref: /netbsd-src/sys/dev/raidframe/rf_driver.c (revision 861ff774d57d6524e280e956bf7bfaaf0149c919)
1 /*	$NetBSD: rf_driver.c,v 1.144 2024/09/19 06:13:03 andvar Exp $	*/
2 /*-
3  * Copyright (c) 1999 The NetBSD Foundation, Inc.
4  * All rights reserved.
5  *
6  * This code is derived from software contributed to The NetBSD Foundation
7  * by Greg Oster
8  *
9  * Redistribution and use in source and binary forms, with or without
10  * modification, are permitted provided that the following conditions
11  * are met:
12  * 1. Redistributions of source code must retain the above copyright
13  *    notice, this list of conditions and the following disclaimer.
14  * 2. Redistributions in binary form must reproduce the above copyright
15  *    notice, this list of conditions and the following disclaimer in the
16  *    documentation and/or other materials provided with the distribution.
17  *
18  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
19  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
20  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
21  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
22  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
23  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
24  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
25  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
26  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
27  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
28  * POSSIBILITY OF SUCH DAMAGE.
29  */
30 
31 /*
32  * Copyright (c) 1995 Carnegie-Mellon University.
33  * All rights reserved.
34  *
35  * Author: Mark Holland, Khalil Amiri, Claudson Bornstein, William V. Courtright II,
36  *         Robby Findler, Daniel Stodolsky, Rachad Youssef, Jim Zelenka
37  *
38  * Permission to use, copy, modify and distribute this software and
39  * its documentation is hereby granted, provided that both the copyright
40  * notice and this permission notice appear in all copies of the
41  * software, derivative works or modified versions, and any portions
42  * thereof, and that both notices appear in supporting documentation.
43  *
44  * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS"
45  * CONDITION.  CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND
46  * FOR ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE.
47  *
48  * Carnegie Mellon requests users of this software to return to
49  *
50  *  Software Distribution Coordinator  or  Software.Distribution@CS.CMU.EDU
51  *  School of Computer Science
52  *  Carnegie Mellon University
53  *  Pittsburgh PA 15213-3890
54  *
55  * any improvements or extensions that they make and grant Carnegie the
56  * rights to redistribute these changes.
57  */
58 
59 /******************************************************************************
60  *
61  * rf_driver.c -- main setup, teardown, and access routines for the RAID driver
62  *
63  * all routines are prefixed with rf_ (raidframe), to avoid conflicts.
64  *
65  ******************************************************************************/
66 
67 
68 #include <sys/cdefs.h>
69 __KERNEL_RCSID(0, "$NetBSD: rf_driver.c,v 1.144 2024/09/19 06:13:03 andvar Exp $");
70 
71 #ifdef _KERNEL_OPT
72 #include "opt_raid_diagnostic.h"
73 #endif
74 
75 #include <sys/param.h>
76 #include <sys/systm.h>
77 #include <sys/ioctl.h>
78 #include <sys/fcntl.h>
79 #include <sys/vnode.h>
80 
81 
82 #include "rf_archs.h"
83 #include "rf_threadstuff.h"
84 
85 #include <sys/errno.h>
86 
87 #include "rf_raid.h"
88 #include "rf_dag.h"
89 #include "rf_aselect.h"
90 #include "rf_diskqueue.h"
91 #include "rf_parityscan.h"
92 #include "rf_alloclist.h"
93 #include "rf_dagutils.h"
94 #include "rf_utils.h"
95 #include "rf_etimer.h"
96 #include "rf_acctrace.h"
97 #include "rf_general.h"
98 #include "rf_desc.h"
99 #include "rf_states.h"
100 #include "rf_decluster.h"
101 #include "rf_map.h"
102 #include "rf_revent.h"
103 #include "rf_callback.h"
104 #include "rf_engine.h"
105 #include "rf_mcpair.h"
106 #include "rf_nwayxor.h"
107 #include "rf_driver.h"
108 #include "rf_options.h"
109 #include "rf_shutdown.h"
110 #include "rf_kintf.h"
111 #include "rf_paritymap.h"
112 
113 #include <sys/buf.h>
114 
115 #ifndef RF_ACCESS_DEBUG
116 #define RF_ACCESS_DEBUG 0
117 #endif
118 
119 /* rad == RF_RaidAccessDesc_t */
120 #define RF_MAX_FREE_RAD 128
121 #define RF_MIN_FREE_RAD  32
122 
123 /* main configuration routines */
124 static int raidframe_booted = 0;
125 
126 static void rf_ConfigureDebug(RF_Config_t * cfgPtr);
127 static void set_debug_option(char *name, long val);
128 static void rf_UnconfigureArray(void);
129 static void rf_ShutdownRDFreeList(void *);
130 static int rf_ConfigureRDFreeList(RF_ShutdownList_t **, RF_Raid_t *, RF_Config_t *);
131 
132 rf_declare_mutex2(rf_printf_mutex);	/* debug only:  avoids interleaved
133 					 * printfs by different stripes */
134 
135 #define SIGNAL_QUIESCENT_COND(_raid_) \
136 	rf_broadcast_cond2((_raid_)->access_suspend_cv)
137 #define WAIT_FOR_QUIESCENCE(_raid_) \
138 	rf_wait_cond2((_raid_)->access_suspend_cv, \
139 		      (_raid_)->access_suspend_mutex)
140 
141 static int configureCount = 0;	/* number of active configurations */
142 static int isconfigged = 0;	/* is basic raidframe (non per-array)
143 				 * stuff configured */
144 static rf_declare_mutex2(configureMutex); /* used to lock the configuration
145 					   * stuff */
146 static RF_ShutdownList_t *globalShutdown;	/* non array-specific
147 						 * stuff */
148 
149 static int rf_ConfigureRDFreeList(RF_ShutdownList_t ** listp, RF_Raid_t *raidPtr, RF_Config_t *cfgPtr);
150 static int rf_AllocEmergBuffers(RF_Raid_t *);
151 static void rf_FreeEmergBuffers(RF_Raid_t *);
152 static void rf_destroy_mutex_cond(RF_Raid_t *);
153 static void rf_alloc_mutex_cond(RF_Raid_t *);
154 
155 /* called at system boot time */
156 int
157 rf_BootRaidframe(bool boot)
158 {
159 
160 	if (boot) {
161 		if (raidframe_booted)
162 			return (EBUSY);
163 		raidframe_booted = 1;
164 		rf_init_mutex2(configureMutex, IPL_NONE);
165  		configureCount = 0;
166 		isconfigged = 0;
167 		globalShutdown = NULL;
168 	} else {
169 		rf_destroy_mutex2(configureMutex);
170 		raidframe_booted = 0;
171 	}
172 	return (0);
173 }
174 
175 /*
176  * Called whenever an array is shutdown
177  */
178 static void
179 rf_UnconfigureArray(void)
180 {
181 
182 	rf_lock_mutex2(configureMutex);
183 	if (--configureCount == 0) {	/* if no active configurations, shut
184 					 * everything down */
185 		rf_destroy_mutex2(rf_printf_mutex);
186 		isconfigged = 0;
187 		rf_ShutdownList(&globalShutdown);
188 
189 		/*
190 	         * We must wait until now, because the AllocList module
191 	         * uses the DebugMem module.
192 	         */
193 #if RF_DEBUG_MEM
194 		if (rf_memDebug)
195 			rf_print_unfreed();
196 #endif
197 	}
198 	rf_unlock_mutex2(configureMutex);
199 }
200 
201 /*
202  * Called to shut down an array.
203  */
204 int
205 rf_Shutdown(RF_Raid_t *raidPtr)
206 {
207 
208 	if (!raidPtr->valid) {
209 		RF_ERRORMSG("Attempt to shut down unconfigured RAIDframe driver.  Aborting shutdown\n");
210 		return (EINVAL);
211 	}
212 	/*
213          * wait for outstanding IOs to land
214          * As described in rf_raid.h, we use the rad_freelist lock
215          * to protect the per-array info about outstanding descs
216          * since we need to do freelist locking anyway, and this
217          * cuts down on the amount of serialization we've got going
218          * on.
219          */
220 	rf_lock_mutex2(raidPtr->rad_lock);
221 	if (raidPtr->waitShutdown) {
222 		rf_unlock_mutex2(raidPtr->rad_lock);
223 		return (EBUSY);
224 	}
225 	raidPtr->waitShutdown = 1;
226 	while (raidPtr->nAccOutstanding) {
227 		rf_wait_cond2(raidPtr->outstandingCond, raidPtr->rad_lock);
228 	}
229 
230 	/* Wait for any parity re-writes to stop... */
231 	while (raidPtr->parity_rewrite_in_progress) {
232 		printf("raid%d: Waiting for parity re-write to exit...\n",
233 		       raidPtr->raidid);
234 		rf_wait_cond2(raidPtr->parity_rewrite_cv, raidPtr->rad_lock);
235 	}
236 	rf_unlock_mutex2(raidPtr->rad_lock);
237 
238 	/* Wait for any reconstruction to stop... */
239 	rf_lock_mutex2(raidPtr->mutex);
240 	while (raidPtr->reconInProgress) {
241 		printf("raid%d: Waiting for reconstruction to stop...\n",
242 		       raidPtr->raidid);
243 		rf_wait_cond2(raidPtr->waitForReconCond, raidPtr->mutex);
244 	}
245 	rf_unlock_mutex2(raidPtr->mutex);
246 
247 	raidPtr->valid = 0;
248 
249 	if (raidPtr->parity_map != NULL)
250 		rf_paritymap_detach(raidPtr);
251 
252 	rf_update_component_labels(raidPtr, RF_FINAL_COMPONENT_UPDATE);
253 
254 	rf_UnconfigureVnodes(raidPtr);
255 
256 	rf_FreeEmergBuffers(raidPtr);
257 
258 	rf_ShutdownList(&raidPtr->shutdownList);
259 
260 	rf_destroy_mutex_cond(raidPtr);
261 
262 	rf_UnconfigureArray();
263 
264 	return (0);
265 }
266 
267 
268 #define DO_INIT_CONFIGURE(f) { \
269 	rc = f (&globalShutdown); \
270 	if (rc) { \
271 		RF_ERRORMSG2("RAIDFRAME: failed %s with %d\n", RF_STRING(f), rc); \
272 		rf_ShutdownList(&globalShutdown); \
273 		configureCount--; \
274 		rf_unlock_mutex2(configureMutex); \
275 		rf_destroy_mutex2(rf_printf_mutex); \
276 		return(rc); \
277 	} \
278 }
279 
280 #define DO_RAID_FAIL() { \
281 	rf_UnconfigureVnodes(raidPtr); \
282 	rf_FreeEmergBuffers(raidPtr); \
283 	rf_ShutdownList(&raidPtr->shutdownList); \
284 	rf_UnconfigureArray(); \
285 	rf_destroy_mutex_cond(raidPtr); \
286 }
287 
288 #define DO_RAID_INIT_CONFIGURE(f) { \
289 	rc = f (&raidPtr->shutdownList, raidPtr, cfgPtr); \
290 	if (rc) { \
291 		RF_ERRORMSG2("RAIDFRAME: failed %s with %d\n", RF_STRING(f), rc); \
292 		DO_RAID_FAIL(); \
293 		return(rc); \
294 	} \
295 }
296 
297 int
298 rf_Configure(RF_Raid_t *raidPtr, RF_Config_t *cfgPtr, RF_AutoConfig_t *ac)
299 {
300 	RF_RowCol_t col;
301 	int rc;
302 	bool swapped = false;
303 	bool first = true;
304 
305 	rf_lock_mutex2(configureMutex);
306 	configureCount++;
307 	if (isconfigged == 0) {
308 		rf_init_mutex2(rf_printf_mutex, IPL_VM);
309 
310 		/* initialize globals */
311 		DO_INIT_CONFIGURE(rf_ConfigureAllocList);
312 
313 		/*
314 	         * Yes, this does make debugging general to the whole
315 	         * system instead of being array specific. Bummer, drag.
316 		 */
317 		rf_ConfigureDebug(cfgPtr);
318 		DO_INIT_CONFIGURE(rf_ConfigureDebugMem);
319 #if RF_ACC_TRACE > 0
320 		DO_INIT_CONFIGURE(rf_ConfigureAccessTrace);
321 #endif
322 		DO_INIT_CONFIGURE(rf_ConfigureNWayXor);
323 		DO_INIT_CONFIGURE(rf_ConfigureDAGFuncs);
324 		isconfigged = 1;
325 	}
326 	rf_unlock_mutex2(configureMutex);
327 
328 	rf_alloc_mutex_cond(raidPtr);
329 
330 	/* set up the cleanup list.  Do this after ConfigureDebug so that
331 	 * value of memDebug will be set */
332 
333 	rf_MakeAllocList(raidPtr->cleanupList);
334 	if (raidPtr->cleanupList == NULL) {
335 		DO_RAID_FAIL();
336 		return (ENOMEM);
337 	}
338 	rf_ShutdownCreate(&raidPtr->shutdownList,
339 			  (void (*) (void *)) rf_FreeAllocList,
340 			  raidPtr->cleanupList);
341 
342 	KASSERT(cfgPtr->numCol < RF_MAXCOL);
343 	KASSERT(cfgPtr->numCol >= 0);
344 	KASSERT(cfgPtr->numSpare < RF_MAXSPARE);
345 	KASSERT(cfgPtr->numSpare >= 0);
346 
347 	raidPtr->numCol = cfgPtr->numCol;
348 	raidPtr->numSpare = cfgPtr->numSpare;
349 	raidPtr->maxQueue = cfgPtr->numSpare;
350 
351 	raidPtr->status = rf_rs_optimal;
352 	raidPtr->reconControl = NULL;
353 
354 	DO_RAID_INIT_CONFIGURE(rf_ConfigureMapModule);
355 	DO_RAID_INIT_CONFIGURE(rf_ConfigureReconEvent);
356 	DO_RAID_INIT_CONFIGURE(rf_ConfigureCallback);
357 	DO_RAID_INIT_CONFIGURE(rf_ConfigureRDFreeList);
358 	DO_RAID_INIT_CONFIGURE(rf_ConfigureStripeLockFreeList);
359 	DO_RAID_INIT_CONFIGURE(rf_ConfigureMCPair);
360 	DO_RAID_INIT_CONFIGURE(rf_ConfigureDAGs);
361 	DO_RAID_INIT_CONFIGURE(rf_ConfigureReconstruction);
362 	DO_RAID_INIT_CONFIGURE(rf_ConfigureDiskQueueSystem);
363 	DO_RAID_INIT_CONFIGURE(rf_ConfigurePSStatus);
364 
365 	DO_RAID_INIT_CONFIGURE(rf_ConfigureEngine);
366 	DO_RAID_INIT_CONFIGURE(rf_ConfigureStripeLocks);
367 
368 	raidPtr->nAccOutstanding = 0;
369 	raidPtr->waitShutdown = 0;
370 
371 	if (ac!=NULL) {
372 		/* We have an AutoConfig structure..  Don't do the
373 		   normal disk configuration... call the auto config
374 		   stuff */
375 		rf_AutoConfigureDisks(raidPtr, cfgPtr, ac);
376 	} else {
377 		DO_RAID_INIT_CONFIGURE(rf_ConfigureDisks);
378 		DO_RAID_INIT_CONFIGURE(rf_ConfigureSpareDisks);
379 	}
380 	/* do this after ConfigureDisks & ConfigureSpareDisks to be sure dev
381 	 * no. is set */
382 	DO_RAID_INIT_CONFIGURE(rf_ConfigureDiskQueues);
383 
384 	DO_RAID_INIT_CONFIGURE(rf_ConfigureLayout);
385 
386 
387 
388 
389 	/* Initialize per-RAID PSS bits */
390 	rf_InitPSStatus(raidPtr);
391 
392 #if RF_INCLUDE_CHAINDECLUSTER > 0
393 	for (col = 0; col < raidPtr->numCol; col++) {
394 		/*
395 		 * XXX better distribution
396 		 */
397 		raidPtr->hist_diskreq[col] = 0;
398 	}
399 #endif
400 	raidPtr->numNewFailures = 0;
401 	raidPtr->parity_rewrite_in_progress = 0;
402 	raidPtr->changing_components = 0;
403 	raidPtr->recon_in_progress = 0;
404 
405 	raidPtr->maxOutstanding = cfgPtr->maxOutstandingDiskReqs;
406 
407 	/* autoconfigure and root_partition will actually get filled in
408 	   after the config is done */
409 	raidPtr->autoconfigure = 0;
410 	raidPtr->root_partition = 0;
411 	raidPtr->last_unit = raidPtr->raidid;
412 	raidPtr->config_order = 0;
413 
414 	if (rf_keepAccTotals) {
415 		raidPtr->keep_acc_totals = 1;
416 	}
417 
418 	/* Allocate a bunch of buffers to be used in low-memory conditions */
419 	raidPtr->iobuf = NULL;
420 
421 	rc = rf_AllocEmergBuffers(raidPtr);
422 	if (rc) {
423 		printf("raid%d: Unable to allocate emergency buffers.\n",
424 		       raidPtr->raidid);
425 		DO_RAID_FAIL();
426 		return(rc);
427 	}
428 
429 	/* Set up parity map stuff, if applicable. */
430 #ifndef RF_NO_PARITY_MAP
431 	rf_paritymap_attach(raidPtr, cfgPtr->force);
432 #endif
433 
434 	raidPtr->valid = 1;
435 
436 	printf("raid%d: %s\n", raidPtr->raidid,
437 	       raidPtr->Layout.map->configName);
438 	printf("raid%d: Components:", raidPtr->raidid);
439 
440 	for (col = 0; col < raidPtr->numCol; col++) {
441 		RF_ComponentLabel_t *clabel;
442 		bool compswapped;
443 
444 		printf(" %s", raidPtr->Disks[col].devname);
445 		if (RF_DEAD_DISK(raidPtr->Disks[col].status)) {
446 			printf("[**FAILED**]");
447 		}
448 		clabel = raidget_component_label(raidPtr, col);
449 		compswapped = clabel->version ==
450 			      bswap32(RF_COMPONENT_LABEL_VERSION);
451 		if (first)
452 			swapped = compswapped;
453 		else if (swapped != compswapped)
454 			printf("raid%d: Component %d has different endian "
455 			       "than first component.", raidPtr->raidid, col);
456 	}
457 	printf("\n");
458 	printf("raid%d: Total Sectors: %" PRIu64 " (%" PRIu64 " MB)\n",
459 	       raidPtr->raidid,
460 	       raidPtr->totalSectors,
461 	       (raidPtr->totalSectors / 1024 *
462 				(1 << raidPtr->logBytesPerSector) / 1024));
463 	if (swapped)
464 		printf("raid%d: Using swapped-endian component labels.\n",
465 		    raidPtr->raidid);
466 
467 	return (0);
468 }
469 
470 
471 /*
472 
473   Routines to allocate and free the "emergency buffers" for a given
474   RAID set.  These emergency buffers will be used when the kernel runs
475   out of kernel memory.
476 
477  */
478 
479 static int
480 rf_AllocEmergBuffers(RF_Raid_t *raidPtr)
481 {
482 	void *tmpbuf;
483 	RF_VoidPointerListElem_t *vple;
484 	int i;
485 
486 	/* XXX next line needs tuning... */
487 	raidPtr->numEmergencyBuffers = 10 * raidPtr->numCol;
488 #if DEBUG
489 	printf("raid%d: allocating %d buffers of %d bytes.\n",
490 	       raidPtr->raidid,
491 	       raidPtr->numEmergencyBuffers,
492 	       (int)(raidPtr->Layout.sectorsPerStripeUnit <<
493 	       raidPtr->logBytesPerSector));
494 #endif
495 	for (i = 0; i < raidPtr->numEmergencyBuffers; i++) {
496 		tmpbuf = malloc( raidPtr->Layout.sectorsPerStripeUnit <<
497 				 raidPtr->logBytesPerSector,
498 				 M_RAIDFRAME, M_WAITOK);
499 		if (tmpbuf) {
500 			vple = rf_AllocVPListElem(raidPtr);
501 			vple->p= tmpbuf;
502 			vple->next = raidPtr->iobuf;
503 			raidPtr->iobuf = vple;
504 			raidPtr->iobuf_count++;
505 		} else {
506 			printf("raid%d: failed to allocate emergency buffer!\n",
507 			       raidPtr->raidid);
508 			return 1;
509 		}
510 	}
511 
512 	/* XXX next line needs tuning too... */
513 	raidPtr->numEmergencyStripeBuffers = 10;
514         for (i = 0; i < raidPtr->numEmergencyStripeBuffers; i++) {
515                 tmpbuf = malloc( raidPtr->numCol * (raidPtr->Layout.sectorsPerStripeUnit <<
516                                  raidPtr->logBytesPerSector),
517                                  M_RAIDFRAME, M_WAITOK);
518                 if (tmpbuf) {
519                         vple = rf_AllocVPListElem(raidPtr);
520                         vple->p= tmpbuf;
521                         vple->next = raidPtr->stripebuf;
522                         raidPtr->stripebuf = vple;
523                         raidPtr->stripebuf_count++;
524                 } else {
525                         printf("raid%d: failed to allocate emergency stripe buffer!\n",
526                                raidPtr->raidid);
527 			return 1;
528                 }
529         }
530 
531 	return (0);
532 }
533 
534 static void
535 rf_FreeEmergBuffers(RF_Raid_t *raidPtr)
536 {
537 	RF_VoidPointerListElem_t *tmp;
538 
539 	/* Free the emergency IO buffers */
540 	while (raidPtr->iobuf != NULL) {
541 		tmp = raidPtr->iobuf;
542 		raidPtr->iobuf = raidPtr->iobuf->next;
543 		free(tmp->p, M_RAIDFRAME);
544 		rf_FreeVPListElem(raidPtr,tmp);
545 	}
546 
547 	/* Free the emergency stripe buffers */
548 	while (raidPtr->stripebuf != NULL) {
549 		tmp = raidPtr->stripebuf;
550 		raidPtr->stripebuf = raidPtr->stripebuf->next;
551 		free(tmp->p, M_RAIDFRAME);
552 		rf_FreeVPListElem(raidPtr, tmp);
553 	}
554 }
555 
556 
557 static void
558 rf_ShutdownRDFreeList(void *arg)
559 {
560 	RF_Raid_t *raidPtr;
561 
562 	raidPtr = (RF_Raid_t *) arg;
563 
564 	pool_destroy(&raidPtr->pools.rad);
565 }
566 
567 static int
568 rf_ConfigureRDFreeList(RF_ShutdownList_t **listp, RF_Raid_t *raidPtr,
569 		       RF_Config_t *cfgPtr)
570 {
571 
572 	rf_pool_init(raidPtr, raidPtr->poolNames.rad, &raidPtr->pools.rad, sizeof(RF_RaidAccessDesc_t),
573 		     "rad", RF_MIN_FREE_RAD, RF_MAX_FREE_RAD);
574 	rf_ShutdownCreate(listp, rf_ShutdownRDFreeList, raidPtr);
575 	return (0);
576 }
577 
578 RF_RaidAccessDesc_t *
579 rf_AllocRaidAccDesc(RF_Raid_t *raidPtr, RF_IoType_t type,
580 		    RF_RaidAddr_t raidAddress, RF_SectorCount_t numBlocks,
581 		    void *bufPtr, void *bp, RF_RaidAccessFlags_t flags,
582 		    const RF_AccessState_t *states)
583 {
584 	RF_RaidAccessDesc_t *desc;
585 
586 	desc = pool_get(&raidPtr->pools.rad, PR_WAITOK);
587 
588 	rf_lock_mutex2(raidPtr->rad_lock);
589 	if (raidPtr->waitShutdown) {
590 		/*
591 	         * Actually, we're shutting the array down. Free the desc
592 	         * and return NULL.
593 	         */
594 
595 		rf_unlock_mutex2(raidPtr->rad_lock);
596 		pool_put(&raidPtr->pools.rad, desc);
597 		return (NULL);
598 	}
599 	raidPtr->nAccOutstanding++;
600 
601 	rf_unlock_mutex2(raidPtr->rad_lock);
602 
603 	desc->raidPtr = (void *) raidPtr;
604 	desc->type = type;
605 	desc->raidAddress = raidAddress;
606 	desc->numBlocks = numBlocks;
607 	desc->bufPtr = bufPtr;
608 	desc->bp = bp;
609 	desc->flags = flags;
610 	desc->states = states;
611 	desc->state = 0;
612 	desc->dagList = NULL;
613 
614 	desc->status = 0;
615 	desc->numRetries = 0;
616 #if RF_ACC_TRACE > 0
617 	memset(&desc->tracerec, 0, sizeof(desc->tracerec));
618 #endif
619 	desc->callbackFunc = NULL;
620 	desc->callbackArg = NULL;
621 	desc->next = NULL;
622 	desc->iobufs = NULL;
623 	desc->stripebufs = NULL;
624 
625 	return (desc);
626 }
627 
628 void
629 rf_FreeRaidAccDesc(RF_RaidAccessDesc_t *desc)
630 {
631 	RF_Raid_t *raidPtr = desc->raidPtr;
632 	RF_DagList_t *dagList, *temp;
633 	RF_VoidPointerListElem_t *tmp;
634 
635 	RF_ASSERT(desc);
636 
637 	/* Cleanup the dagList(s) */
638 	dagList = desc->dagList;
639 	while(dagList != NULL) {
640 		temp = dagList;
641 		dagList = dagList->next;
642 		rf_FreeDAGList(raidPtr, temp);
643 	}
644 
645 	while (desc->iobufs) {
646 		tmp = desc->iobufs;
647 		desc->iobufs = desc->iobufs->next;
648 		rf_FreeIOBuffer(raidPtr, tmp);
649 	}
650 
651 	while (desc->stripebufs) {
652 		tmp = desc->stripebufs;
653 		desc->stripebufs = desc->stripebufs->next;
654 		rf_FreeStripeBuffer(raidPtr, tmp);
655 	}
656 
657 	pool_put(&raidPtr->pools.rad, desc);
658 	rf_lock_mutex2(raidPtr->rad_lock);
659 	raidPtr->nAccOutstanding--;
660 	if (raidPtr->waitShutdown) {
661 		rf_signal_cond2(raidPtr->outstandingCond);
662 	}
663 	rf_unlock_mutex2(raidPtr->rad_lock);
664 }
665 /*********************************************************************
666  * Main routine for performing an access.
667  * Accesses are retried until a DAG can not be selected.  This occurs
668  * when either the DAG library is incomplete or there are too many
669  * failures in a parity group.
670  *
671  * type should be read or write.  bp_in is a buf pointer.  void *to
672  * facilitate ignoring it outside the kernel
673  ********************************************************************/
674 int
675 rf_DoAccess(RF_Raid_t * raidPtr, RF_IoType_t type, RF_RaidAddr_t raidAddress, RF_SectorCount_t numBlocks,
676 	    void *bufPtr, struct buf *bp, RF_RaidAccessFlags_t flags)
677 {
678 	RF_RaidAccessDesc_t *desc;
679 	void *lbufPtr = bufPtr;
680 
681 	raidAddress += rf_raidSectorOffset;
682 
683 #if RF_ACCESS_DEBUG
684 	if (rf_accessDebug) {
685 
686 		printf("logBytes is: %d %d %d\n", raidPtr->raidid,
687 		    raidPtr->logBytesPerSector,
688 		    (int) rf_RaidAddressToByte(raidPtr, numBlocks));
689 		printf("raid%d: %s raidAddr %d (stripeid %d-%d) numBlocks %d (%d bytes) buf 0x%lx\n", raidPtr->raidid,
690 		    (type == RF_IO_TYPE_READ) ? "READ" : "WRITE", (int) raidAddress,
691 		    (int) rf_RaidAddressToStripeID(&raidPtr->Layout, raidAddress),
692 		    (int) rf_RaidAddressToStripeID(&raidPtr->Layout, raidAddress + numBlocks - 1),
693 		    (int) numBlocks,
694 		    (int) rf_RaidAddressToByte(raidPtr, numBlocks),
695 		    (long) bufPtr);
696 	}
697 #endif
698 
699 	desc = rf_AllocRaidAccDesc(raidPtr, type, raidAddress,
700 	    numBlocks, lbufPtr, bp, flags, raidPtr->Layout.map->states);
701 
702 	if (desc == NULL) {
703 		return (ENOMEM);
704 	}
705 #if RF_ACC_TRACE > 0
706 	RF_ETIMER_START(desc->tracerec.tot_timer);
707 #endif
708 
709 	if (raidPtr->parity_map != NULL &&
710 	    type == RF_IO_TYPE_WRITE)
711 		rf_paritymap_begin(raidPtr->parity_map, raidAddress,
712 		    numBlocks);
713 
714 	rf_ContinueRaidAccess(desc);
715 
716 	return (0);
717 }
718 #if 0
719 /* force the array into reconfigured mode without doing reconstruction */
720 int
721 rf_SetReconfiguredMode(RF_Raid_t *raidPtr, int col)
722 {
723 	if (!(raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE)) {
724 		printf("Can't set reconfigured mode in dedicated-spare array\n");
725 		RF_PANIC();
726 	}
727 	rf_lock_mutex2(raidPtr->mutex);
728 	raidPtr->numFailures++;
729 	raidPtr->Disks[col].status = rf_ds_dist_spared;
730 	raidPtr->status = rf_rs_reconfigured;
731 	rf_update_component_labels(raidPtr, RF_NORMAL_COMPONENT_UPDATE);
732 	/* install spare table only if declustering + distributed sparing
733 	 * architecture. */
734 	if (raidPtr->Layout.map->flags & RF_BD_DECLUSTERED)
735 		rf_InstallSpareTable(raidPtr, col);
736 	rf_unlock_mutex2(raidPtr->mutex);
737 	return (0);
738 }
739 #endif
740 
741 int
742 rf_FailDisk(RF_Raid_t *raidPtr, int fcol, int initRecon)
743 {
744 
745 	/* need to suspend IO's here -- if there are DAGs in flight
746 	   and we pull the rug out from under ci_vp, Bad Things
747 	   can happen.  */
748 
749 	rf_SuspendNewRequestsAndWait(raidPtr);
750 
751 	rf_lock_mutex2(raidPtr->mutex);
752 	if (raidPtr->Disks[fcol].status != rf_ds_failed) {
753 		/* must be failing something that is valid, or else it's
754 		   already marked as failed (in which case we don't
755 		   want to mark it failed again!) */
756 		raidPtr->numFailures++;
757 		raidPtr->Disks[fcol].status = rf_ds_failed;
758 		raidPtr->status = rf_rs_degraded;
759 	}
760 	rf_unlock_mutex2(raidPtr->mutex);
761 
762 	rf_update_component_labels(raidPtr, RF_NORMAL_COMPONENT_UPDATE);
763 
764 	/* Close the component, so that it's not "locked" if someone
765 	   else want's to use it! */
766 
767 	rf_close_component(raidPtr, raidPtr->raid_cinfo[fcol].ci_vp,
768 			   raidPtr->Disks[fcol].auto_configured);
769 
770 	rf_lock_mutex2(raidPtr->mutex);
771 	raidPtr->raid_cinfo[fcol].ci_vp = NULL;
772 
773 	/* Need to mark the component as not being auto_configured
774 	   (in case it was previously). */
775 
776 	raidPtr->Disks[fcol].auto_configured = 0;
777 	rf_unlock_mutex2(raidPtr->mutex);
778 	/* now we can allow IO to continue -- we'll be suspending it
779 	   again in rf_ReconstructFailedDisk() if we have to.. */
780 
781 	rf_ResumeNewRequests(raidPtr);
782 
783 	if (initRecon)
784 		rf_ReconstructFailedDisk(raidPtr, fcol);
785 	return (0);
786 }
787 /* releases a thread that is waiting for the array to become quiesced.
788  * access_suspend_mutex should be locked upon calling this
789  */
790 void
791 rf_SignalQuiescenceLock(RF_Raid_t *raidPtr)
792 {
793 #if RF_DEBUG_QUIESCE
794 	if (rf_quiesceDebug) {
795 		printf("raid%d: Signalling quiescence lock\n",
796 		       raidPtr->raidid);
797 	}
798 #endif
799 	raidPtr->access_suspend_release = 1;
800 
801 	if (raidPtr->waiting_for_quiescence) {
802 		SIGNAL_QUIESCENT_COND(raidPtr);
803 	}
804 }
805 /* suspends all new requests to the array.  No effect on accesses that are in flight.  */
806 int
807 rf_SuspendNewRequestsAndWait(RF_Raid_t *raidPtr)
808 {
809 #if RF_DEBUG_QUIESCE
810 	if (rf_quiesceDebug)
811 		printf("raid%d: Suspending new reqs\n", raidPtr->raidid);
812 #endif
813 	rf_lock_mutex2(raidPtr->access_suspend_mutex);
814 	raidPtr->accesses_suspended++;
815 	raidPtr->waiting_for_quiescence = (raidPtr->accs_in_flight == 0) ? 0 : 1;
816 
817 	if (raidPtr->waiting_for_quiescence) {
818 		raidPtr->access_suspend_release = 0;
819 		while (!raidPtr->access_suspend_release) {
820 #if RF_DEBUG_QUIESCE
821 			printf("raid%d: Suspending: Waiting for Quiescence\n",
822 			       raidPtr->raidid);
823 #endif
824 			WAIT_FOR_QUIESCENCE(raidPtr);
825 			raidPtr->waiting_for_quiescence = 0;
826 		}
827 	}
828 #if RF_DEBUG_QUIESCE
829 	printf("raid%d: Quiescence reached..\n", raidPtr->raidid);
830 #endif
831 
832 	rf_unlock_mutex2(raidPtr->access_suspend_mutex);
833 	return (raidPtr->waiting_for_quiescence);
834 }
835 /* wake up everyone waiting for quiescence to be released */
836 void
837 rf_ResumeNewRequests(RF_Raid_t *raidPtr)
838 {
839 	RF_CallbackFuncDesc_t *t, *cb;
840 
841 #if RF_DEBUG_QUIESCE
842 	if (rf_quiesceDebug)
843 		printf("raid%d: Resuming new requests\n", raidPtr->raidid);
844 #endif
845 
846 	rf_lock_mutex2(raidPtr->access_suspend_mutex);
847 	raidPtr->accesses_suspended--;
848 	if (raidPtr->accesses_suspended == 0)
849 		cb = raidPtr->quiesce_wait_list;
850 	else
851 		cb = NULL;
852 	raidPtr->quiesce_wait_list = NULL;
853 	rf_unlock_mutex2(raidPtr->access_suspend_mutex);
854 
855 	while (cb) {
856 		t = cb;
857 		cb = cb->next;
858 		(t->callbackFunc) (t->callbackArg);
859 		rf_FreeCallbackFuncDesc(raidPtr, t);
860 	}
861 }
862 /*****************************************************************************************
863  *
864  * debug routines
865  *
866  ****************************************************************************************/
867 
868 static void
869 set_debug_option(char *name, long val)
870 {
871 	RF_DebugName_t *p;
872 
873 	for (p = rf_debugNames; p->name; p++) {
874 		if (!strcmp(p->name, name)) {
875 			*(p->ptr) = val;
876 			printf("[Set debug variable %s to %ld]\n", name, val);
877 			return;
878 		}
879 	}
880 	RF_ERRORMSG1("Unknown debug string \"%s\"\n", name);
881 }
882 
883 
884 /* would like to use sscanf here, but apparently not available in kernel */
885 /*ARGSUSED*/
886 static void
887 rf_ConfigureDebug(RF_Config_t *cfgPtr)
888 {
889 	char   *val_p, *name_p, *white_p;
890 	long    val;
891 	int     i;
892 
893 	rf_ResetDebugOptions();
894 	for (i = 0; i < RF_MAXDBGV && cfgPtr->debugVars[i][0]; i++) {
895 		name_p = rf_find_non_white(&cfgPtr->debugVars[i][0]);
896 		white_p = rf_find_white(name_p);	/* skip to start of 2nd
897 							 * word */
898 		val_p = rf_find_non_white(white_p);
899 		if (*val_p == '0' && *(val_p + 1) == 'x')
900 			val = rf_htoi(val_p + 2);
901 		else
902 			val = rf_atoi(val_p);
903 		*white_p = '\0';
904 		set_debug_option(name_p, val);
905 	}
906 }
907 
908 void
909 rf_print_panic_message(int line, const char *file)
910 {
911 	kern_assert("raidframe error at line %d file %s", line, file);
912 }
913 
914 #ifdef RAID_DIAGNOSTIC
915 void
916 rf_print_assert_panic_message(int line,	const char *file, const char *condition)
917 {
918 	kern_assert("raidframe error at line %d file %s (failed asserting %s)\n",
919 	    line, file, condition);
920 }
921 #endif
922 
923 void
924 rf_print_unable_to_init_mutex(const char *file, int line, int rc)
925 {
926 	RF_ERRORMSG3("Unable to init mutex file %s line %d rc=%d\n",
927 		     file, line, rc);
928 }
929 
930 void
931 rf_print_unable_to_add_shutdown(const char *file, int line, int rc)
932 {
933 	RF_ERRORMSG3("Unable to add to shutdown list file %s line %d rc=%d\n",
934 		     file, line, rc);
935 }
936 
937 static void
938 rf_alloc_mutex_cond(RF_Raid_t *raidPtr)
939 {
940 
941 	rf_init_mutex2(raidPtr->mutex, IPL_VM);
942 
943 	rf_init_cond2(raidPtr->outstandingCond, "rfocond");
944 	rf_init_cond2(raidPtr->parity_rewrite_cv, "rfprwshutdown");
945 	rf_init_mutex2(raidPtr->rad_lock, IPL_VM);
946 
947 	rf_init_mutex2(raidPtr->access_suspend_mutex, IPL_VM);
948 	rf_init_cond2(raidPtr->access_suspend_cv, "rfquiesce");
949 
950 	rf_init_cond2(raidPtr->waitForReconCond, "rfrcnw");
951 
952 	rf_init_cond2(raidPtr->changing_components_cv, "rfcc");
953 }
954 
955 static void
956 rf_destroy_mutex_cond(RF_Raid_t *raidPtr)
957 {
958 
959 	rf_destroy_cond2(raidPtr->waitForReconCond);
960 	rf_destroy_cond2(raidPtr->changing_components_cv);
961 
962 	rf_destroy_mutex2(raidPtr->access_suspend_mutex);
963 	rf_destroy_cond2(raidPtr->access_suspend_cv);
964 
965 	rf_destroy_cond2(raidPtr->parity_rewrite_cv);
966 	rf_destroy_cond2(raidPtr->outstandingCond);
967 	rf_destroy_mutex2(raidPtr->rad_lock);
968 
969 	rf_destroy_mutex2(raidPtr->mutex);
970 }
971