xref: /netbsd-src/sys/dev/dksubr.c (revision 82d56013d7b633d116a93943de88e08335357a7c)
1 /* $NetBSD: dksubr.c,v 1.113 2021/04/15 00:32:50 rin Exp $ */
2 
3 /*-
4  * Copyright (c) 1996, 1997, 1998, 1999, 2002, 2008 The NetBSD Foundation, Inc.
5  * All rights reserved.
6  *
7  * This code is derived from software contributed to The NetBSD Foundation
8  * by Jason R. Thorpe and Roland C. Dowdeswell.
9  *
10  * Redistribution and use in source and binary forms, with or without
11  * modification, are permitted provided that the following conditions
12  * are met:
13  * 1. Redistributions of source code must retain the above copyright
14  *    notice, this list of conditions and the following disclaimer.
15  * 2. Redistributions in binary form must reproduce the above copyright
16  *    notice, this list of conditions and the following disclaimer in the
17  *    documentation and/or other materials provided with the distribution.
18  *
19  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
20  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
21  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
22  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
23  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29  * POSSIBILITY OF SUCH DAMAGE.
30  */
31 
32 #include <sys/cdefs.h>
33 __KERNEL_RCSID(0, "$NetBSD: dksubr.c,v 1.113 2021/04/15 00:32:50 rin Exp $");
34 
35 #include <sys/param.h>
36 #include <sys/systm.h>
37 #include <sys/stat.h>
38 #include <sys/proc.h>
39 #include <sys/ioctl.h>
40 #include <sys/device.h>
41 #include <sys/disk.h>
42 #include <sys/disklabel.h>
43 #include <sys/buf.h>
44 #include <sys/bufq.h>
45 #include <sys/vnode.h>
46 #include <sys/fcntl.h>
47 #include <sys/namei.h>
48 #include <sys/module.h>
49 #include <sys/syslog.h>
50 
51 #include <dev/dkvar.h>
52 #include <miscfs/specfs/specdev.h> /* for v_rdev */
53 
54 int	dkdebug = 0;
55 
56 #ifdef DEBUG
57 #define DKDB_FOLLOW	0x1
58 #define DKDB_INIT	0x2
59 #define DKDB_VNODE	0x4
60 #define DKDB_DUMP	0x8
61 
62 #define IFDEBUG(x,y)		if (dkdebug & (x)) y
63 #define DPRINTF(x,y)		IFDEBUG(x, printf y)
64 #define DPRINTF_FOLLOW(y)	DPRINTF(DKDB_FOLLOW, y)
65 #else
66 #define IFDEBUG(x,y)
67 #define DPRINTF(x,y)
68 #define DPRINTF_FOLLOW(y)
69 #endif
70 
71 #define DKF_READYFORDUMP	(DKF_INITED|DKF_TAKEDUMP)
72 
73 static int dk_subr_modcmd(modcmd_t, void *);
74 
75 #define DKLABELDEV(dev)	\
76 	(MAKEDISKDEV(major((dev)), DISKUNIT((dev)), RAW_PART))
77 
78 static void	dk_makedisklabel(struct dk_softc *);
79 static int	dk_translate(struct dk_softc *, struct buf *);
80 static void	dk_done1(struct dk_softc *, struct buf *, bool);
81 
82 void
83 dk_init(struct dk_softc *dksc, device_t dev, int dtype)
84 {
85 
86 	memset(dksc, 0x0, sizeof(*dksc));
87 	dksc->sc_dtype = dtype;
88 	dksc->sc_dev = dev;
89 
90 	strlcpy(dksc->sc_xname, device_xname(dev), DK_XNAME_SIZE);
91 	dksc->sc_dkdev.dk_name = dksc->sc_xname;
92 }
93 
94 void
95 dk_attach(struct dk_softc *dksc)
96 {
97 	KASSERT(dksc->sc_dev != NULL);
98 
99 	mutex_init(&dksc->sc_iolock, MUTEX_DEFAULT, IPL_VM);
100 	dksc->sc_flags |= DKF_READYFORDUMP;
101 #ifdef DIAGNOSTIC
102 	dksc->sc_flags |= DKF_WARNLABEL | DKF_LABELSANITY;
103 #endif
104 
105 	if ((dksc->sc_flags & DKF_NO_RND) == 0) {
106 		/* Attach the device into the rnd source list. */
107 		rnd_attach_source(&dksc->sc_rnd_source, dksc->sc_xname,
108 		    RND_TYPE_DISK, RND_FLAG_DEFAULT);
109 	}
110 }
111 
112 void
113 dk_detach(struct dk_softc *dksc)
114 {
115 	if ((dksc->sc_flags & DKF_NO_RND) == 0) {
116 		/* Unhook the entropy source. */
117 		rnd_detach_source(&dksc->sc_rnd_source);
118 	}
119 
120 	dksc->sc_flags &= ~DKF_READYFORDUMP;
121 	mutex_destroy(&dksc->sc_iolock);
122 }
123 
124 /* ARGSUSED */
125 int
126 dk_open(struct dk_softc *dksc, dev_t dev,
127     int flags, int fmt, struct lwp *l)
128 {
129 	const struct dkdriver *dkd = dksc->sc_dkdev.dk_driver;
130 	struct	disklabel *lp = dksc->sc_dkdev.dk_label;
131 	int	part = DISKPART(dev);
132 	int	pmask = 1 << part;
133 	int	ret = 0;
134 	struct disk *dk = &dksc->sc_dkdev;
135 
136 	DPRINTF_FOLLOW(("%s(%s, %p, 0x%"PRIx64", 0x%x)\n", __func__,
137 	    dksc->sc_xname, dksc, dev, flags));
138 
139 	mutex_enter(&dk->dk_openlock);
140 
141 	/*
142 	 * If there are wedges, and this is not RAW_PART, then we
143 	 * need to fail.
144 	 */
145 	if (dk->dk_nwedges != 0 && part != RAW_PART) {
146 		ret = EBUSY;
147 		goto done;
148 	}
149 
150 	/* If no dkdriver attached, bail */
151 	if (dkd == NULL) {
152 		ret = ENXIO;
153 		goto done;
154 	}
155 
156 	/*
157 	 * initialize driver for the first opener
158 	 */
159 	if (dk->dk_openmask == 0 && dkd->d_firstopen != NULL) {
160 		ret = (*dkd->d_firstopen)(dksc->sc_dev, dev, flags, fmt);
161 		if (ret)
162 			goto done;
163 	}
164 
165 	/*
166 	 * If we're init'ed and there are no other open partitions then
167 	 * update the in-core disklabel.
168 	 */
169 	if ((dksc->sc_flags & DKF_INITED)) {
170 		if ((dksc->sc_flags & DKF_VLABEL) == 0) {
171 			dksc->sc_flags |= DKF_VLABEL;
172 			dk_getdisklabel(dksc, dev);
173 		}
174 	}
175 
176 	/* Fail if we can't find the partition. */
177 	if (part != RAW_PART &&
178 	    ((dksc->sc_flags & DKF_VLABEL) == 0 ||
179 	     part >= lp->d_npartitions ||
180 	     lp->d_partitions[part].p_fstype == FS_UNUSED)) {
181 		ret = ENXIO;
182 		goto done;
183 	}
184 
185 	/* Mark our unit as open. */
186 	switch (fmt) {
187 	case S_IFCHR:
188 		dk->dk_copenmask |= pmask;
189 		break;
190 	case S_IFBLK:
191 		dk->dk_bopenmask |= pmask;
192 		break;
193 	}
194 
195 	dk->dk_openmask = dk->dk_copenmask | dk->dk_bopenmask;
196 
197 done:
198 	mutex_exit(&dk->dk_openlock);
199 	return ret;
200 }
201 
202 /* ARGSUSED */
203 int
204 dk_close(struct dk_softc *dksc, dev_t dev,
205     int flags, int fmt, struct lwp *l)
206 {
207 	const struct dkdriver *dkd = dksc->sc_dkdev.dk_driver;
208 	int	part = DISKPART(dev);
209 	int	pmask = 1 << part;
210 	struct disk *dk = &dksc->sc_dkdev;
211 
212 	DPRINTF_FOLLOW(("%s(%s, %p, 0x%"PRIx64", 0x%x)\n", __func__,
213 	    dksc->sc_xname, dksc, dev, flags));
214 
215 	mutex_enter(&dk->dk_openlock);
216 
217 	switch (fmt) {
218 	case S_IFCHR:
219 		dk->dk_copenmask &= ~pmask;
220 		break;
221 	case S_IFBLK:
222 		dk->dk_bopenmask &= ~pmask;
223 		break;
224 	}
225 	dk->dk_openmask = dk->dk_copenmask | dk->dk_bopenmask;
226 
227 	if (dk->dk_openmask == 0) {
228 		if (dkd->d_lastclose != NULL)
229 			(*dkd->d_lastclose)(dksc->sc_dev);
230 		if ((dksc->sc_flags & DKF_KLABEL) == 0)
231 			dksc->sc_flags &= ~DKF_VLABEL;
232 	}
233 
234 	mutex_exit(&dk->dk_openlock);
235 	return 0;
236 }
237 
238 static int
239 dk_translate(struct dk_softc *dksc, struct buf *bp)
240 {
241 	int	part;
242 	int	wlabel;
243 	daddr_t	blkno;
244 	struct disklabel *lp;
245 	struct disk *dk;
246 	uint64_t numsecs;
247 	unsigned secsize;
248 
249 	lp = dksc->sc_dkdev.dk_label;
250 	dk = &dksc->sc_dkdev;
251 
252 	part = DISKPART(bp->b_dev);
253 	numsecs = dk->dk_geom.dg_secperunit;
254 	secsize = dk->dk_geom.dg_secsize;
255 
256 	/*
257 	 * The transfer must be a whole number of blocks and the offset must
258 	 * not be negative.
259 	 */
260 	if ((bp->b_bcount % secsize) != 0 || bp->b_blkno < 0) {
261 		bp->b_error = EINVAL;
262 		goto done;
263 	}
264 
265 	/* If there is nothing to do, then we are done */
266 	if (bp->b_bcount == 0)
267 		goto done;
268 
269 	wlabel = dksc->sc_flags & (DKF_WLABEL|DKF_LABELLING);
270 	if (part == RAW_PART) {
271 		uint64_t numblocks = btodb(numsecs * secsize);
272 		if (bounds_check_with_mediasize(bp, DEV_BSIZE, numblocks) <= 0)
273 			goto done;
274 	} else {
275 		if (bounds_check_with_label(&dksc->sc_dkdev, bp, wlabel) <= 0)
276 			goto done;
277 	}
278 
279 	/*
280 	 * Convert the block number to absolute and put it in terms
281 	 * of the device's logical block size.
282 	 */
283 	if (secsize >= DEV_BSIZE)
284 		blkno = bp->b_blkno / (secsize / DEV_BSIZE);
285 	else
286 		blkno = bp->b_blkno * (DEV_BSIZE / secsize);
287 
288 	if (part != RAW_PART)
289 		blkno += lp->d_partitions[DISKPART(bp->b_dev)].p_offset;
290 	bp->b_rawblkno = blkno;
291 
292 	return -1;
293 
294 done:
295 	bp->b_resid = bp->b_bcount;
296 	return bp->b_error;
297 }
298 
299 static int
300 dk_strategy1(struct dk_softc *dksc, struct buf *bp)
301 {
302 	int error;
303 
304 	DPRINTF_FOLLOW(("%s(%s, %p, %p)\n", __func__,
305 	    dksc->sc_xname, dksc, bp));
306 
307 	if (!(dksc->sc_flags & DKF_INITED)) {
308 		DPRINTF_FOLLOW(("%s: not inited\n", __func__));
309 		bp->b_error = ENXIO;
310 		bp->b_resid = bp->b_bcount;
311 		biodone(bp);
312 		return 1;
313 	}
314 
315 	error = dk_translate(dksc, bp);
316 	if (error >= 0) {
317 		biodone(bp);
318 		return 1;
319 	}
320 
321 	return 0;
322 }
323 
324 void
325 dk_strategy(struct dk_softc *dksc, struct buf *bp)
326 {
327 	int error;
328 
329 	error = dk_strategy1(dksc, bp);
330 	if (error)
331 		return;
332 
333 	/*
334 	 * Queue buffer and start unit
335 	 */
336 	dk_start(dksc, bp);
337 }
338 
339 int
340 dk_strategy_defer(struct dk_softc *dksc, struct buf *bp)
341 {
342 	int error;
343 
344 	error = dk_strategy1(dksc, bp);
345 	if (error)
346 		return error;
347 
348 	/*
349 	 * Queue buffer only
350 	 */
351 	mutex_enter(&dksc->sc_iolock);
352 	disk_wait(&dksc->sc_dkdev);
353 	bufq_put(dksc->sc_bufq, bp);
354 	mutex_exit(&dksc->sc_iolock);
355 
356 	return 0;
357 }
358 
359 int
360 dk_strategy_pending(struct dk_softc *dksc)
361 {
362 	struct buf *bp;
363 
364 	if (!(dksc->sc_flags & DKF_INITED)) {
365 		DPRINTF_FOLLOW(("%s: not inited\n", __func__));
366 		return 0;
367 	}
368 
369 	mutex_enter(&dksc->sc_iolock);
370 	bp = bufq_peek(dksc->sc_bufq);
371 	mutex_exit(&dksc->sc_iolock);
372 
373 	return bp != NULL;
374 }
375 
376 void
377 dk_start(struct dk_softc *dksc, struct buf *bp)
378 {
379 	const struct dkdriver *dkd = dksc->sc_dkdev.dk_driver;
380 	int error;
381 
382 	if (!(dksc->sc_flags & DKF_INITED)) {
383 		DPRINTF_FOLLOW(("%s: not inited\n", __func__));
384 		return;
385 	}
386 
387 	mutex_enter(&dksc->sc_iolock);
388 
389 	if (bp != NULL) {
390 		bp->b_ci = curcpu();
391 		disk_wait(&dksc->sc_dkdev);
392 		bufq_put(dksc->sc_bufq, bp);
393 	}
394 
395 	/*
396 	 * If another thread is running the queue, increment
397 	 * busy counter to 2 so that the queue is retried,
398 	 * because the driver may now accept additional
399 	 * requests.
400 	 */
401 	if (dksc->sc_busy < 2)
402 		dksc->sc_busy++;
403 	if (dksc->sc_busy > 1)
404 		goto done;
405 
406 	/*
407 	 * Peeking at the buffer queue and committing the operation
408 	 * only after success isn't atomic.
409 	 *
410 	 * So when a diskstart fails, the buffer is saved
411 	 * and tried again before the next buffer is fetched.
412 	 * dk_drain() handles flushing of a saved buffer.
413 	 *
414 	 * This keeps order of I/O operations, unlike bufq_put.
415 	 */
416 
417 	while (dksc->sc_busy > 0) {
418 
419 		bp = dksc->sc_deferred;
420 		dksc->sc_deferred = NULL;
421 
422 		if (bp == NULL)
423 			bp = bufq_get(dksc->sc_bufq);
424 
425 		while (bp != NULL) {
426 
427 			disk_busy(&dksc->sc_dkdev);
428 			mutex_exit(&dksc->sc_iolock);
429 			error = dkd->d_diskstart(dksc->sc_dev, bp);
430 			mutex_enter(&dksc->sc_iolock);
431 			if (error == EAGAIN || error == ENOMEM) {
432 				/*
433 				 * Not a disk error. Retry later.
434 				 */
435 				KASSERT(dksc->sc_deferred == NULL);
436 				dksc->sc_deferred = bp;
437 				disk_unbusy(&dksc->sc_dkdev, 0, (bp->b_flags & B_READ));
438 				disk_wait(&dksc->sc_dkdev);
439 				break;
440 			}
441 
442 			if (error != 0) {
443 				bp->b_error = error;
444 				bp->b_resid = bp->b_bcount;
445 				dk_done1(dksc, bp, false);
446 			}
447 
448 			bp = bufq_get(dksc->sc_bufq);
449 		}
450 
451 		dksc->sc_busy--;
452 	}
453 done:
454 	mutex_exit(&dksc->sc_iolock);
455 }
456 
457 static void
458 dk_done1(struct dk_softc *dksc, struct buf *bp, bool lock)
459 {
460 	struct disk *dk = &dksc->sc_dkdev;
461 
462 	if (bp->b_error != 0) {
463 		struct cfdriver *cd = device_cfdriver(dksc->sc_dev);
464 
465 		diskerr(bp, cd->cd_name, "error", LOG_PRINTF, 0,
466 			dk->dk_label);
467 		printf("\n");
468 	}
469 
470 	if (lock)
471 		mutex_enter(&dksc->sc_iolock);
472 	disk_unbusy(dk, bp->b_bcount - bp->b_resid, (bp->b_flags & B_READ));
473 
474 	if ((dksc->sc_flags & DKF_NO_RND) == 0)
475 		rnd_add_uint32(&dksc->sc_rnd_source, bp->b_rawblkno);
476 	if (lock)
477 		mutex_exit(&dksc->sc_iolock);
478 
479 	biodone(bp);
480 }
481 
482 void
483 dk_done(struct dk_softc *dksc, struct buf *bp)
484 {
485 	dk_done1(dksc, bp, true);
486 }
487 
488 void
489 dk_drain(struct dk_softc *dksc)
490 {
491 	struct buf *bp;
492 
493 	mutex_enter(&dksc->sc_iolock);
494 	bp = dksc->sc_deferred;
495 	dksc->sc_deferred = NULL;
496 	if (bp != NULL) {
497 		bp->b_error = EIO;
498 		bp->b_resid = bp->b_bcount;
499 		biodone(bp);
500 	}
501 	bufq_drain(dksc->sc_bufq);
502 	mutex_exit(&dksc->sc_iolock);
503 }
504 
505 int
506 dk_discard(struct dk_softc *dksc, dev_t dev, off_t pos, off_t len)
507 {
508 	const struct dkdriver *dkd = dksc->sc_dkdev.dk_driver;
509 	unsigned secsize = dksc->sc_dkdev.dk_geom.dg_secsize;
510 	struct buf tmp, *bp = &tmp;
511 	int maxsz;
512 	int error = 0;
513 
514 	KASSERT(len >= 0);
515 
516 	DPRINTF_FOLLOW(("%s(%s, %p, 0x"PRIx64", %jd, %jd)\n", __func__,
517 	    dksc->sc_xname, dksc, (intmax_t)pos, (intmax_t)len));
518 
519 	if (!(dksc->sc_flags & DKF_INITED)) {
520 		DPRINTF_FOLLOW(("%s: not inited\n", __func__));
521 		return ENXIO;
522 	}
523 
524 	if (secsize == 0 || (pos % secsize) != 0 || (len % secsize) != 0)
525 		return EINVAL;
526 
527 	/* largest value that b_bcount can store */
528 	maxsz = rounddown(INT_MAX, secsize);
529 
530 	while (len > 0) {
531 		/* enough data to please the bounds checking code */
532 		bp->b_dev = dev;
533 		bp->b_blkno = (daddr_t)(pos / secsize);
534 		bp->b_bcount = uimin(len, maxsz);
535 		bp->b_flags = B_WRITE;
536 
537 		error = dk_translate(dksc, bp);
538 		if (error >= 0)
539 			break;
540 
541 		error = dkd->d_discard(dksc->sc_dev,
542 			(off_t)bp->b_rawblkno * secsize,
543 			(off_t)bp->b_bcount);
544 		if (error)
545 			break;
546 
547 		pos += bp->b_bcount;
548 		len -= bp->b_bcount;
549 	}
550 
551 	return error;
552 }
553 
554 int
555 dk_size(struct dk_softc *dksc, dev_t dev)
556 {
557 	const struct dkdriver *dkd = dksc->sc_dkdev.dk_driver;
558 	struct	disklabel *lp;
559 	int	is_open;
560 	int	part;
561 	int	size;
562 
563 	if ((dksc->sc_flags & DKF_INITED) == 0)
564 		return -1;
565 
566 	part = DISKPART(dev);
567 	is_open = dksc->sc_dkdev.dk_openmask & (1 << part);
568 
569 	if (!is_open && dkd->d_open(dev, 0, S_IFBLK, curlwp))
570 		return -1;
571 
572 	lp = dksc->sc_dkdev.dk_label;
573 	if (lp->d_partitions[part].p_fstype != FS_SWAP)
574 		size = -1;
575 	else
576 		size = lp->d_partitions[part].p_size *
577 		    (lp->d_secsize / DEV_BSIZE);
578 
579 	if (!is_open && dkd->d_close(dev, 0, S_IFBLK, curlwp))
580 		return -1;
581 
582 	return size;
583 }
584 
585 int
586 dk_ioctl(struct dk_softc *dksc, dev_t dev,
587 	    u_long cmd, void *data, int flag, struct lwp *l)
588 {
589 	const struct dkdriver *dkd = dksc->sc_dkdev.dk_driver;
590 	struct	disklabel *lp;
591 	struct	disk *dk = &dksc->sc_dkdev;
592 #ifdef __HAVE_OLD_DISKLABEL
593 	struct	disklabel newlabel;
594 #endif
595 	int	error;
596 
597 	DPRINTF_FOLLOW(("%s(%s, %p, 0x%"PRIx64", 0x%lx)\n", __func__,
598 	    dksc->sc_xname, dksc, dev, cmd));
599 
600 	/* ensure that the pseudo disk is open for writes for these commands */
601 	switch (cmd) {
602 	case DIOCSDINFO:
603 	case DIOCWDINFO:
604 #ifdef __HAVE_OLD_DISKLABEL
605 	case ODIOCSDINFO:
606 	case ODIOCWDINFO:
607 #endif
608 	case DIOCKLABEL:
609 	case DIOCWLABEL:
610 	case DIOCAWEDGE:
611 	case DIOCDWEDGE:
612 	case DIOCSSTRATEGY:
613 		if ((flag & FWRITE) == 0)
614 			return EBADF;
615 	}
616 
617 	/* ensure that the pseudo-disk is initialized for these */
618 	switch (cmd) {
619 	case DIOCGDINFO:
620 	case DIOCSDINFO:
621 	case DIOCWDINFO:
622 	case DIOCGPARTINFO:
623 	case DIOCKLABEL:
624 	case DIOCWLABEL:
625 	case DIOCGDEFLABEL:
626 	case DIOCAWEDGE:
627 	case DIOCDWEDGE:
628 	case DIOCLWEDGES:
629 	case DIOCMWEDGES:
630 	case DIOCRMWEDGES:
631 	case DIOCCACHESYNC:
632 #ifdef __HAVE_OLD_DISKLABEL
633 	case ODIOCGDINFO:
634 	case ODIOCSDINFO:
635 	case ODIOCWDINFO:
636 	case ODIOCGDEFLABEL:
637 #endif
638 		if ((dksc->sc_flags & DKF_INITED) == 0)
639 			return ENXIO;
640 	}
641 
642 	error = disk_ioctl(dk, dev, cmd, data, flag, l);
643 	if (error != EPASSTHROUGH)
644 		return error;
645 	else
646 		error = 0;
647 
648 	switch (cmd) {
649 	case DIOCWDINFO:
650 	case DIOCSDINFO:
651 #ifdef __HAVE_OLD_DISKLABEL
652 	case ODIOCWDINFO:
653 	case ODIOCSDINFO:
654 #endif
655 #ifdef __HAVE_OLD_DISKLABEL
656 		if (cmd == ODIOCSDINFO || cmd == ODIOCWDINFO) {
657 			memset(&newlabel, 0, sizeof newlabel);
658 			memcpy(&newlabel, data, sizeof (struct olddisklabel));
659 			lp = &newlabel;
660 		} else
661 #endif
662 		lp = (struct disklabel *)data;
663 
664 		mutex_enter(&dk->dk_openlock);
665 		dksc->sc_flags |= DKF_LABELLING;
666 
667 		error = setdisklabel(dksc->sc_dkdev.dk_label,
668 		    lp, 0, dksc->sc_dkdev.dk_cpulabel);
669 		if (error == 0) {
670 			if (cmd == DIOCWDINFO
671 #ifdef __HAVE_OLD_DISKLABEL
672 			    || cmd == ODIOCWDINFO
673 #endif
674 			   )
675 				error = writedisklabel(DKLABELDEV(dev),
676 				    dkd->d_strategy, dksc->sc_dkdev.dk_label,
677 				    dksc->sc_dkdev.dk_cpulabel);
678 		}
679 
680 		dksc->sc_flags &= ~DKF_LABELLING;
681 		mutex_exit(&dk->dk_openlock);
682 		break;
683 
684 	case DIOCKLABEL:
685 		if (*(int *)data != 0)
686 			dksc->sc_flags |= DKF_KLABEL;
687 		else
688 			dksc->sc_flags &= ~DKF_KLABEL;
689 		break;
690 
691 	case DIOCWLABEL:
692 		if (*(int *)data != 0)
693 			dksc->sc_flags |= DKF_WLABEL;
694 		else
695 			dksc->sc_flags &= ~DKF_WLABEL;
696 		break;
697 
698 	case DIOCGDEFLABEL:
699 		dk_getdefaultlabel(dksc, (struct disklabel *)data);
700 		break;
701 
702 #ifdef __HAVE_OLD_DISKLABEL
703 	case ODIOCGDEFLABEL:
704 		dk_getdefaultlabel(dksc, &newlabel);
705 		if (newlabel.d_npartitions > OLDMAXPARTITIONS)
706 			return ENOTTY;
707 		memcpy(data, &newlabel, sizeof (struct olddisklabel));
708 		break;
709 #endif
710 
711 	case DIOCGSTRATEGY:
712 	    {
713 		struct disk_strategy *dks = (void *)data;
714 
715 		mutex_enter(&dksc->sc_iolock);
716 		if (dksc->sc_bufq != NULL)
717 			strlcpy(dks->dks_name,
718 			    bufq_getstrategyname(dksc->sc_bufq),
719 			    sizeof(dks->dks_name));
720 		else
721 			error = EINVAL;
722 		mutex_exit(&dksc->sc_iolock);
723 		dks->dks_paramlen = 0;
724 		break;
725 	    }
726 
727 	case DIOCSSTRATEGY:
728 	    {
729 		struct disk_strategy *dks = (void *)data;
730 		struct bufq_state *new;
731 		struct bufq_state *old;
732 
733 		if (dks->dks_param != NULL) {
734 			return EINVAL;
735 		}
736 		dks->dks_name[sizeof(dks->dks_name) - 1] = 0; /* ensure term */
737 		error = bufq_alloc(&new, dks->dks_name,
738 		    BUFQ_EXACT|BUFQ_SORT_RAWBLOCK);
739 		if (error) {
740 			return error;
741 		}
742 		mutex_enter(&dksc->sc_iolock);
743 		old = dksc->sc_bufq;
744 		if (old)
745 			bufq_move(new, old);
746 		dksc->sc_bufq = new;
747 		mutex_exit(&dksc->sc_iolock);
748 		if (old)
749 			bufq_free(old);
750 		break;
751 	    }
752 
753 	default:
754 		error = ENOTTY;
755 	}
756 
757 	return error;
758 }
759 
760 /*
761  * dk_dump dumps all of physical memory into the partition specified.
762  * This requires substantially more framework than {s,w}ddump, and hence
763  * is probably much more fragile.
764  *
765  */
766 
767 #define DKFF_READYFORDUMP(x)	(((x) & DKF_READYFORDUMP) == DKF_READYFORDUMP)
768 static volatile int	dk_dumping = 0;
769 
770 /* ARGSUSED */
771 int
772 dk_dump(struct dk_softc *dksc, dev_t dev,
773     daddr_t blkno, void *vav, size_t size, int flags)
774 {
775 	const struct dkdriver *dkd = dksc->sc_dkdev.dk_driver;
776 	struct disk_geom *dg = &dksc->sc_dkdev.dk_geom;
777 	char *va = vav;
778 	struct disklabel *lp;
779 	struct partition *p;
780 	int part, towrt, maxblkcnt, nblk;
781 	int maxxfer, rv = 0;
782 
783 	/*
784 	 * ensure that we consider this device to be safe for dumping,
785 	 * and that the device is configured.
786 	 */
787 	if (!DKFF_READYFORDUMP(dksc->sc_flags)) {
788 		DPRINTF(DKDB_DUMP, ("%s: bad dump flags 0x%x\n", __func__,
789 		    dksc->sc_flags));
790 		return ENXIO;
791 	}
792 
793 	/* ensure that we are not already dumping */
794 	if (dk_dumping)
795 		return EFAULT;
796 	if ((flags & DK_DUMP_RECURSIVE) == 0)
797 		dk_dumping = 1;
798 
799 	if (dkd->d_dumpblocks == NULL) {
800 		DPRINTF(DKDB_DUMP, ("%s: no dumpblocks\n", __func__));
801 		return ENXIO;
802 	}
803 
804 	/* device specific max transfer size */
805 	maxxfer = MAXPHYS;
806 	if (dkd->d_iosize != NULL)
807 		(*dkd->d_iosize)(dksc->sc_dev, &maxxfer);
808 
809 	/* Convert to disk sectors.  Request must be a multiple of size. */
810 	part = DISKPART(dev);
811 	lp = dksc->sc_dkdev.dk_label;
812 	if ((size % lp->d_secsize) != 0) {
813 		DPRINTF(DKDB_DUMP, ("%s: odd size %zu\n", __func__, size));
814 		return EFAULT;
815 	}
816 	towrt = size / lp->d_secsize;
817 	blkno = dbtob(blkno) / lp->d_secsize;   /* blkno in secsize units */
818 
819 	p = &lp->d_partitions[part];
820 	if (part == RAW_PART) {
821 		if (p->p_fstype != FS_UNUSED) {
822 			DPRINTF(DKDB_DUMP, ("%s: bad fstype %d\n", __func__,
823 			    p->p_fstype));
824 			return ENXIO;
825 		}
826 		/* Check whether dump goes to a wedge */
827 		if (dksc->sc_dkdev.dk_nwedges == 0) {
828 			DPRINTF(DKDB_DUMP, ("%s: dump to raw\n", __func__));
829 			return ENXIO;
830 		}
831 		/* Check transfer bounds against media size */
832 		if (blkno < 0 || (blkno + towrt) > dg->dg_secperunit) {
833 			DPRINTF(DKDB_DUMP, ("%s: out of bounds blkno=%jd, towrt=%d, "
834 			    "nsects=%jd\n", __func__, (intmax_t)blkno, towrt, dg->dg_secperunit));
835 			return EINVAL;
836 		}
837 	} else {
838 		int nsects, sectoff;
839 
840 		if (p->p_fstype != FS_SWAP) {
841 			DPRINTF(DKDB_DUMP, ("%s: bad fstype %d\n", __func__,
842 			    p->p_fstype));
843 			return ENXIO;
844 		}
845 		nsects = p->p_size;
846 		sectoff = p->p_offset;
847 
848 		/* Check transfer bounds against partition size. */
849 		if ((blkno < 0) || ((blkno + towrt) > nsects)) {
850 			DPRINTF(DKDB_DUMP, ("%s: out of bounds blkno=%jd, towrt=%d, "
851 			    "nsects=%d\n", __func__, (intmax_t)blkno, towrt, nsects));
852 			return EINVAL;
853 		}
854 
855 		/* Offset block number to start of partition. */
856 		blkno += sectoff;
857 	}
858 
859 	/* Start dumping and return when done. */
860 	maxblkcnt = howmany(maxxfer, lp->d_secsize);
861 	while (towrt > 0) {
862 		nblk = uimin(maxblkcnt, towrt);
863 
864 		if ((rv = (*dkd->d_dumpblocks)(dksc->sc_dev, va, blkno, nblk))
865 		    != 0) {
866 			DPRINTF(DKDB_DUMP, ("%s: dumpblocks %d\n", __func__,
867 			    rv));
868 			return rv;
869 		}
870 
871 		towrt -= nblk;
872 		blkno += nblk;
873 		va += nblk * lp->d_secsize;
874 	}
875 
876 	if ((flags & DK_DUMP_RECURSIVE) == 0)
877 		dk_dumping = 0;
878 
879 	return 0;
880 }
881 
882 /* ARGSUSED */
883 void
884 dk_getdefaultlabel(struct dk_softc *dksc, struct disklabel *lp)
885 {
886 	const struct dkdriver *dkd = dksc->sc_dkdev.dk_driver;
887 	struct disk_geom *dg = &dksc->sc_dkdev.dk_geom;
888 
889 	memset(lp, 0, sizeof(*lp));
890 
891 	if (dg->dg_secperunit > UINT32_MAX)
892 		lp->d_secperunit = UINT32_MAX;
893 	else
894 		lp->d_secperunit = dg->dg_secperunit;
895 	lp->d_secsize = dg->dg_secsize;
896 	lp->d_nsectors = dg->dg_nsectors;
897 	lp->d_ntracks = dg->dg_ntracks;
898 	lp->d_ncylinders = dg->dg_ncylinders;
899 	lp->d_secpercyl = lp->d_ntracks * lp->d_nsectors;
900 
901 	strlcpy(lp->d_typename, dksc->sc_xname, sizeof(lp->d_typename));
902 	lp->d_type = dksc->sc_dtype;
903 	strlcpy(lp->d_packname, "fictitious", sizeof(lp->d_packname));
904 	lp->d_rpm = 3600;
905 	lp->d_interleave = 1;
906 	lp->d_flags = 0;
907 
908 	lp->d_partitions[RAW_PART].p_offset = 0;
909 	lp->d_partitions[RAW_PART].p_size = lp->d_secperunit;
910 	lp->d_partitions[RAW_PART].p_fstype = FS_UNUSED;
911 	lp->d_npartitions = RAW_PART + 1;
912 
913 	lp->d_magic = DISKMAGIC;
914 	lp->d_magic2 = DISKMAGIC;
915 
916 	if (dkd->d_label)
917 		dkd->d_label(dksc->sc_dev, lp);
918 
919 	lp->d_checksum = dkcksum(lp);
920 }
921 
922 /* ARGSUSED */
923 void
924 dk_getdisklabel(struct dk_softc *dksc, dev_t dev)
925 {
926 	const struct dkdriver *dkd = dksc->sc_dkdev.dk_driver;
927 	struct	 disklabel *lp = dksc->sc_dkdev.dk_label;
928 	struct	 cpu_disklabel *clp = dksc->sc_dkdev.dk_cpulabel;
929 	struct   disk_geom *dg = &dksc->sc_dkdev.dk_geom;
930 	struct	 partition *pp;
931 	int	 i, lpratio, dgratio;
932 	const char	*errstring;
933 
934 	memset(clp, 0x0, sizeof(*clp));
935 	dk_getdefaultlabel(dksc, lp);
936 	errstring = readdisklabel(DKLABELDEV(dev), dkd->d_strategy,
937 	    dksc->sc_dkdev.dk_label, dksc->sc_dkdev.dk_cpulabel);
938 	if (errstring) {
939 		dk_makedisklabel(dksc);
940 		if (dksc->sc_flags & DKF_WARNLABEL)
941 			printf("%s: %s\n", dksc->sc_xname, errstring);
942 		return;
943 	}
944 
945 	if ((dksc->sc_flags & DKF_LABELSANITY) == 0)
946 		return;
947 
948 	/* Convert sector counts to multiple of DEV_BSIZE for comparison */
949 	lpratio = dgratio = 1;
950 	if (lp->d_secsize > DEV_BSIZE)
951 		lpratio = lp->d_secsize / DEV_BSIZE;
952 	if (dg->dg_secsize > DEV_BSIZE)
953 		dgratio = dg->dg_secsize / DEV_BSIZE;
954 
955 	/* Sanity check */
956 	if ((uint64_t)lp->d_secperunit * lpratio > dg->dg_secperunit * dgratio)
957 		printf("WARNING: %s: "
958 		    "total unit size in disklabel (%" PRIu64 ") "
959 		    "!= the size of %s (%" PRIu64 ")\n", dksc->sc_xname,
960 		    (uint64_t)lp->d_secperunit * lpratio, dksc->sc_xname,
961 		    dg->dg_secperunit * dgratio);
962 	else if (lp->d_secperunit < UINT32_MAX &&
963 	    (uint64_t)lp->d_secperunit * lpratio < dg->dg_secperunit * dgratio)
964 		printf("%s: %" PRIu64 " trailing sectors not covered"
965 		    " by disklabel\n", dksc->sc_xname,
966 		    (dg->dg_secperunit * dgratio)
967 		    - (lp->d_secperunit * lpratio));
968 
969 	for (i=0; i < lp->d_npartitions; i++) {
970 		uint64_t pend;
971 
972 		pp = &lp->d_partitions[i];
973 		pend = pp->p_offset + pp->p_size;
974 		if (pend * lpratio > dg->dg_secperunit * dgratio)
975 			printf("WARNING: %s: end of partition `%c' exceeds "
976 			    "the size of %s (%" PRIu64 ")\n", dksc->sc_xname,
977 			    'a' + i, dksc->sc_xname,
978 			    dg->dg_secperunit * dgratio);
979 	}
980 }
981 
982 /*
983  * Heuristic to conjure a disklabel if reading a disklabel failed.
984  *
985  * This is to allow the raw partition to be used for a filesystem
986  * without caring about the write protected label sector.
987  *
988  * If the driver provides it's own callback, use that instead.
989  */
990 /* ARGSUSED */
991 static void
992 dk_makedisklabel(struct dk_softc *dksc)
993 {
994 	const struct dkdriver *dkd = dksc->sc_dkdev.dk_driver;
995 	struct  disklabel *lp = dksc->sc_dkdev.dk_label;
996 
997 	strlcpy(lp->d_packname, "default label", sizeof(lp->d_packname));
998 
999 	if (dkd->d_label)
1000 		dkd->d_label(dksc->sc_dev, lp);
1001 	else
1002 		lp->d_partitions[RAW_PART].p_fstype = FS_BSDFFS;
1003 
1004 	lp->d_checksum = dkcksum(lp);
1005 }
1006 
1007 MODULE(MODULE_CLASS_MISC, dk_subr, NULL);
1008 
1009 static int
1010 dk_subr_modcmd(modcmd_t cmd, void *arg)
1011 {
1012 	switch (cmd) {
1013 	case MODULE_CMD_INIT:
1014 	case MODULE_CMD_FINI:
1015 		return 0;
1016 	case MODULE_CMD_STAT:
1017 	case MODULE_CMD_AUTOUNLOAD:
1018 	default:
1019 		return ENOTTY;
1020 	}
1021 }
1022