xref: /netbsd-src/sys/arch/sun2/sun2/machdep.c (revision 3816d47b2c42fcd6e549e3407f842a5b1a1d23ad)
1 /*	$NetBSD: machdep.c,v 1.64 2009/12/10 14:13:52 matt Exp $	*/
2 
3 /*
4  * Copyright (c) 1982, 1986, 1990, 1993
5  *	The Regents of the University of California.  All rights reserved.
6  *
7  * This code is derived from software contributed to Berkeley by
8  * the Systems Programming Group of the University of Utah Computer
9  * Science Department.
10  *
11  * Redistribution and use in source and binary forms, with or without
12  * modification, are permitted provided that the following conditions
13  * are met:
14  * 1. Redistributions of source code must retain the above copyright
15  *    notice, this list of conditions and the following disclaimer.
16  * 2. Redistributions in binary form must reproduce the above copyright
17  *    notice, this list of conditions and the following disclaimer in the
18  *    documentation and/or other materials provided with the distribution.
19  * 3. Neither the name of the University nor the names of its contributors
20  *    may be used to endorse or promote products derived from this software
21  *    without specific prior written permission.
22  *
23  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
24  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
25  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
26  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
27  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
28  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
29  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
30  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
31  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
32  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
33  * SUCH DAMAGE.
34  *
35  *	from: Utah Hdr: machdep.c 1.74 92/12/20
36  *	from: @(#)machdep.c	8.10 (Berkeley) 4/20/94
37  */
38 
39 /*
40  * Copyright (c) 2001 Matthew Fredette.
41  * Copyright (c) 1994, 1995 Gordon W. Ross
42  * Copyright (c) 1993 Adam Glass
43  * Copyright (c) 1988 University of Utah.
44  *
45  * This code is derived from software contributed to Berkeley by
46  * the Systems Programming Group of the University of Utah Computer
47  * Science Department.
48  *
49  * Redistribution and use in source and binary forms, with or without
50  * modification, are permitted provided that the following conditions
51  * are met:
52  * 1. Redistributions of source code must retain the above copyright
53  *    notice, this list of conditions and the following disclaimer.
54  * 2. Redistributions in binary form must reproduce the above copyright
55  *    notice, this list of conditions and the following disclaimer in the
56  *    documentation and/or other materials provided with the distribution.
57  * 3. All advertising materials mentioning features or use of this software
58  *    must display the following acknowledgement:
59  *	This product includes software developed by the University of
60  *	California, Berkeley and its contributors.
61  * 4. Neither the name of the University nor the names of its contributors
62  *    may be used to endorse or promote products derived from this software
63  *    without specific prior written permission.
64  *
65  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
66  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
67  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
68  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
69  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
70  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
71  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
72  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
73  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
74  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
75  * SUCH DAMAGE.
76  *
77  *	from: Utah Hdr: machdep.c 1.74 92/12/20
78  *	from: @(#)machdep.c	8.10 (Berkeley) 4/20/94
79  */
80 
81 /*-
82  * Copyright (c) 1996, 1997, 1998 The NetBSD Foundation, Inc.
83  * All rights reserved.
84  *
85  * This code is derived from software contributed to The NetBSD Foundation
86  * by Jason R. Thorpe of the Numerical Aerospace Simulation Facility,
87  * NASA Ames Research Center.
88  *
89  * Redistribution and use in source and binary forms, with or without
90  * modification, are permitted provided that the following conditions
91  * are met:
92  * 1. Redistributions of source code must retain the above copyright
93  *    notice, this list of conditions and the following disclaimer.
94  * 2. Redistributions in binary form must reproduce the above copyright
95  *    notice, this list of conditions and the following disclaimer in the
96  *    documentation and/or other materials provided with the distribution.
97  *
98  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
99  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
100  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
101  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
102  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
103  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
104  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
105  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
106  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
107  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
108  * POSSIBILITY OF SUCH DAMAGE.
109  */
110 
111 /*
112  * Copyright (c) 1992, 1993
113  *	The Regents of the University of California.  All rights reserved.
114  *
115  * This software was developed by the Computer Systems Engineering group
116  * at Lawrence Berkeley Laboratory under DARPA contract BG 91-66 and
117  * contributed to Berkeley.
118  *
119  * All advertising materials mentioning features or use of this software
120  * must display the following acknowledgement:
121  *	This product includes software developed by the University of
122  *	California, Lawrence Berkeley Laboratory.
123  *
124  * Redistribution and use in source and binary forms, with or without
125  * modification, are permitted provided that the following conditions
126  * are met:
127  * 1. Redistributions of source code must retain the above copyright
128  *    notice, this list of conditions and the following disclaimer.
129  * 2. Redistributions in binary form must reproduce the above copyright
130  *    notice, this list of conditions and the following disclaimer in the
131  *    documentation and/or other materials provided with the distribution.
132  * 3. All advertising materials mentioning features or use of this software
133  *    must display the following acknowledgement:
134  *	This product includes software developed by the University of
135  *	California, Berkeley and its contributors.
136  * 4. Neither the name of the University nor the names of its contributors
137  *    may be used to endorse or promote products derived from this software
138  *    without specific prior written permission.
139  *
140  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
141  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
142  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
143  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
144  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
145  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
146  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
147  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
148  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
149  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
150  * SUCH DAMAGE.
151  *
152  *	@(#)machdep.c	8.6 (Berkeley) 1/14/94
153  */
154 
155 #include <sys/cdefs.h>
156 __KERNEL_RCSID(0, "$NetBSD: machdep.c,v 1.64 2009/12/10 14:13:52 matt Exp $");
157 
158 #include "opt_ddb.h"
159 #include "opt_kgdb.h"
160 #include "opt_fpu_emulate.h"
161 #include "opt_modular.h"
162 
163 #include <sys/param.h>
164 #include <sys/systm.h>
165 #include <sys/kernel.h>
166 #include <sys/proc.h>
167 #include <sys/buf.h>
168 #include <sys/reboot.h>
169 #include <sys/conf.h>
170 #include <sys/file.h>
171 #include <sys/device.h>
172 #include <sys/malloc.h>
173 #include <sys/extent.h>
174 #include <sys/mbuf.h>
175 #include <sys/msgbuf.h>
176 #include <sys/ioctl.h>
177 #include <sys/tty.h>
178 #include <sys/mount.h>
179 #include <sys/exec.h>
180 #include <sys/exec_aout.h>		/* for MID_* */
181 #include <sys/core.h>
182 #include <sys/kcore.h>
183 #include <sys/vnode.h>
184 #include <sys/syscallargs.h>
185 #include <sys/ksyms.h>
186 #ifdef	KGDB
187 #include <sys/kgdb.h>
188 #endif
189 
190 #include <uvm/uvm.h> /* XXX: not _extern ... need vm_map_create */
191 
192 #include <sys/sysctl.h>
193 
194 #include <dev/cons.h>
195 
196 #include <machine/promlib.h>
197 #include <machine/cpu.h>
198 #include <machine/dvma.h>
199 #include <machine/idprom.h>
200 #include <machine/kcore.h>
201 #include <machine/reg.h>
202 #include <machine/psl.h>
203 #include <machine/pte.h>
204 #define _SUN68K_BUS_DMA_PRIVATE
205 #include <machine/autoconf.h>
206 #include <machine/bus.h>
207 #include <machine/intr.h>
208 #include <machine/pmap.h>
209 
210 #if defined(DDB)
211 #include <machine/db_machdep.h>
212 #include <ddb/db_sym.h>
213 #include <ddb/db_extern.h>
214 #endif
215 
216 #include <dev/vme/vmereg.h>
217 #include <dev/vme/vmevar.h>
218 
219 #include <sun2/sun2/control.h>
220 #include <sun2/sun2/enable.h>
221 #include <sun2/sun2/machdep.h>
222 
223 #include <sun68k/sun68k/vme_sun68k.h>
224 
225 #include "ksyms.h"
226 
227 /* Defined in locore.s */
228 extern char kernel_text[];
229 /* Defined by the linker */
230 extern char etext[];
231 /* Defined in vfs_bio.c */
232 extern u_int bufpages;
233 
234 /* Our exported CPU info; we can have only one. */
235 struct cpu_info cpu_info_store;
236 
237 struct vm_map *mb_map = NULL;
238 struct vm_map *phys_map = NULL;
239 
240 int	physmem;
241 int	fputype;
242 void *	msgbufaddr;
243 
244 /* Virtual page frame for /dev/mem (see mem.c) */
245 vaddr_t vmmap;
246 
247 /*
248  * safepri is a safe priority for sleep to set for a spin-wait
249  * during autoconfiguration or after a panic.
250  */
251 int	safepri = PSL_LOWIPL;
252 
253 /* Soft copy of the enable register. */
254 volatile u_short enable_reg_soft = ENABLE_REG_SOFT_UNDEF;
255 
256 /*
257  * Our no-fault fault handler.
258  */
259 label_t *nofault;
260 
261 /*
262  * dvmamap is used to manage DVMA memory.
263  */
264 static struct extent *dvmamap;
265 
266 /* Our private scratch page for dumping the MMU. */
267 static vaddr_t dumppage;
268 
269 static void identifycpu(void);
270 static void initcpu(void);
271 
272 /*
273  * cpu_startup: allocate memory for variable-sized tables,
274  * initialize CPU, and do autoconfiguration.
275  *
276  * This is called early in init_main.c:main(), after the
277  * kernel memory allocator is ready for use, but before
278  * the creation of processes 1,2, and mountroot, etc.
279  */
280 void
281 cpu_startup(void)
282 {
283 	void *v;
284 	vaddr_t minaddr, maxaddr;
285 	char pbuf[9];
286 
287 	/*
288 	 * Initialize message buffer (for kernel printf).
289 	 * This is put in physical pages four through seven
290 	 * so it will always be in the same place after a
291 	 * reboot. (physical pages 0-3 are reserved by the PROM
292 	 * for its vector table and other stuff.)
293 	 * Its mapping was prepared in pmap_bootstrap().
294 	 * Also, offset some to avoid PROM scribbles.
295 	 */
296 	v = (void *) (PAGE_SIZE * 4);
297 	msgbufaddr = (void *)((char *)v + MSGBUFOFF);
298 	initmsgbuf(msgbufaddr, MSGBUFSIZE);
299 
300 #if NKSYMS || defined(DDB) || defined(MODULAR)
301 	{
302 		extern int nsym;
303 		extern char *ssym, *esym;
304 
305 		ksyms_addsyms_elf(nsym, ssym, esym);
306 	}
307 #endif /* DDB */
308 
309 	/*
310 	 * Good {morning,afternoon,evening,night}.
311 	 */
312 	printf("%s%s", copyright, version);
313 	identifycpu();
314 	fputype = FPU_NONE;
315 #ifdef  FPU_EMULATE
316 	printf("fpu: emulator\n");
317 #else
318 	printf("fpu: no math support\n");
319 #endif
320 
321 	format_bytes(pbuf, sizeof(pbuf), ctob(physmem));
322 	printf("total memory = %s\n", pbuf);
323 
324 	/*
325 	 * XXX fredette - we force a small number of buffers
326 	 * to help me debug this on my low-memory machine.
327 	 * this should go away at some point, allowing the
328 	 * normal automatic buffer-sizing to happen.
329 	 */
330 	bufpages = 37;
331 
332 	/*
333 	 * Get scratch page for dumpsys().
334 	 */
335 	if ((dumppage = uvm_km_alloc(kernel_map, PAGE_SIZE,0, UVM_KMF_WIRED))
336 	    == 0)
337 		panic("startup: alloc dumppage");
338 
339 
340 	minaddr = 0;
341 
342 	/*
343 	 * Allocate a submap for physio
344 	 */
345 	phys_map = uvm_km_suballoc(kernel_map, &minaddr, &maxaddr,
346 				   VM_PHYS_SIZE, 0, false, NULL);
347 
348 	/*
349 	 * Finally, allocate mbuf cluster submap.
350 	 */
351 	mb_map = uvm_km_suballoc(kernel_map, &minaddr, &maxaddr,
352 				 nmbclusters * mclbytes, VM_MAP_INTRSAFE,
353 				 false, NULL);
354 
355 	format_bytes(pbuf, sizeof(pbuf), ptoa(uvmexp.free));
356 	printf("avail memory = %s\n", pbuf);
357 
358 	/*
359 	 * Allocate a virtual page (for use by /dev/mem)
360 	 * This page is handed to pmap_enter() therefore
361 	 * it has to be in the normal kernel VA range.
362 	 */
363 	vmmap = uvm_km_alloc(kernel_map, PAGE_SIZE, 0,
364 	    UVM_KMF_VAONLY | UVM_KMF_WAITVA);
365 
366 	/*
367 	 * Allocate DMA map for devices on the bus.
368 	 */
369 	dvmamap = extent_create("dvmamap",
370 	    DVMA_MAP_BASE, DVMA_MAP_BASE + DVMA_MAP_AVAIL,
371 	    M_DEVBUF, 0, 0, EX_NOWAIT);
372 	if (dvmamap == NULL)
373 		panic("unable to allocate DVMA map");
374 
375 	/*
376 	 * Set up CPU-specific registers, cache, etc.
377 	 */
378 	initcpu();
379 }
380 
381 /*
382  * Set registers on exec.
383  */
384 void
385 setregs(struct lwp *l, struct exec_package *pack, vaddr_t stack)
386 {
387 	struct trapframe *tf = (struct trapframe *)l->l_md.md_regs;
388 	struct pcb *pcb = lwp_getpcb(l);
389 
390 	tf->tf_sr = PSL_USERSET;
391 	tf->tf_pc = pack->ep_entry & ~1;
392 	tf->tf_regs[D0] = 0;
393 	tf->tf_regs[D1] = 0;
394 	tf->tf_regs[D2] = 0;
395 	tf->tf_regs[D3] = 0;
396 	tf->tf_regs[D4] = 0;
397 	tf->tf_regs[D5] = 0;
398 	tf->tf_regs[D6] = 0;
399 	tf->tf_regs[D7] = 0;
400 	tf->tf_regs[A0] = 0;
401 	tf->tf_regs[A1] = 0;
402 	tf->tf_regs[A2] = (int)l->l_proc->p_psstr;
403 	tf->tf_regs[A3] = 0;
404 	tf->tf_regs[A4] = 0;
405 	tf->tf_regs[A5] = 0;
406 	tf->tf_regs[A6] = 0;
407 	tf->tf_regs[SP] = stack;
408 
409 	/* restore a null state frame */
410 	pcb->pcb_fpregs.fpf_null = 0;
411 
412 	l->l_md.md_flags = 0;
413 }
414 
415 /*
416  * Info for CTL_HW
417  */
418 char	machine[16] = MACHINE;		/* from <machine/param.h> */
419 char	kernel_arch[16] = "sun2";	/* XXX needs a sysctl node */
420 char	cpu_model[120];
421 
422 /*
423  * Determine which Sun2 model we are running on.
424  */
425 void
426 identifycpu(void)
427 {
428 	extern char *cpu_string;	/* XXX */
429 
430 	/* Other stuff? (VAC, mc6888x version, etc.) */
431 	/* Note: miniroot cares about the kernel_arch part. */
432 	sprintf(cpu_model, "%s %s", kernel_arch, cpu_string);
433 
434 	printf("Model: %s\n", cpu_model);
435 }
436 
437 /*
438  * machine dependent system variables.
439  */
440 #if 0	/* XXX - Not yet... */
441 static int
442 sysctl_machdep_root_device(SYSCTLFN_ARGS)
443 {
444 	struct sysctlnode node = *rnode;
445 
446 	node.sysctl_data = some permutation on root_device;
447 	node.sysctl_size = strlen(root_device) + 1;
448 	return (sysctl_lookup(SYSCTLFN_CALL(&node)));
449 }
450 #endif
451 
452 static int
453 sysctl_machdep_booted_kernel(SYSCTLFN_ARGS)
454 {
455 	struct sysctlnode node = *rnode;
456 	char *cp;
457 
458 	cp = prom_getbootfile();
459 	if (cp == NULL || cp[0] == '\0')
460 		return (ENOENT);
461 
462 	node.sysctl_data = cp;
463 	node.sysctl_size = strlen(cp) + 1;
464 	return (sysctl_lookup(SYSCTLFN_CALL(&node)));
465 }
466 
467 SYSCTL_SETUP(sysctl_machdep_setup, "sysctl machdep subtree setup")
468 {
469 
470 	sysctl_createv(clog, 0, NULL, NULL,
471 		       CTLFLAG_PERMANENT,
472 		       CTLTYPE_NODE, "machdep", NULL,
473 		       NULL, 0, NULL, 0,
474 		       CTL_MACHDEP, CTL_EOL);
475 
476 	sysctl_createv(clog, 0, NULL, NULL,
477 		       CTLFLAG_PERMANENT,
478 		       CTLTYPE_STRUCT, "console_device", NULL,
479 		       sysctl_consdev, 0, NULL, sizeof(dev_t),
480 		       CTL_MACHDEP, CPU_CONSDEV, CTL_EOL);
481 #if 0	/* XXX - Not yet... */
482 	sysctl_createv(clog, 0, NULL, NULL,
483 		       CTLFLAG_PERMANENT,
484 		       CTLTYPE_STRING, "root_device", NULL,
485 		       sysctl_machdep_root_device, 0, NULL, 0,
486 		       CTL_MACHDEP, CPU_ROOT_DEVICE, CTL_EOL);
487 #endif
488 	sysctl_createv(clog, 0, NULL, NULL,
489 		       CTLFLAG_PERMANENT,
490 		       CTLTYPE_STRING, "booted_kernel", NULL,
491 		       sysctl_machdep_booted_kernel, 0, NULL, 0,
492 		       CTL_MACHDEP, CPU_BOOTED_KERNEL, CTL_EOL);
493 }
494 
495 /* See: sig_machdep.c */
496 
497 /*
498  * Do a sync in preparation for a reboot.
499  * XXX - This could probably be common code.
500  * XXX - And now, most of it is in vfs_shutdown()
501  * XXX - Put waittime checks in there too?
502  */
503 int waittime = -1;	/* XXX - Who else looks at this? -gwr */
504 static void
505 reboot_sync(void)
506 {
507 
508 	/* Check waittime here to localize its use to this function. */
509 	if (waittime >= 0)
510 		return;
511 	waittime = 0;
512 	vfs_shutdown();
513 }
514 
515 /*
516  * Common part of the BSD and SunOS reboot system calls.
517  */
518 __dead void
519 cpu_reboot(int howto, char *user_boot_string)
520 {
521 	char *bs, *p;
522 	char default_boot_string[8];
523 
524 	/* If system is cold, just halt. (early panic?) */
525 	if (cold)
526 		goto haltsys;
527 
528 	/* Un-blank the screen if appropriate. */
529 	cnpollc(1);
530 
531 	if ((howto & RB_NOSYNC) == 0) {
532 		reboot_sync();
533 		/*
534 		 * If we've been adjusting the clock, the todr
535 		 * will be out of synch; adjust it now.
536 		 *
537 		 * XXX - However, if the kernel has been sitting in ddb,
538 		 * the time will be way off, so don't set the HW clock!
539 		 * XXX - Should do sanity check against HW clock. -gwr
540 		 */
541 		/* resettodr(); */
542 	}
543 
544 	/* Disable interrupts. */
545 	splhigh();
546 
547 	/* Write out a crash dump if asked. */
548 	if (howto & RB_DUMP)
549 		dumpsys();
550 
551 	/* run any shutdown hooks */
552 	doshutdownhooks();
553 
554 	pmf_system_shutdown(boothowto);
555 
556 	if (howto & RB_HALT) {
557 	haltsys:
558 		printf("halted.\n");
559 		prom_halt();
560 	}
561 
562 	/*
563 	 * Automatic reboot.
564 	 */
565 	bs = user_boot_string;
566 	if (bs == NULL) {
567 		/*
568 		 * Build our own boot string with an empty
569 		 * boot device/file and (maybe) some flags.
570 		 * The PROM will supply the device/file name.
571 		 */
572 		bs = default_boot_string;
573 		*bs = '\0';
574 		if (howto & (RB_KDB|RB_ASKNAME|RB_SINGLE)) {
575 			/* Append the boot flags. */
576 			p = bs;
577 			*p++ = ' ';
578 			*p++ = '-';
579 			if (howto & RB_KDB)
580 				*p++ = 'd';
581 			if (howto & RB_ASKNAME)
582 				*p++ = 'a';
583 			if (howto & RB_SINGLE)
584 				*p++ = 's';
585 			*p = '\0';
586 		}
587 	}
588 	printf("rebooting...\n");
589 	prom_boot(bs);
590 	for (;;) ;
591 	/*NOTREACHED*/
592 }
593 
594 /*
595  * These variables are needed by /sbin/savecore
596  */
597 uint32_t dumpmag = 0x8fca0101;	/* magic number */
598 int 	dumpsize = 0;		/* pages */
599 long	dumplo = 0; 		/* blocks */
600 
601 #define	DUMP_EXTRA 	3	/* CPU-dependent extra pages */
602 
603 /*
604  * This is called by main to set dumplo, dumpsize.
605  * Dumps always skip the first PAGE_SIZE of disk space
606  * in case there might be a disk label stored there.
607  * If there is extra space, put dump at the end to
608  * reduce the chance that swapping trashes it.
609  */
610 void
611 cpu_dumpconf(void)
612 {
613 	const struct bdevsw *bdev;
614 	int devblks;	/* size of dump device in blocks */
615 	int dumpblks;	/* size of dump image in blocks */
616 	int (*getsize)(dev_t);
617 
618 	if (dumpdev == NODEV)
619 		return;
620 
621 	bdev = bdevsw_lookup(dumpdev);
622 	if (bdev == NULL) {
623 		dumpdev = NODEV;
624 		return;
625 	}
626 	getsize = bdev->d_psize;
627 	if (getsize == NULL)
628 		return;
629 	devblks = (*getsize)(dumpdev);
630 	if (devblks <= ctod(1))
631 		return;
632 	devblks &= ~(ctod(1)-1);
633 
634 	/*
635 	 * Note: savecore expects dumpsize to be the
636 	 * number of pages AFTER the dump header.
637 	 */
638 	dumpsize = physmem;
639 
640 	/* Position dump image near end of space, page aligned. */
641 	dumpblks = ctod(physmem + DUMP_EXTRA);
642 	dumplo = devblks - dumpblks;
643 
644 	/* If it does not fit, truncate it by moving dumplo. */
645 	/* Note: Must force signed comparison. */
646 	if (dumplo < ((long)ctod(1))) {
647 		dumplo = ctod(1);
648 		dumpsize = dtoc(devblks - dumplo) - DUMP_EXTRA;
649 	}
650 }
651 
652 /* Note: gdb looks for "dumppcb" in a kernel crash dump. */
653 struct pcb dumppcb;
654 extern paddr_t avail_start;
655 
656 /*
657  * Write a crash dump.  The format while in swap is:
658  *   kcore_seg_t cpu_hdr;
659  *   cpu_kcore_hdr_t cpu_data;
660  *   padding (PAGE_SIZE-sizeof(kcore_seg_t))
661  *   pagemap (2*PAGE_SIZE)
662  *   physical memory...
663  */
664 void
665 dumpsys(void)
666 {
667 	const struct bdevsw *dsw;
668 	kcore_seg_t	*kseg_p;
669 	cpu_kcore_hdr_t *chdr_p;
670 	struct sun2_kcore_hdr *sh;
671 	char *vaddr;
672 	paddr_t paddr;
673 	int psize, todo, chunk;
674 	daddr_t blkno;
675 	int error = 0;
676 
677 	if (dumpdev == NODEV)
678 		return;
679 	dsw = bdevsw_lookup(dumpdev);
680 	if (dsw == NULL || dsw->d_psize == NULL)
681 		return;
682 	if (dumppage == 0)
683 		return;
684 
685 	/*
686 	 * For dumps during autoconfiguration,
687 	 * if dump device has already configured...
688 	 */
689 	if (dumpsize == 0)
690 		cpu_dumpconf();
691 	if (dumplo <= 0) {
692 		printf("\ndump to dev %u,%u not possible\n",
693 		    major(dumpdev), minor(dumpdev));
694 		return;
695 	}
696 	savectx(&dumppcb);
697 
698 	psize = (*(dsw->d_psize))(dumpdev);
699 	if (psize == -1) {
700 		printf("dump area unavailable\n");
701 		return;
702 	}
703 
704 	printf("\ndumping to dev %u,%u offset %ld\n",
705 	    major(dumpdev), minor(dumpdev), dumplo);
706 
707 	/*
708 	 * Prepare the dump header, including MMU state.
709 	 */
710 	blkno = dumplo;
711 	todo = dumpsize;	/* pages */
712 	vaddr = (char*)dumppage;
713 	memset(vaddr, 0, PAGE_SIZE);
714 
715 	/* Set pointers to all three parts. */
716 	kseg_p = (kcore_seg_t *)vaddr;
717 	chdr_p = (cpu_kcore_hdr_t *) (kseg_p + 1);
718 	sh = &chdr_p->un._sun2;
719 
720 	/* Fill in kcore_seg_t part. */
721 	CORE_SETMAGIC(*kseg_p, KCORE_MAGIC, MID_MACHINE, CORE_CPU);
722 	kseg_p->c_size = (ctob(DUMP_EXTRA) - sizeof(*kseg_p));
723 
724 	/* Fill in cpu_kcore_hdr_t part. */
725 	strncpy(chdr_p->name, kernel_arch, sizeof(chdr_p->name));
726 	chdr_p->page_size = PAGE_SIZE;
727 	chdr_p->kernbase = KERNBASE;
728 
729 	/* Fill in the sun2_kcore_hdr part (MMU state). */
730 	pmap_kcore_hdr(sh);
731 
732 	/* Write out the dump header. */
733 	error = (*dsw->d_dump)(dumpdev, blkno, vaddr, PAGE_SIZE);
734 	if (error)
735 		goto fail;
736 	blkno += btodb(PAGE_SIZE);
737 
738 	/* translation RAM (pages zero through seven) */
739 	for(chunk = 0; chunk < (PAGE_SIZE * 8); chunk += PAGE_SIZE) {
740 		pmap_get_pagemap((int*)vaddr, chunk);
741 		error = (*dsw->d_dump)(dumpdev, blkno, vaddr, PAGE_SIZE);
742 		if (error)
743 			goto fail;
744 		blkno += btodb(PAGE_SIZE);
745 	}
746 
747 	/*
748 	 * Now dump physical memory.  Have to do it in two chunks.
749 	 * The first chunk is "unmanaged" (by the VM code) and its
750 	 * range of physical addresses is not allow in pmap_enter.
751 	 * However, that segment is mapped linearly, so we can just
752 	 * use the virtual mappings already in place.  The second
753 	 * chunk is done the normal way, using pmap_enter.
754 	 *
755 	 * Note that vaddr==(paddr+KERNBASE) for paddr=0 through etext.
756 	 */
757 
758 	/* Do the first chunk (0 <= PA < avail_start) */
759 	paddr = 0;
760 	chunk = btoc(avail_start);
761 	if (chunk > todo)
762 		chunk = todo;
763 	do {
764 		if ((todo & 0xf) == 0)
765 			printf_nolog("\r%4d", todo);
766 		vaddr = (char*)(paddr + KERNBASE);
767 		error = (*dsw->d_dump)(dumpdev, blkno, vaddr, PAGE_SIZE);
768 		if (error)
769 			goto fail;
770 		paddr += PAGE_SIZE;
771 		blkno += btodb(PAGE_SIZE);
772 		--todo;
773 	} while (--chunk > 0);
774 
775 	/* Do the second chunk (avail_start <= PA < dumpsize) */
776 	vaddr = (char*)vmmap;	/* Borrow /dev/mem VA */
777 	do {
778 		if ((todo & 0xf) == 0)
779 			printf_nolog("\r%4d", todo);
780 		pmap_kenter_pa(vmmap, paddr | PMAP_NC, VM_PROT_READ, 0);
781 		pmap_update(pmap_kernel());
782 		error = (*dsw->d_dump)(dumpdev, blkno, vaddr, PAGE_SIZE);
783 		pmap_kremove(vmmap, PAGE_SIZE);
784 		pmap_update(pmap_kernel());
785 		if (error)
786 			goto fail;
787 		paddr += PAGE_SIZE;
788 		blkno += btodb(PAGE_SIZE);
789 	} while (--todo > 0);
790 
791 	printf("\rdump succeeded\n");
792 	return;
793 fail:
794 	printf(" dump error=%d\n", error);
795 }
796 
797 static void
798 initcpu(void)
799 {
800 	/* XXX: Enable RAM parity/ECC checking? */
801 	/* XXX: parityenable(); */
802 
803 }
804 
805 /* straptrap() in trap.c */
806 
807 /* from hp300: badaddr() */
808 
809 /* XXX: parityenable() ? */
810 /* regdump() moved to regdump.c */
811 
812 /*
813  * cpu_exec_aout_makecmds():
814  *	CPU-dependent a.out format hook for execve().
815  *
816  * Determine if the given exec package refers to something which we
817  * understand and, if so, set up the vmcmds for it.
818  */
819 int
820 cpu_exec_aout_makecmds(struct lwp *l, struct exec_package *epp)
821 {
822 	return ENOEXEC;
823 }
824 
825 #if 0
826 /*
827  * Soft interrupt support.
828  */
829 void
830 isr_soft_request(int level)
831 {
832 	u_char bit;
833 
834 	if ((level < _IPL_SOFT_LEVEL_MIN) || (level > _IPL_SOFT_LEVEL_MAX))
835 		return;
836 
837 	bit = 1 << level;
838 	enable_reg_or(bit);
839 }
840 
841 void
842 isr_soft_clear(int level)
843 {
844 	u_char bit;
845 
846 	if ((level < _IPL_SOFT_LEVEL_MIN) || (level > _IPL_SOFT_LEVEL_MAX))
847 		return;
848 
849 	bit = 1 << level;
850 	enable_reg_and(~bit);
851 }
852 #endif
853 
854 /*
855  * Like _bus_dmamap_load(), but for raw memory allocated with
856  * bus_dmamem_alloc().
857  */
858 int
859 _bus_dmamap_load_raw(bus_dma_tag_t t, bus_dmamap_t map, bus_dma_segment_t *segs,
860     int nsegs, bus_size_t size, int flags)
861 {
862 	struct vm_page *m;
863 	paddr_t pa;
864 	bus_addr_t dva;
865 	bus_size_t sgsize;
866 	struct pglist *mlist;
867 	int pagesz = PAGE_SIZE;
868 	int error;
869 
870 	/*
871 	 * Make sure that on error condition we return "no valid mappings".
872 	 */
873 	map->dm_nsegs = 0;
874 	map->dm_mapsize = 0;
875 
876 	/* Allocate DVMA addresses */
877 	sgsize = (size + pagesz - 1) & -pagesz;
878 
879 	/*
880 	 * If the device can see our entire 24-bit address space,
881 	 * we can use any properly aligned virtual addresses.
882 	 */
883 	if ((map->_dm_flags & BUS_DMA_24BIT) != 0) {
884 		dva = _bus_dma_valloc_skewed(sgsize, map->_dm_boundary,
885 					     pagesz, 0);
886 		if (dva == 0)
887 			return (ENOMEM);
888 	}
889 
890 	/*
891 	 * Otherwise, we need virtual addresses in DVMA space.
892 	 */
893 	else {
894 		error = extent_alloc(dvmamap, sgsize, pagesz,
895 					map->_dm_boundary,
896 					(flags & BUS_DMA_NOWAIT) == 0
897 						? EX_WAITOK : EX_NOWAIT,
898 					(u_long *)&dva);
899 		if (error)
900 			return (error);
901 	}
902 
903 	/* Fill in the segment. */
904 	map->dm_segs[0].ds_addr = dva;
905 	map->dm_segs[0].ds_len = size;
906 	map->dm_segs[0]._ds_va = dva;
907 	map->dm_segs[0]._ds_sgsize = sgsize;
908 
909 	/* Map physical pages into MMU */
910 	mlist = segs[0]._ds_mlist;
911 	for (m = TAILQ_FIRST(mlist); m != NULL; m = TAILQ_NEXT(m,pageq.queue)) {
912 		if (sgsize == 0)
913 			panic("_bus_dmamap_load_raw: size botch");
914 		pa = VM_PAGE_TO_PHYS(m);
915 		pmap_enter(pmap_kernel(), dva,
916 			   (pa & -pagesz) | PMAP_NC,
917 			   VM_PROT_READ|VM_PROT_WRITE, PMAP_WIRED);
918 
919 		dva += pagesz;
920 		sgsize -= pagesz;
921 	}
922 	pmap_update(pmap_kernel());
923 
924 	/* Make the map truly valid. */
925 	map->dm_nsegs = 1;
926 	map->dm_mapsize = size;
927 
928 	return (0);
929 }
930 
931 /*
932  * load DMA map with a linear buffer.
933  */
934 int
935 _bus_dmamap_load(bus_dma_tag_t t, bus_dmamap_t map, void *buf,
936     bus_size_t buflen, struct proc *p, int flags)
937 {
938 	bus_size_t sgsize;
939 	vaddr_t va = (vaddr_t)buf;
940 	int pagesz = PAGE_SIZE;
941 	bus_addr_t dva;
942 	pmap_t pmap;
943 	int rv;
944 
945 	/*
946 	 * Make sure that on error condition we return "no valid mappings".
947 	 */
948 	map->dm_nsegs = 0;
949 	map->dm_mapsize = 0;
950 
951 	if (buflen > map->_dm_size)
952 		return (EINVAL);
953 
954 	/*
955 	 * A 24-bit device can see all of our kernel address space, so
956 	 * if we have KVAs, we can just load them as-is, no mapping
957 	 * necessary.
958 	 */
959 	if ((map->_dm_flags & BUS_DMA_24BIT) != 0 && p == NULL) {
960 		/*
961 		 * XXX Need to implement "don't DMA across this boundry".
962 		 */
963 		if (map->_dm_boundary != 0)
964 			panic("bus_dmamap_load: boundaries not implemented");
965 		map->dm_mapsize = buflen;
966 		map->dm_nsegs = 1;
967 		map->dm_segs[0].ds_addr = (bus_addr_t)va;
968 		map->dm_segs[0].ds_len = buflen;
969 		map->_dm_flags |= _BUS_DMA_DIRECTMAP;
970 		return (0);
971 	}
972 
973 	/*
974 	 * Allocate a region in DVMA space.
975 	 */
976 	sgsize = m68k_round_page(buflen + (va & (pagesz - 1)));
977 
978 	if (extent_alloc(dvmamap, sgsize, pagesz, map->_dm_boundary,
979 			 (flags & BUS_DMA_NOWAIT) == 0 ? EX_WAITOK : EX_NOWAIT,
980 			 (u_long *)&dva) != 0) {
981 		return (ENOMEM);
982 	}
983 
984 	/* Fill in the segment. */
985 	map->dm_segs[0].ds_addr = dva + (va & (pagesz - 1));
986 	map->dm_segs[0].ds_len = buflen;
987 	map->dm_segs[0]._ds_va = dva;
988 	map->dm_segs[0]._ds_sgsize = sgsize;
989 
990 	/*
991 	 * Now map the DVMA addresses we allocated to point to the
992 	 * pages of the caller's buffer.
993 	 */
994 	if (p != NULL)
995 		pmap = p->p_vmspace->vm_map.pmap;
996 	else
997 		pmap = pmap_kernel();
998 
999 	for (; buflen > 0; ) {
1000 		paddr_t pa;
1001 		/*
1002 		 * Get the physical address for this page.
1003 		 */
1004 		rv = pmap_extract(pmap, va, &pa);
1005 #ifdef	DIAGNOSTIC
1006 		if (!rv)
1007 			panic("_bus_dmamap_load: no page");
1008 #endif	/* DIAGNOSTIC */
1009 
1010 		/*
1011 		 * Compute the segment size, and adjust counts.
1012 		 */
1013 		sgsize = pagesz - (va & (pagesz - 1));
1014 		if (buflen < sgsize)
1015 			sgsize = buflen;
1016 
1017 		pmap_enter(pmap_kernel(), dva,
1018 			   (pa & -pagesz) | PMAP_NC,
1019 			   VM_PROT_READ|VM_PROT_WRITE, PMAP_WIRED);
1020 
1021 		dva += pagesz;
1022 		va += sgsize;
1023 		buflen -= sgsize;
1024 	}
1025 	pmap_update(pmap_kernel());
1026 
1027 	/* Make the map truly valid. */
1028 	map->dm_nsegs = 1;
1029 	map->dm_mapsize = map->dm_segs[0].ds_len;
1030 
1031 	return (0);
1032 }
1033 
1034 /*
1035  * unload a DMA map.
1036  */
1037 void
1038 _bus_dmamap_unload(bus_dma_tag_t t, bus_dmamap_t map)
1039 {
1040 	bus_dma_segment_t *segs = map->dm_segs;
1041 	int nsegs = map->dm_nsegs;
1042 	int flags = map->_dm_flags;
1043 	bus_addr_t dva;
1044 	bus_size_t len;
1045 	int s, error;
1046 
1047 	if (nsegs != 1)
1048 		panic("_bus_dmamem_unload: nsegs = %d", nsegs);
1049 
1050 	/*
1051 	 * _BUS_DMA_DIRECTMAP is set iff this map was loaded using
1052 	 * _bus_dmamap_load for a 24-bit device.
1053 	 */
1054 	if ((flags & _BUS_DMA_DIRECTMAP) != 0) {
1055 		/* Nothing to release */
1056 		map->_dm_flags &= ~_BUS_DMA_DIRECTMAP;
1057 	}
1058 
1059 	/*
1060 	 * Otherwise, this map was loaded using _bus_dmamap_load for a
1061 	 * non-24-bit device, or using _bus_dmamap_load_raw.
1062 	 */
1063 	else {
1064 		dva = segs[0]._ds_va & -PAGE_SIZE;
1065 		len = segs[0]._ds_sgsize;
1066 
1067 		/*
1068 		 * Unmap the DVMA addresses.
1069 		 */
1070 		pmap_remove(pmap_kernel(), dva, dva + len);
1071 		pmap_update(pmap_kernel());
1072 
1073 		/*
1074 		 * Free the DVMA addresses.
1075 		 */
1076 		if ((flags & BUS_DMA_24BIT) != 0) {
1077 			/*
1078 			 * This map was loaded using _bus_dmamap_load_raw
1079 			 * for a 24-bit device.
1080 			 */
1081 			uvm_unmap(kernel_map, dva, dva + len);
1082 		} else {
1083 			/*
1084 			 * This map was loaded using _bus_dmamap_load or
1085 			 * _bus_dmamap_load_raw for a non-24-bit device.
1086 			 */
1087 			s = splhigh();
1088 			error = extent_free(dvmamap, dva, len, EX_NOWAIT);
1089 			splx(s);
1090 			if (error != 0)
1091 				printf("warning: %ld of DVMA space lost\n", len);
1092 		}
1093 	}
1094 
1095 	/* Mark the mappings as invalid. */
1096 	map->dm_mapsize = 0;
1097 	map->dm_nsegs = 0;
1098 }
1099 
1100 /*
1101  * Translate a VME address and address modifier
1102  * into a CPU physical address and page type.
1103  */
1104 int
1105 vmebus_translate(vme_am_t mod, vme_addr_t addr, bus_type_t *btp,
1106     bus_addr_t *bap)
1107 {
1108 	bus_addr_t base;
1109 
1110 	switch(mod) {
1111 #define _DS (VME_AM_MBO | VME_AM_SUPER | VME_AM_DATA)
1112 
1113 	case (VME_AM_A16|_DS):
1114 		base = 0x00ff0000;
1115 		break;
1116 
1117 	case (VME_AM_A24|_DS):
1118 		base = 0;
1119 		break;
1120 
1121 	default:
1122 		return (ENOENT);
1123 #undef _DS
1124 	}
1125 
1126 	*bap = base | addr;
1127 	*btp = (*bap & 0x800000 ? PMAP_VME8 : PMAP_VME0);
1128 	return (0);
1129 }
1130 
1131 /*
1132  * If we can find a mapping that was established by the PROM, use it.
1133  */
1134 int
1135 find_prom_map(paddr_t pa, bus_type_t iospace, int len, vaddr_t *vap)
1136 {
1137 	u_long	pf;
1138 	int	pgtype;
1139 	vaddr_t	va, eva;
1140 	int	sme;
1141 	u_long	pte;
1142 	int	saved_ctx;
1143 
1144 	/*
1145 	 * The mapping must fit entirely within one page.
1146 	 */
1147 	if ((((u_long)pa & PGOFSET) + len) > PAGE_SIZE)
1148 		return EINVAL;
1149 
1150 	pf = PA_PGNUM(pa);
1151 	pgtype = iospace << PG_MOD_SHIFT;
1152 	saved_ctx = kernel_context();
1153 
1154 	/*
1155 	 * Walk the PROM address space, looking for a page with the
1156 	 * mapping we want.
1157 	 */
1158 	for (va = SUN_MONSTART; va < SUN_MONEND; ) {
1159 
1160 		/*
1161 		 * Make sure this segment is mapped.
1162 		 */
1163 		sme = get_segmap(va);
1164 		if (sme == SEGINV) {
1165 			va += NBSG;
1166 			continue;			/* next segment */
1167 		}
1168 
1169 		/*
1170 		 * Walk the pages of this segment.
1171 		 */
1172 		for(eva = va + NBSG; va < eva; va += PAGE_SIZE) {
1173 			pte = get_pte(va);
1174 
1175 			if ((pte & (PG_VALID | PG_TYPE)) ==
1176 				(PG_VALID | pgtype) &&
1177 			    PG_PFNUM(pte) == pf)
1178 			{
1179 				/*
1180 				 * Found the PROM mapping.
1181 				 * note: preserve page offset
1182 				 */
1183 				*vap = (va | ((vaddr_t)pa & PGOFSET));
1184 				restore_context(saved_ctx);
1185 				return 0;
1186 			}
1187 		}
1188 	}
1189 	restore_context(saved_ctx);
1190 	return ENOENT;
1191 }
1192