xref: /onnv-gate/usr/src/uts/sun4/vm/sfmmu.c (revision 3764:74844940a161)
10Sstevel@tonic-gate /*
20Sstevel@tonic-gate  * CDDL HEADER START
30Sstevel@tonic-gate  *
40Sstevel@tonic-gate  * The contents of this file are subject to the terms of the
52241Shuah  * Common Development and Distribution License (the "License").
62241Shuah  * You may not use this file except in compliance with the License.
70Sstevel@tonic-gate  *
80Sstevel@tonic-gate  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
90Sstevel@tonic-gate  * or http://www.opensolaris.org/os/licensing.
100Sstevel@tonic-gate  * See the License for the specific language governing permissions
110Sstevel@tonic-gate  * and limitations under the License.
120Sstevel@tonic-gate  *
130Sstevel@tonic-gate  * When distributing Covered Code, include this CDDL HEADER in each
140Sstevel@tonic-gate  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
150Sstevel@tonic-gate  * If applicable, add the following below this CDDL HEADER, with the
160Sstevel@tonic-gate  * fields enclosed by brackets "[]" replaced with your own identifying
170Sstevel@tonic-gate  * information: Portions Copyright [yyyy] [name of copyright owner]
180Sstevel@tonic-gate  *
190Sstevel@tonic-gate  * CDDL HEADER END
200Sstevel@tonic-gate  */
210Sstevel@tonic-gate /*
22*3764Sdp78419  * Copyright 2007 Sun Microsystems, Inc.  All rights reserved.
230Sstevel@tonic-gate  * Use is subject to license terms.
240Sstevel@tonic-gate  */
250Sstevel@tonic-gate 
260Sstevel@tonic-gate #pragma ident	"%Z%%M%	%I%	%E% SMI"
270Sstevel@tonic-gate 
280Sstevel@tonic-gate #include <sys/types.h>
290Sstevel@tonic-gate #include <vm/hat.h>
300Sstevel@tonic-gate #include <vm/hat_sfmmu.h>
310Sstevel@tonic-gate #include <vm/page.h>
320Sstevel@tonic-gate #include <sys/pte.h>
330Sstevel@tonic-gate #include <sys/systm.h>
340Sstevel@tonic-gate #include <sys/mman.h>
350Sstevel@tonic-gate #include <sys/sysmacros.h>
360Sstevel@tonic-gate #include <sys/machparam.h>
370Sstevel@tonic-gate #include <sys/vtrace.h>
380Sstevel@tonic-gate #include <sys/kmem.h>
390Sstevel@tonic-gate #include <sys/mmu.h>
400Sstevel@tonic-gate #include <sys/cmn_err.h>
410Sstevel@tonic-gate #include <sys/cpu.h>
420Sstevel@tonic-gate #include <sys/cpuvar.h>
430Sstevel@tonic-gate #include <sys/debug.h>
440Sstevel@tonic-gate #include <sys/lgrp.h>
450Sstevel@tonic-gate #include <sys/archsystm.h>
460Sstevel@tonic-gate #include <sys/machsystm.h>
470Sstevel@tonic-gate #include <sys/vmsystm.h>
480Sstevel@tonic-gate #include <sys/bitmap.h>
490Sstevel@tonic-gate #include <vm/as.h>
500Sstevel@tonic-gate #include <vm/seg.h>
510Sstevel@tonic-gate #include <vm/seg_kmem.h>
520Sstevel@tonic-gate #include <vm/seg_kp.h>
530Sstevel@tonic-gate #include <vm/seg_kpm.h>
540Sstevel@tonic-gate #include <vm/rm.h>
550Sstevel@tonic-gate #include <vm/vm_dep.h>
560Sstevel@tonic-gate #include <sys/t_lock.h>
570Sstevel@tonic-gate #include <sys/vm_machparam.h>
580Sstevel@tonic-gate #include <sys/promif.h>
590Sstevel@tonic-gate #include <sys/prom_isa.h>
600Sstevel@tonic-gate #include <sys/prom_plat.h>
610Sstevel@tonic-gate #include <sys/prom_debug.h>
620Sstevel@tonic-gate #include <sys/privregs.h>
630Sstevel@tonic-gate #include <sys/bootconf.h>
640Sstevel@tonic-gate #include <sys/memlist.h>
650Sstevel@tonic-gate #include <sys/memlist_plat.h>
660Sstevel@tonic-gate #include <sys/cpu_module.h>
670Sstevel@tonic-gate #include <sys/reboot.h>
680Sstevel@tonic-gate #include <sys/kdi.h>
690Sstevel@tonic-gate 
700Sstevel@tonic-gate /*
710Sstevel@tonic-gate  * Static routines
720Sstevel@tonic-gate  */
730Sstevel@tonic-gate static void	sfmmu_map_prom_mappings(struct translation *, size_t);
740Sstevel@tonic-gate static struct translation *read_prom_mappings(size_t *);
750Sstevel@tonic-gate static void	sfmmu_reloc_trap_handler(void *, void *, size_t);
760Sstevel@tonic-gate 
770Sstevel@tonic-gate /*
780Sstevel@tonic-gate  * External routines
790Sstevel@tonic-gate  */
800Sstevel@tonic-gate extern void sfmmu_remap_kernel(void);
810Sstevel@tonic-gate extern void sfmmu_patch_utsb(void);
820Sstevel@tonic-gate 
830Sstevel@tonic-gate /*
840Sstevel@tonic-gate  * Global Data:
850Sstevel@tonic-gate  */
860Sstevel@tonic-gate extern caddr_t	textva, datava;
870Sstevel@tonic-gate extern tte_t	ktext_tte, kdata_tte;	/* ttes for kernel text and data */
880Sstevel@tonic-gate extern int	enable_bigktsb;
890Sstevel@tonic-gate 
900Sstevel@tonic-gate uint64_t memsegspa = (uintptr_t)MSEG_NULLPTR_PA; /* memsegs physical linkage */
910Sstevel@tonic-gate uint64_t memseg_phash[N_MEM_SLOTS];	/* use physical memseg addresses */
920Sstevel@tonic-gate 
930Sstevel@tonic-gate int	sfmmu_kern_mapped = 0;
940Sstevel@tonic-gate 
950Sstevel@tonic-gate /*
960Sstevel@tonic-gate  * DMMU primary context register for the kernel context. Machine specific code
970Sstevel@tonic-gate  * inserts correct page size codes when necessary
980Sstevel@tonic-gate  */
990Sstevel@tonic-gate uint64_t kcontextreg = KCONTEXT;
1000Sstevel@tonic-gate 
1010Sstevel@tonic-gate /* Extern Global Data */
1020Sstevel@tonic-gate 
1030Sstevel@tonic-gate extern int page_relocate_ready;
1040Sstevel@tonic-gate 
1050Sstevel@tonic-gate /*
1060Sstevel@tonic-gate  * Controls the logic which enables the use of the
1070Sstevel@tonic-gate  * QUAD_LDD_PHYS ASI for TSB accesses.
1080Sstevel@tonic-gate  */
1090Sstevel@tonic-gate extern int	ktsb_phys;
1100Sstevel@tonic-gate 
1110Sstevel@tonic-gate /*
1120Sstevel@tonic-gate  * Global Routines called from within:
1130Sstevel@tonic-gate  *	usr/src/uts/sun4u
1140Sstevel@tonic-gate  *	usr/src/uts/sfmmu
1150Sstevel@tonic-gate  *	usr/src/uts/sun
1160Sstevel@tonic-gate  */
1170Sstevel@tonic-gate 
1180Sstevel@tonic-gate pfn_t
1190Sstevel@tonic-gate va_to_pfn(void *vaddr)
1200Sstevel@tonic-gate {
1210Sstevel@tonic-gate 	u_longlong_t physaddr;
1220Sstevel@tonic-gate 	int mode, valid;
1230Sstevel@tonic-gate 
1240Sstevel@tonic-gate 	if (tba_taken_over)
1250Sstevel@tonic-gate 		return (hat_getpfnum(kas.a_hat, (caddr_t)vaddr));
1260Sstevel@tonic-gate 
127*3764Sdp78419 #if !defined(C_OBP)
128*3764Sdp78419 	if ((caddr_t)vaddr >= kmem64_base && (caddr_t)vaddr < kmem64_end) {
129*3764Sdp78419 		if (kmem64_pabase == (uint64_t)-1)
130*3764Sdp78419 			prom_panic("va_to_pfn: kmem64_pabase not init");
131*3764Sdp78419 		physaddr = kmem64_pabase + ((caddr_t)vaddr - kmem64_base);
132*3764Sdp78419 		return ((pfn_t)physaddr >> MMU_PAGESHIFT);
133*3764Sdp78419 	}
134*3764Sdp78419 #endif	/* !C_OBP */
135*3764Sdp78419 
1360Sstevel@tonic-gate 	if ((prom_translate_virt(vaddr, &valid, &physaddr, &mode) != -1) &&
1370Sstevel@tonic-gate 	    (valid == -1)) {
1380Sstevel@tonic-gate 		return ((pfn_t)(physaddr >> MMU_PAGESHIFT));
1390Sstevel@tonic-gate 	}
1400Sstevel@tonic-gate 	return (PFN_INVALID);
1410Sstevel@tonic-gate }
1420Sstevel@tonic-gate 
1430Sstevel@tonic-gate uint64_t
1440Sstevel@tonic-gate va_to_pa(void *vaddr)
1450Sstevel@tonic-gate {
1460Sstevel@tonic-gate 	pfn_t pfn;
1470Sstevel@tonic-gate 
1480Sstevel@tonic-gate 	if ((pfn = va_to_pfn(vaddr)) == PFN_INVALID)
1490Sstevel@tonic-gate 		return ((uint64_t)-1);
1500Sstevel@tonic-gate 	return (((uint64_t)pfn << MMU_PAGESHIFT) |
1510Sstevel@tonic-gate 		((uint64_t)vaddr & MMU_PAGEOFFSET));
1520Sstevel@tonic-gate }
1530Sstevel@tonic-gate 
1540Sstevel@tonic-gate void
1550Sstevel@tonic-gate hat_kern_setup(void)
1560Sstevel@tonic-gate {
1570Sstevel@tonic-gate 	struct translation *trans_root;
1580Sstevel@tonic-gate 	size_t ntrans_root;
1590Sstevel@tonic-gate 	extern void startup_fixup_physavail(void);
1600Sstevel@tonic-gate 
1610Sstevel@tonic-gate 	/*
1620Sstevel@tonic-gate 	 * These are the steps we take to take over the mmu from the prom.
1630Sstevel@tonic-gate 	 *
1640Sstevel@tonic-gate 	 * (1)	Read the prom's mappings through the translation property.
1650Sstevel@tonic-gate 	 * (2)	Remap the kernel text and kernel data with 2 locked 4MB ttes.
1660Sstevel@tonic-gate 	 *	Create the the hmeblks for these 2 ttes at this time.
1670Sstevel@tonic-gate 	 * (3)	Create hat structures for all other prom mappings.  Since the
1680Sstevel@tonic-gate 	 *	kernel text and data hme_blks have already been created we
1690Sstevel@tonic-gate 	 *	skip the equivalent prom's mappings.
1700Sstevel@tonic-gate 	 * (4)	Initialize the tsb and its corresponding hardware regs.
1710Sstevel@tonic-gate 	 * (5)	Take over the trap table (currently in startup).
1720Sstevel@tonic-gate 	 * (6)	Up to this point it is possible the prom required some of its
1730Sstevel@tonic-gate 	 *	locked tte's.  Now that we own the trap table we remove them.
1740Sstevel@tonic-gate 	 */
1750Sstevel@tonic-gate 
1760Sstevel@tonic-gate 	ktsb_pbase = va_to_pa(ktsb_base);
1770Sstevel@tonic-gate 	ktsb4m_pbase = va_to_pa(ktsb4m_base);
1780Sstevel@tonic-gate 	PRM_DEBUG(ktsb_pbase);
1790Sstevel@tonic-gate 	PRM_DEBUG(ktsb4m_pbase);
1800Sstevel@tonic-gate 
1810Sstevel@tonic-gate 	sfmmu_patch_ktsb();
1820Sstevel@tonic-gate 	sfmmu_patch_utsb();
1830Sstevel@tonic-gate 	sfmmu_patch_mmu_asi(ktsb_phys);
1840Sstevel@tonic-gate 
1850Sstevel@tonic-gate 	sfmmu_init_tsbs();
1860Sstevel@tonic-gate 
1870Sstevel@tonic-gate 	if (kpm_enable) {
1880Sstevel@tonic-gate 		sfmmu_kpm_patch_tlbm();
1890Sstevel@tonic-gate 		if (kpm_smallpages == 0) {
1900Sstevel@tonic-gate 			sfmmu_kpm_patch_tsbm();
1910Sstevel@tonic-gate 		}
1920Sstevel@tonic-gate 	}
1930Sstevel@tonic-gate 
1940Sstevel@tonic-gate 	/*
1950Sstevel@tonic-gate 	 * The 8K-indexed kernel TSB space is used to hold
1960Sstevel@tonic-gate 	 * translations below...
1970Sstevel@tonic-gate 	 */
1980Sstevel@tonic-gate 	trans_root = read_prom_mappings(&ntrans_root);
1990Sstevel@tonic-gate 	sfmmu_remap_kernel();
2000Sstevel@tonic-gate 	startup_fixup_physavail();
2010Sstevel@tonic-gate 	mmu_init_kernel_pgsz(kas.a_hat);
2020Sstevel@tonic-gate 	sfmmu_map_prom_mappings(trans_root, ntrans_root);
2030Sstevel@tonic-gate 
2040Sstevel@tonic-gate 	/*
2050Sstevel@tonic-gate 	 * We invalidate 8K kernel TSB because we used it in
2060Sstevel@tonic-gate 	 * sfmmu_map_prom_mappings()
2070Sstevel@tonic-gate 	 */
2080Sstevel@tonic-gate 	sfmmu_inv_tsb(ktsb_base, ktsb_sz);
2090Sstevel@tonic-gate 	sfmmu_inv_tsb(ktsb4m_base, ktsb4m_sz);
2100Sstevel@tonic-gate 
2110Sstevel@tonic-gate 	sfmmu_init_ktsbinfo();
2120Sstevel@tonic-gate 
2130Sstevel@tonic-gate 
2140Sstevel@tonic-gate 	sfmmu_kern_mapped = 1;
2150Sstevel@tonic-gate 
2160Sstevel@tonic-gate 	/*
2170Sstevel@tonic-gate 	 * hments have been created for mapped pages, and thus we're ready
2180Sstevel@tonic-gate 	 * for kmdb to start using its own trap table.  It walks the hments
2190Sstevel@tonic-gate 	 * to resolve TLB misses, and can't be used until they're ready.
2200Sstevel@tonic-gate 	 */
2210Sstevel@tonic-gate 	if (boothowto & RB_DEBUG)
2220Sstevel@tonic-gate 		kdi_dvec_vmready();
2230Sstevel@tonic-gate }
2240Sstevel@tonic-gate 
2250Sstevel@tonic-gate /*
2260Sstevel@tonic-gate  * Macro used below to convert the prom's 32-bit high and low fields into
2270Sstevel@tonic-gate  * a value appropriate for the 64-bit kernel.
2280Sstevel@tonic-gate  */
2290Sstevel@tonic-gate 
2300Sstevel@tonic-gate #define	COMBINE(hi, lo) (((uint64_t)(uint32_t)(hi) << 32) | (uint32_t)(lo))
2310Sstevel@tonic-gate 
2320Sstevel@tonic-gate /*
233*3764Sdp78419  * Track larges pages used.
234*3764Sdp78419  * Provides observability for this feature on non-debug kernels.
235*3764Sdp78419  */
236*3764Sdp78419 ulong_t map_prom_lpcount[MMU_PAGE_SIZES];
237*3764Sdp78419 
238*3764Sdp78419 /*
2390Sstevel@tonic-gate  * This function traverses the prom mapping list and creates equivalent
2400Sstevel@tonic-gate  * mappings in the sfmmu mapping hash.
2410Sstevel@tonic-gate  */
2420Sstevel@tonic-gate static void
2430Sstevel@tonic-gate sfmmu_map_prom_mappings(struct translation *trans_root, size_t ntrans_root)
2440Sstevel@tonic-gate {
2450Sstevel@tonic-gate 	struct translation *promt;
2460Sstevel@tonic-gate 	tte_t	tte, oldtte, *ttep;
2470Sstevel@tonic-gate 	pfn_t	pfn, oldpfn, basepfn;
2480Sstevel@tonic-gate 	caddr_t vaddr;
2490Sstevel@tonic-gate 	size_t	size, offset;
2500Sstevel@tonic-gate 	unsigned long i;
2510Sstevel@tonic-gate 	uint_t	attr;
2520Sstevel@tonic-gate 	page_t *pp;
2530Sstevel@tonic-gate 	extern struct memlist *virt_avail;
2540Sstevel@tonic-gate 
2550Sstevel@tonic-gate 	ttep = &tte;
2560Sstevel@tonic-gate 	for (i = 0, promt = trans_root; i < ntrans_root; i++, promt++) {
2570Sstevel@tonic-gate 		ASSERT(promt->tte_hi != 0);
2580Sstevel@tonic-gate 		ASSERT32(promt->virt_hi == 0 && promt->size_hi == 0);
2590Sstevel@tonic-gate 
260*3764Sdp78419 		vaddr = (caddr_t)COMBINE(promt->virt_hi, promt->virt_lo);
261*3764Sdp78419 
2620Sstevel@tonic-gate 		/*
2630Sstevel@tonic-gate 		 * hack until we get rid of map-for-unix
2640Sstevel@tonic-gate 		 */
265*3764Sdp78419 		if (vaddr < (caddr_t)KERNELBASE)
2660Sstevel@tonic-gate 			continue;
2670Sstevel@tonic-gate 
2680Sstevel@tonic-gate 		ttep->tte_inthi = promt->tte_hi;
2690Sstevel@tonic-gate 		ttep->tte_intlo = promt->tte_lo;
2700Sstevel@tonic-gate 		attr = PROC_DATA | HAT_NOSYNC;
2710Sstevel@tonic-gate #if defined(TTE_IS_GLOBAL)
2720Sstevel@tonic-gate 		if (TTE_IS_GLOBAL(ttep)) {
2730Sstevel@tonic-gate 			/*
2740Sstevel@tonic-gate 			 * The prom better not use global translations
2750Sstevel@tonic-gate 			 * because a user process might use the same
2760Sstevel@tonic-gate 			 * virtual addresses
2770Sstevel@tonic-gate 			 */
2780Sstevel@tonic-gate 			cmn_err(CE_PANIC, "map_prom: global translation");
2790Sstevel@tonic-gate 			TTE_SET_LOFLAGS(ttep, TTE_GLB_INT, 0);
2800Sstevel@tonic-gate 		}
2810Sstevel@tonic-gate #endif
2820Sstevel@tonic-gate 		if (TTE_IS_LOCKED(ttep)) {
2830Sstevel@tonic-gate 			/* clear the lock bits */
2840Sstevel@tonic-gate 			TTE_CLR_LOCKED(ttep);
2850Sstevel@tonic-gate 		}
2860Sstevel@tonic-gate 		attr |= (TTE_IS_VCACHEABLE(ttep)) ? 0 : SFMMU_UNCACHEVTTE;
2870Sstevel@tonic-gate 		attr |= (TTE_IS_PCACHEABLE(ttep)) ? 0 : SFMMU_UNCACHEPTTE;
2880Sstevel@tonic-gate 		attr |= (TTE_IS_SIDEFFECT(ttep)) ? SFMMU_SIDEFFECT : 0;
2890Sstevel@tonic-gate 		attr |= (TTE_IS_IE(ttep)) ? HAT_STRUCTURE_LE : 0;
2900Sstevel@tonic-gate 
2910Sstevel@tonic-gate 		size = COMBINE(promt->size_hi, promt->size_lo);
2920Sstevel@tonic-gate 		offset = 0;
2930Sstevel@tonic-gate 		basepfn = TTE_TO_PFN((caddr_t)COMBINE(promt->virt_hi,
2940Sstevel@tonic-gate 		    promt->virt_lo), ttep);
2950Sstevel@tonic-gate 		while (size) {
2960Sstevel@tonic-gate 			vaddr = (caddr_t)(COMBINE(promt->virt_hi,
2970Sstevel@tonic-gate 			    promt->virt_lo) + offset);
2980Sstevel@tonic-gate 
2990Sstevel@tonic-gate 			/*
3000Sstevel@tonic-gate 			 * make sure address is not in virt-avail list
3010Sstevel@tonic-gate 			 */
3020Sstevel@tonic-gate 			if (address_in_memlist(virt_avail, (uint64_t)vaddr,
3030Sstevel@tonic-gate 			    size)) {
3040Sstevel@tonic-gate 				cmn_err(CE_PANIC, "map_prom: inconsistent "
3050Sstevel@tonic-gate 				    "translation/avail lists");
3060Sstevel@tonic-gate 			}
3070Sstevel@tonic-gate 
3080Sstevel@tonic-gate 			pfn = basepfn + mmu_btop(offset);
3090Sstevel@tonic-gate 			if (pf_is_memory(pfn)) {
3100Sstevel@tonic-gate 				if (attr & SFMMU_UNCACHEPTTE) {
3110Sstevel@tonic-gate 					cmn_err(CE_PANIC, "map_prom: "
3120Sstevel@tonic-gate 					    "uncached prom memory page");
3130Sstevel@tonic-gate 				}
3140Sstevel@tonic-gate 			} else {
3150Sstevel@tonic-gate 				if (!(attr & SFMMU_SIDEFFECT)) {
3160Sstevel@tonic-gate 					cmn_err(CE_PANIC, "map_prom: prom "
3170Sstevel@tonic-gate 					    "i/o page without side-effect");
3180Sstevel@tonic-gate 				}
3190Sstevel@tonic-gate 			}
320*3764Sdp78419 
321*3764Sdp78419 			/*
322*3764Sdp78419 			 * skip kmem64 area
323*3764Sdp78419 			 */
324*3764Sdp78419 			if (vaddr >= kmem64_base &&
325*3764Sdp78419 			    vaddr < kmem64_aligned_end) {
326*3764Sdp78419 #if !defined(C_OBP)
327*3764Sdp78419 				cmn_err(CE_PANIC,
328*3764Sdp78419 				    "unexpected kmem64 prom mapping\n");
329*3764Sdp78419 #else	/* !C_OBP */
330*3764Sdp78419 				size_t mapsz;
331*3764Sdp78419 
332*3764Sdp78419 				if (ptob(pfn) !=
333*3764Sdp78419 				    kmem64_pabase + (vaddr - kmem64_base)) {
334*3764Sdp78419 					cmn_err(CE_PANIC,
335*3764Sdp78419 					    "unexpected kmem64 prom mapping\n");
336*3764Sdp78419 				}
337*3764Sdp78419 
338*3764Sdp78419 				mapsz = kmem64_aligned_end - vaddr;
339*3764Sdp78419 				if (mapsz >= size) {
340*3764Sdp78419 					break;
341*3764Sdp78419 				}
342*3764Sdp78419 				size -= mapsz;
343*3764Sdp78419 				offset += mapsz;
344*3764Sdp78419 				continue;
345*3764Sdp78419 #endif	/* !C_OBP */
346*3764Sdp78419 			}
347*3764Sdp78419 
3480Sstevel@tonic-gate 			oldpfn = sfmmu_vatopfn(vaddr, KHATID, &oldtte);
3490Sstevel@tonic-gate 			ASSERT(oldpfn != PFN_SUSPENDED);
3500Sstevel@tonic-gate 			ASSERT(page_relocate_ready == 0);
3510Sstevel@tonic-gate 
3520Sstevel@tonic-gate 			if (oldpfn != PFN_INVALID) {
3530Sstevel@tonic-gate 				/*
3540Sstevel@tonic-gate 				 * mapping already exists.
3550Sstevel@tonic-gate 				 * Verify they are equal
3560Sstevel@tonic-gate 				 */
3570Sstevel@tonic-gate 				if (pfn != oldpfn) {
3580Sstevel@tonic-gate 					cmn_err(CE_PANIC, "map_prom: mapping "
3590Sstevel@tonic-gate 					    "conflict (va=0x%p pfn=%p, "
3600Sstevel@tonic-gate 					    "oldpfn=%p)",
3610Sstevel@tonic-gate 					    (void *)vaddr, (void *)pfn,
3620Sstevel@tonic-gate 					    (void *)oldpfn);
3630Sstevel@tonic-gate 				}
3640Sstevel@tonic-gate 				size -= MMU_PAGESIZE;
3650Sstevel@tonic-gate 				offset += MMU_PAGESIZE;
3660Sstevel@tonic-gate 				continue;
3670Sstevel@tonic-gate 			}
3680Sstevel@tonic-gate 
3690Sstevel@tonic-gate 			pp = page_numtopp_nolock(pfn);
3700Sstevel@tonic-gate 			if ((pp != NULL) && PP_ISFREE((page_t *)pp)) {
3710Sstevel@tonic-gate 				cmn_err(CE_PANIC, "map_prom: "
3720Sstevel@tonic-gate 				    "prom-mapped page (va 0x%p, pfn 0x%p) "
3730Sstevel@tonic-gate 				    "on free list", (void *)vaddr, (void *)pfn);
3740Sstevel@tonic-gate 			}
3750Sstevel@tonic-gate 
3760Sstevel@tonic-gate 			sfmmu_memtte(ttep, pfn, attr, TTE8K);
3770Sstevel@tonic-gate 			sfmmu_tteload(kas.a_hat, ttep, vaddr, pp,
3780Sstevel@tonic-gate 			    HAT_LOAD_LOCK | SFMMU_NO_TSBLOAD);
3790Sstevel@tonic-gate 			size -= MMU_PAGESIZE;
3800Sstevel@tonic-gate 			offset += MMU_PAGESIZE;
3810Sstevel@tonic-gate 		}
3820Sstevel@tonic-gate 	}
383*3764Sdp78419 
384*3764Sdp78419 	/*
385*3764Sdp78419 	 * We claimed kmem64 from prom, so now we need to load tte.
386*3764Sdp78419 	 */
387*3764Sdp78419 	if (kmem64_base != NULL) {
388*3764Sdp78419 		pgcnt_t pages;
389*3764Sdp78419 		size_t psize;
390*3764Sdp78419 		int pszc;
391*3764Sdp78419 
392*3764Sdp78419 		pszc = kmem64_szc;
393*3764Sdp78419 #ifdef sun4u
394*3764Sdp78419 		if (pszc > TTE8K) {
395*3764Sdp78419 			pszc = segkmem_lpszc;
396*3764Sdp78419 		}
397*3764Sdp78419 #endif	/* sun4u */
398*3764Sdp78419 		psize = TTEBYTES(pszc);
399*3764Sdp78419 		pages = btop(psize);
400*3764Sdp78419 		basepfn = kmem64_pabase >> MMU_PAGESHIFT;
401*3764Sdp78419 		vaddr = kmem64_base;
402*3764Sdp78419 		while (vaddr < kmem64_end) {
403*3764Sdp78419 			sfmmu_memtte(ttep, basepfn,
404*3764Sdp78419 			    PROC_DATA | HAT_NOSYNC, pszc);
405*3764Sdp78419 			sfmmu_tteload(kas.a_hat, ttep, vaddr, NULL,
406*3764Sdp78419 			    HAT_LOAD_LOCK | SFMMU_NO_TSBLOAD);
407*3764Sdp78419 			vaddr += psize;
408*3764Sdp78419 			basepfn += pages;
409*3764Sdp78419 		}
410*3764Sdp78419 		map_prom_lpcount[pszc] =
411*3764Sdp78419 		    ((caddr_t)P2ROUNDUP((uintptr_t)kmem64_end, psize) -
412*3764Sdp78419 			kmem64_base) >> TTE_PAGE_SHIFT(pszc);
413*3764Sdp78419 	}
4140Sstevel@tonic-gate }
4150Sstevel@tonic-gate 
4160Sstevel@tonic-gate #undef COMBINE	/* local to previous routine */
4170Sstevel@tonic-gate 
4180Sstevel@tonic-gate /*
4190Sstevel@tonic-gate  * This routine reads in the "translations" property in to a buffer and
4200Sstevel@tonic-gate  * returns a pointer to this buffer and the number of translations.
4210Sstevel@tonic-gate  */
4220Sstevel@tonic-gate static struct translation *
4230Sstevel@tonic-gate read_prom_mappings(size_t *ntransrootp)
4240Sstevel@tonic-gate {
4250Sstevel@tonic-gate 	char *prop = "translations";
4260Sstevel@tonic-gate 	size_t translen;
427789Sahrens 	pnode_t node;
4280Sstevel@tonic-gate 	struct translation *transroot;
4290Sstevel@tonic-gate 
4300Sstevel@tonic-gate 	/*
4310Sstevel@tonic-gate 	 * the "translations" property is associated with the mmu node
4320Sstevel@tonic-gate 	 */
433789Sahrens 	node = (pnode_t)prom_getphandle(prom_mmu_ihandle());
4340Sstevel@tonic-gate 
4350Sstevel@tonic-gate 	/*
4360Sstevel@tonic-gate 	 * We use the TSB space to read in the prom mappings.  This space
4370Sstevel@tonic-gate 	 * is currently not being used because we haven't taken over the
4380Sstevel@tonic-gate 	 * trap table yet.  It should be big enough to hold the mappings.
4390Sstevel@tonic-gate 	 */
4400Sstevel@tonic-gate 	if ((translen = prom_getproplen(node, prop)) == -1)
4410Sstevel@tonic-gate 		cmn_err(CE_PANIC, "no translations property");
4420Sstevel@tonic-gate 	*ntransrootp = translen / sizeof (*transroot);
4430Sstevel@tonic-gate 	translen = roundup(translen, MMU_PAGESIZE);
4440Sstevel@tonic-gate 	PRM_DEBUG(translen);
4450Sstevel@tonic-gate 	if (translen > TSB_BYTES(ktsb_szcode))
4460Sstevel@tonic-gate 		cmn_err(CE_PANIC, "not enough space for translations");
4470Sstevel@tonic-gate 
4480Sstevel@tonic-gate 	transroot = (struct translation *)ktsb_base;
4490Sstevel@tonic-gate 	ASSERT(transroot);
4500Sstevel@tonic-gate 	if (prom_getprop(node, prop, (caddr_t)transroot) == -1) {
4510Sstevel@tonic-gate 		cmn_err(CE_PANIC, "translations getprop failed");
4520Sstevel@tonic-gate 	}
4530Sstevel@tonic-gate 	return (transroot);
4540Sstevel@tonic-gate }
4550Sstevel@tonic-gate 
4560Sstevel@tonic-gate /*
4570Sstevel@tonic-gate  * Init routine of the nucleus data memory allocator.
4580Sstevel@tonic-gate  *
4590Sstevel@tonic-gate  * The nucleus data memory allocator is organized in ecache_alignsize'd
4600Sstevel@tonic-gate  * memory chunks. Memory allocated by ndata_alloc() will never be freed.
4610Sstevel@tonic-gate  *
4620Sstevel@tonic-gate  * The ndata argument is used as header of the ndata freelist.
4630Sstevel@tonic-gate  * Other freelist nodes are placed in the nucleus memory itself
4640Sstevel@tonic-gate  * at the beginning of a free memory chunk. Therefore a freelist
4650Sstevel@tonic-gate  * node (struct memlist) must fit into the smallest allocatable
4660Sstevel@tonic-gate  * memory chunk (ecache_alignsize bytes).
4670Sstevel@tonic-gate  *
4680Sstevel@tonic-gate  * The memory interval [base, end] passed to ndata_alloc_init() must be
4690Sstevel@tonic-gate  * bzero'd to allow the allocator to return bzero'd memory easily.
4700Sstevel@tonic-gate  */
4710Sstevel@tonic-gate void
4720Sstevel@tonic-gate ndata_alloc_init(struct memlist *ndata, uintptr_t base, uintptr_t end)
4730Sstevel@tonic-gate {
4740Sstevel@tonic-gate 	ASSERT(sizeof (struct memlist) <= ecache_alignsize);
4750Sstevel@tonic-gate 
4760Sstevel@tonic-gate 	base = roundup(base, ecache_alignsize);
4770Sstevel@tonic-gate 	end = end - end % ecache_alignsize;
4780Sstevel@tonic-gate 
4790Sstevel@tonic-gate 	ASSERT(base < end);
4800Sstevel@tonic-gate 
4810Sstevel@tonic-gate 	ndata->address = base;
4820Sstevel@tonic-gate 	ndata->size = end - base;
4830Sstevel@tonic-gate 	ndata->next = NULL;
4840Sstevel@tonic-gate 	ndata->prev = NULL;
4850Sstevel@tonic-gate }
4860Sstevel@tonic-gate 
4870Sstevel@tonic-gate /*
4880Sstevel@tonic-gate  * Deliver the size of the largest free memory chunk.
4890Sstevel@tonic-gate  */
4900Sstevel@tonic-gate size_t
4910Sstevel@tonic-gate ndata_maxsize(struct memlist *ndata)
4920Sstevel@tonic-gate {
4930Sstevel@tonic-gate 	size_t chunksize = ndata->size;
4940Sstevel@tonic-gate 
4950Sstevel@tonic-gate 	while ((ndata = ndata->next) != NULL) {
4960Sstevel@tonic-gate 		if (chunksize < ndata->size)
4970Sstevel@tonic-gate 			chunksize = ndata->size;
4980Sstevel@tonic-gate 	}
4990Sstevel@tonic-gate 
5000Sstevel@tonic-gate 	return (chunksize);
5010Sstevel@tonic-gate }
5020Sstevel@tonic-gate 
5030Sstevel@tonic-gate /*
5040Sstevel@tonic-gate  * This is a special function to figure out if the memory chunk needed
5050Sstevel@tonic-gate  * for the page structs can fit in the nucleus or not. If it fits the
5060Sstevel@tonic-gate  * function calculates and returns the possible remaining ndata size
5070Sstevel@tonic-gate  * in the last element if the size needed for page structs would be
5080Sstevel@tonic-gate  * allocated from the nucleus.
5090Sstevel@tonic-gate  */
5100Sstevel@tonic-gate size_t
5110Sstevel@tonic-gate ndata_spare(struct memlist *ndata, size_t wanted, size_t alignment)
5120Sstevel@tonic-gate {
5130Sstevel@tonic-gate 	struct memlist *frlist;
5140Sstevel@tonic-gate 	uintptr_t base;
5150Sstevel@tonic-gate 	uintptr_t end;
5160Sstevel@tonic-gate 
5170Sstevel@tonic-gate 	for (frlist = ndata; frlist != NULL; frlist = frlist->next) {
5180Sstevel@tonic-gate 		base = roundup(frlist->address, alignment);
5190Sstevel@tonic-gate 		end = roundup(base + wanted, ecache_alignsize);
5200Sstevel@tonic-gate 
5210Sstevel@tonic-gate 		if (end <= frlist->address + frlist->size) {
5220Sstevel@tonic-gate 			if (frlist->next == NULL)
5230Sstevel@tonic-gate 				return (frlist->address + frlist->size - end);
5240Sstevel@tonic-gate 
5250Sstevel@tonic-gate 			while (frlist->next != NULL)
5260Sstevel@tonic-gate 				frlist = frlist->next;
5270Sstevel@tonic-gate 
5280Sstevel@tonic-gate 			return (frlist->size);
5290Sstevel@tonic-gate 		}
5300Sstevel@tonic-gate 	}
5310Sstevel@tonic-gate 
5320Sstevel@tonic-gate 	return (0);
5330Sstevel@tonic-gate }
5340Sstevel@tonic-gate 
5350Sstevel@tonic-gate /*
5360Sstevel@tonic-gate  * Allocate the last properly aligned memory chunk.
5370Sstevel@tonic-gate  * This function is called when no more large nucleus memory chunks
5380Sstevel@tonic-gate  * will be allocated.  The remaining free nucleus memory at the end
5390Sstevel@tonic-gate  * of the nucleus can be added to the phys_avail list.
5400Sstevel@tonic-gate  */
5410Sstevel@tonic-gate void *
5420Sstevel@tonic-gate ndata_extra_base(struct memlist *ndata, size_t alignment)
5430Sstevel@tonic-gate {
5440Sstevel@tonic-gate 	uintptr_t base;
5450Sstevel@tonic-gate 	size_t wasteage = 0;
5460Sstevel@tonic-gate #ifdef	DEBUG
5470Sstevel@tonic-gate 	static int called = 0;
5480Sstevel@tonic-gate 
5490Sstevel@tonic-gate 	if (called++ > 0)
5500Sstevel@tonic-gate 		cmn_err(CE_PANIC, "ndata_extra_base() called more than once");
5510Sstevel@tonic-gate #endif /* DEBUG */
5520Sstevel@tonic-gate 
5530Sstevel@tonic-gate 	/*
5540Sstevel@tonic-gate 	 * The alignment needs to be a multiple of ecache_alignsize.
5550Sstevel@tonic-gate 	 */
5560Sstevel@tonic-gate 	ASSERT((alignment % ecache_alignsize) ==  0);
5570Sstevel@tonic-gate 
5580Sstevel@tonic-gate 	while (ndata->next != NULL) {
5590Sstevel@tonic-gate 		wasteage += ndata->size;
5600Sstevel@tonic-gate 		ndata = ndata->next;
5610Sstevel@tonic-gate 	}
5620Sstevel@tonic-gate 
5630Sstevel@tonic-gate 	base = roundup(ndata->address, alignment);
5640Sstevel@tonic-gate 
5650Sstevel@tonic-gate 	if (base >= ndata->address + ndata->size)
5660Sstevel@tonic-gate 		return (NULL);
5670Sstevel@tonic-gate 
5680Sstevel@tonic-gate 	if (base == ndata->address) {
5690Sstevel@tonic-gate 		if (ndata->prev != NULL)
5700Sstevel@tonic-gate 			ndata->prev->next = NULL;
5710Sstevel@tonic-gate 		else
5720Sstevel@tonic-gate 			ndata->size = 0;
5730Sstevel@tonic-gate 
5740Sstevel@tonic-gate 		bzero((void *)base, sizeof (struct memlist));
5750Sstevel@tonic-gate 
5760Sstevel@tonic-gate 	} else {
5770Sstevel@tonic-gate 		ndata->size = base - ndata->address;
5780Sstevel@tonic-gate 		wasteage += ndata->size;
5790Sstevel@tonic-gate 	}
5800Sstevel@tonic-gate 	PRM_DEBUG(wasteage);
5810Sstevel@tonic-gate 
5820Sstevel@tonic-gate 	return ((void *)base);
5830Sstevel@tonic-gate }
5840Sstevel@tonic-gate 
5850Sstevel@tonic-gate /*
5860Sstevel@tonic-gate  * Select the best matching buffer, avoid memory fragmentation.
5870Sstevel@tonic-gate  */
5880Sstevel@tonic-gate static struct memlist *
5890Sstevel@tonic-gate ndata_select_chunk(struct memlist *ndata, size_t wanted, size_t alignment)
5900Sstevel@tonic-gate {
5910Sstevel@tonic-gate 	struct memlist *fnd_below = NULL;
5920Sstevel@tonic-gate 	struct memlist *fnd_above = NULL;
5930Sstevel@tonic-gate 	struct memlist *fnd_unused = NULL;
5940Sstevel@tonic-gate 	struct memlist *frlist;
5950Sstevel@tonic-gate 	uintptr_t base;
5960Sstevel@tonic-gate 	uintptr_t end;
5970Sstevel@tonic-gate 	size_t below;
5980Sstevel@tonic-gate 	size_t above;
5990Sstevel@tonic-gate 	size_t unused;
6000Sstevel@tonic-gate 	size_t best_below = ULONG_MAX;
6010Sstevel@tonic-gate 	size_t best_above = ULONG_MAX;
6020Sstevel@tonic-gate 	size_t best_unused = ULONG_MAX;
6030Sstevel@tonic-gate 
6040Sstevel@tonic-gate 	ASSERT(ndata != NULL);
6050Sstevel@tonic-gate 
6060Sstevel@tonic-gate 	/*
6070Sstevel@tonic-gate 	 * Look for the best matching buffer, avoid memory fragmentation.
6080Sstevel@tonic-gate 	 * The following strategy is used, try to find
6090Sstevel@tonic-gate 	 *   1. an exact fitting buffer
6100Sstevel@tonic-gate 	 *   2. avoid wasting any space below the buffer, take first
6110Sstevel@tonic-gate 	 *	fitting buffer
6120Sstevel@tonic-gate 	 *   3. avoid wasting any space above the buffer, take first
6130Sstevel@tonic-gate 	 *	fitting buffer
6140Sstevel@tonic-gate 	 *   4. avoid wasting space, take first fitting buffer
6150Sstevel@tonic-gate 	 *   5. take the last buffer in chain
6160Sstevel@tonic-gate 	 */
6170Sstevel@tonic-gate 	for (frlist = ndata; frlist != NULL; frlist = frlist->next) {
6180Sstevel@tonic-gate 		base = roundup(frlist->address, alignment);
6190Sstevel@tonic-gate 		end = roundup(base + wanted, ecache_alignsize);
6200Sstevel@tonic-gate 
6210Sstevel@tonic-gate 		if (end > frlist->address + frlist->size)
6220Sstevel@tonic-gate 			continue;
6230Sstevel@tonic-gate 
6240Sstevel@tonic-gate 		below = (base - frlist->address) / ecache_alignsize;
6250Sstevel@tonic-gate 		above = (frlist->address + frlist->size - end) /
6260Sstevel@tonic-gate 		    ecache_alignsize;
6270Sstevel@tonic-gate 		unused = below + above;
6280Sstevel@tonic-gate 
6290Sstevel@tonic-gate 		if (unused == 0)
6300Sstevel@tonic-gate 			return (frlist);
6310Sstevel@tonic-gate 
6320Sstevel@tonic-gate 		if (frlist->next == NULL)
6330Sstevel@tonic-gate 			break;
6340Sstevel@tonic-gate 
6350Sstevel@tonic-gate 		if (below < best_below) {
6360Sstevel@tonic-gate 			best_below = below;
6370Sstevel@tonic-gate 			fnd_below = frlist;
6380Sstevel@tonic-gate 		}
6390Sstevel@tonic-gate 
6400Sstevel@tonic-gate 		if (above < best_above) {
6410Sstevel@tonic-gate 			best_above = above;
6420Sstevel@tonic-gate 			fnd_above = frlist;
6430Sstevel@tonic-gate 		}
6440Sstevel@tonic-gate 
6450Sstevel@tonic-gate 		if (unused < best_unused) {
6460Sstevel@tonic-gate 			best_unused = unused;
6470Sstevel@tonic-gate 			fnd_unused = frlist;
6480Sstevel@tonic-gate 		}
6490Sstevel@tonic-gate 	}
6500Sstevel@tonic-gate 
6510Sstevel@tonic-gate 	if (best_below == 0)
6520Sstevel@tonic-gate 		return (fnd_below);
6530Sstevel@tonic-gate 	if (best_above == 0)
6540Sstevel@tonic-gate 		return (fnd_above);
6550Sstevel@tonic-gate 	if (best_unused < ULONG_MAX)
6560Sstevel@tonic-gate 		return (fnd_unused);
6570Sstevel@tonic-gate 
6580Sstevel@tonic-gate 	return (frlist);
6590Sstevel@tonic-gate }
6600Sstevel@tonic-gate 
6610Sstevel@tonic-gate /*
6620Sstevel@tonic-gate  * Nucleus data memory allocator.
6630Sstevel@tonic-gate  * The granularity of the allocator is ecache_alignsize.
6640Sstevel@tonic-gate  * See also comment for ndata_alloc_init().
6650Sstevel@tonic-gate  */
6660Sstevel@tonic-gate void *
6670Sstevel@tonic-gate ndata_alloc(struct memlist *ndata, size_t wanted, size_t alignment)
6680Sstevel@tonic-gate {
6690Sstevel@tonic-gate 	struct memlist *found;
6700Sstevel@tonic-gate 	struct memlist *fnd_above;
6710Sstevel@tonic-gate 	uintptr_t base;
6720Sstevel@tonic-gate 	uintptr_t end;
6730Sstevel@tonic-gate 	size_t below;
6740Sstevel@tonic-gate 	size_t above;
6750Sstevel@tonic-gate 
6760Sstevel@tonic-gate 	/*
6770Sstevel@tonic-gate 	 * Look for the best matching buffer, avoid memory fragmentation.
6780Sstevel@tonic-gate 	 */
6790Sstevel@tonic-gate 	if ((found = ndata_select_chunk(ndata, wanted, alignment)) == NULL)
6800Sstevel@tonic-gate 		return (NULL);
6810Sstevel@tonic-gate 
6820Sstevel@tonic-gate 	/*
6830Sstevel@tonic-gate 	 * Allocate the nucleus data buffer.
6840Sstevel@tonic-gate 	 */
6850Sstevel@tonic-gate 	base = roundup(found->address, alignment);
6860Sstevel@tonic-gate 	end = roundup(base + wanted, ecache_alignsize);
6870Sstevel@tonic-gate 	ASSERT(end <= found->address + found->size);
6880Sstevel@tonic-gate 
6890Sstevel@tonic-gate 	below = base - found->address;
6900Sstevel@tonic-gate 	above = found->address + found->size - end;
6910Sstevel@tonic-gate 	ASSERT(above == 0 || (above % ecache_alignsize) == 0);
6920Sstevel@tonic-gate 
6930Sstevel@tonic-gate 	if (below >= ecache_alignsize) {
6940Sstevel@tonic-gate 		/*
6950Sstevel@tonic-gate 		 * There is free memory below the allocated memory chunk.
6960Sstevel@tonic-gate 		 */
6970Sstevel@tonic-gate 		found->size = below - below % ecache_alignsize;
6980Sstevel@tonic-gate 
6990Sstevel@tonic-gate 		if (above) {
7000Sstevel@tonic-gate 			fnd_above = (struct memlist *)end;
7010Sstevel@tonic-gate 			fnd_above->address = end;
7020Sstevel@tonic-gate 			fnd_above->size = above;
7030Sstevel@tonic-gate 
7040Sstevel@tonic-gate 			if ((fnd_above->next = found->next) != NULL)
7050Sstevel@tonic-gate 				found->next->prev = fnd_above;
7060Sstevel@tonic-gate 			fnd_above->prev = found;
7070Sstevel@tonic-gate 			found->next = fnd_above;
7080Sstevel@tonic-gate 		}
7090Sstevel@tonic-gate 
7100Sstevel@tonic-gate 		return ((void *)base);
7110Sstevel@tonic-gate 	}
7120Sstevel@tonic-gate 
7130Sstevel@tonic-gate 	if (found->prev == NULL) {
7140Sstevel@tonic-gate 		/*
7150Sstevel@tonic-gate 		 * The first chunk (ndata) is selected.
7160Sstevel@tonic-gate 		 */
7170Sstevel@tonic-gate 		ASSERT(found == ndata);
7180Sstevel@tonic-gate 		if (above) {
7190Sstevel@tonic-gate 			found->address = end;
7200Sstevel@tonic-gate 			found->size = above;
7210Sstevel@tonic-gate 		} else if (found->next != NULL) {
7220Sstevel@tonic-gate 			found->address = found->next->address;
7230Sstevel@tonic-gate 			found->size = found->next->size;
7240Sstevel@tonic-gate 			if ((found->next = found->next->next) != NULL)
7250Sstevel@tonic-gate 				found->next->prev = found;
7260Sstevel@tonic-gate 
7270Sstevel@tonic-gate 			bzero((void *)found->address, sizeof (struct memlist));
7280Sstevel@tonic-gate 		} else {
7290Sstevel@tonic-gate 			found->address = end;
7300Sstevel@tonic-gate 			found->size = 0;
7310Sstevel@tonic-gate 		}
7320Sstevel@tonic-gate 
7330Sstevel@tonic-gate 		return ((void *)base);
7340Sstevel@tonic-gate 	}
7350Sstevel@tonic-gate 
7360Sstevel@tonic-gate 	/*
7370Sstevel@tonic-gate 	 * Not the first chunk.
7380Sstevel@tonic-gate 	 */
7390Sstevel@tonic-gate 	if (above) {
7400Sstevel@tonic-gate 		fnd_above = (struct memlist *)end;
7410Sstevel@tonic-gate 		fnd_above->address = end;
7420Sstevel@tonic-gate 		fnd_above->size = above;
7430Sstevel@tonic-gate 
7440Sstevel@tonic-gate 		if ((fnd_above->next = found->next) != NULL)
7450Sstevel@tonic-gate 			fnd_above->next->prev = fnd_above;
7460Sstevel@tonic-gate 		fnd_above->prev = found->prev;
7470Sstevel@tonic-gate 		found->prev->next = fnd_above;
7480Sstevel@tonic-gate 
7490Sstevel@tonic-gate 	} else {
7500Sstevel@tonic-gate 		if ((found->prev->next = found->next) != NULL)
7510Sstevel@tonic-gate 			found->next->prev = found->prev;
7520Sstevel@tonic-gate 	}
7530Sstevel@tonic-gate 
7540Sstevel@tonic-gate 	bzero((void *)found->address, sizeof (struct memlist));
7550Sstevel@tonic-gate 
7560Sstevel@tonic-gate 	return ((void *)base);
7570Sstevel@tonic-gate }
7580Sstevel@tonic-gate 
7590Sstevel@tonic-gate /*
7600Sstevel@tonic-gate  * Size the kernel TSBs based upon the amount of physical
7610Sstevel@tonic-gate  * memory in the system.
7620Sstevel@tonic-gate  */
7630Sstevel@tonic-gate static void
7640Sstevel@tonic-gate calc_tsb_sizes(pgcnt_t npages)
7650Sstevel@tonic-gate {
7660Sstevel@tonic-gate 	PRM_DEBUG(npages);
7670Sstevel@tonic-gate 
7680Sstevel@tonic-gate 	if (npages <= TSB_FREEMEM_MIN) {
7690Sstevel@tonic-gate 		ktsb_szcode = TSB_128K_SZCODE;
7700Sstevel@tonic-gate 		enable_bigktsb = 0;
7710Sstevel@tonic-gate 	} else if (npages <= TSB_FREEMEM_LARGE / 2) {
7720Sstevel@tonic-gate 		ktsb_szcode = TSB_256K_SZCODE;
7730Sstevel@tonic-gate 		enable_bigktsb = 0;
7740Sstevel@tonic-gate 	} else if (npages <= TSB_FREEMEM_LARGE) {
7750Sstevel@tonic-gate 		ktsb_szcode = TSB_512K_SZCODE;
7760Sstevel@tonic-gate 		enable_bigktsb = 0;
7770Sstevel@tonic-gate 	} else if (npages <= TSB_FREEMEM_LARGE * 2 ||
7780Sstevel@tonic-gate 	    enable_bigktsb == 0) {
7790Sstevel@tonic-gate 		ktsb_szcode = TSB_1M_SZCODE;
7800Sstevel@tonic-gate 		enable_bigktsb = 0;
7810Sstevel@tonic-gate 	} else {
7820Sstevel@tonic-gate 		ktsb_szcode = highbit(npages - 1);
7830Sstevel@tonic-gate 		ktsb_szcode -= TSB_START_SIZE;
7840Sstevel@tonic-gate 		ktsb_szcode = MAX(ktsb_szcode, MIN_BIGKTSB_SZCODE);
7850Sstevel@tonic-gate 		ktsb_szcode = MIN(ktsb_szcode, MAX_BIGKTSB_SZCODE);
7860Sstevel@tonic-gate 	}
7870Sstevel@tonic-gate 
7880Sstevel@tonic-gate 	/*
7890Sstevel@tonic-gate 	 * We choose the TSB to hold kernel 4M mappings to have twice
7900Sstevel@tonic-gate 	 * the reach as the primary kernel TSB since this TSB will
7910Sstevel@tonic-gate 	 * potentially (currently) be shared by both mappings to all of
792*3764Sdp78419 	 * physical memory plus user TSBs. If this TSB has to be in nucleus
793*3764Sdp78419 	 * (only for Spitfire and Cheetah) limit its size to 64K.
7940Sstevel@tonic-gate 	 */
795*3764Sdp78419 	ktsb4m_szcode = highbit((2 * npages) / TTEPAGES(TTE4M) - 1);
796*3764Sdp78419 	ktsb4m_szcode -= TSB_START_SIZE;
797*3764Sdp78419 	ktsb4m_szcode = MAX(ktsb4m_szcode, TSB_MIN_SZCODE);
798*3764Sdp78419 	ktsb4m_szcode = MIN(ktsb4m_szcode, TSB_SOFTSZ_MASK);
799*3764Sdp78419 	if ((enable_bigktsb == 0 || ktsb_phys == 0) && ktsb4m_szcode >
800*3764Sdp78419 	    TSB_64K_SZCODE) {
801*3764Sdp78419 		ktsb4m_szcode = TSB_64K_SZCODE;
802*3764Sdp78419 		max_bootlp_tteszc = TTE8K;
803*3764Sdp78419 	}
8040Sstevel@tonic-gate 
8050Sstevel@tonic-gate 	ktsb_sz = TSB_BYTES(ktsb_szcode);	/* kernel 8K tsb size */
8060Sstevel@tonic-gate 	ktsb4m_sz = TSB_BYTES(ktsb4m_szcode);	/* kernel 4M tsb size */
8070Sstevel@tonic-gate }
8080Sstevel@tonic-gate 
8090Sstevel@tonic-gate /*
8100Sstevel@tonic-gate  * Allocate kernel TSBs from nucleus data memory.
8110Sstevel@tonic-gate  * The function return 0 on success and -1 on failure.
8120Sstevel@tonic-gate  */
8130Sstevel@tonic-gate int
8140Sstevel@tonic-gate ndata_alloc_tsbs(struct memlist *ndata, pgcnt_t npages)
8150Sstevel@tonic-gate {
8160Sstevel@tonic-gate 	/*
817*3764Sdp78419 	 * Set ktsb_phys to 1 if the processor supports ASI_QUAD_LDD_PHYS.
818*3764Sdp78419 	 */
819*3764Sdp78419 	sfmmu_setup_4lp();
820*3764Sdp78419 
821*3764Sdp78419 	/*
8220Sstevel@tonic-gate 	 * Size the kernel TSBs based upon the amount of physical
8230Sstevel@tonic-gate 	 * memory in the system.
8240Sstevel@tonic-gate 	 */
8250Sstevel@tonic-gate 	calc_tsb_sizes(npages);
8260Sstevel@tonic-gate 
8270Sstevel@tonic-gate 	/*
8280Sstevel@tonic-gate 	 * Allocate the 8K kernel TSB if it belongs inside the nucleus.
8290Sstevel@tonic-gate 	 */
8300Sstevel@tonic-gate 	if (enable_bigktsb == 0) {
8310Sstevel@tonic-gate 		if ((ktsb_base = ndata_alloc(ndata, ktsb_sz, ktsb_sz)) == NULL)
8320Sstevel@tonic-gate 			return (-1);
8330Sstevel@tonic-gate 		ASSERT(!((uintptr_t)ktsb_base & (ktsb_sz - 1)));
8340Sstevel@tonic-gate 
8350Sstevel@tonic-gate 		PRM_DEBUG(ktsb_base);
8360Sstevel@tonic-gate 		PRM_DEBUG(ktsb_sz);
8370Sstevel@tonic-gate 		PRM_DEBUG(ktsb_szcode);
8380Sstevel@tonic-gate 	}
8390Sstevel@tonic-gate 
8400Sstevel@tonic-gate 	/*
8410Sstevel@tonic-gate 	 * Next, allocate 4M kernel TSB from the nucleus since it's small.
8420Sstevel@tonic-gate 	 */
843*3764Sdp78419 	if (ktsb4m_szcode <= TSB_64K_SZCODE) {
8440Sstevel@tonic-gate 
845*3764Sdp78419 		ktsb4m_base = ndata_alloc(ndata, ktsb4m_sz, ktsb4m_sz);
846*3764Sdp78419 		if (ktsb4m_base == NULL)
847*3764Sdp78419 			return (-1);
848*3764Sdp78419 		ASSERT(!((uintptr_t)ktsb4m_base & (ktsb4m_sz - 1)));
849*3764Sdp78419 
850*3764Sdp78419 		PRM_DEBUG(ktsb4m_base);
851*3764Sdp78419 		PRM_DEBUG(ktsb4m_sz);
852*3764Sdp78419 		PRM_DEBUG(ktsb4m_szcode);
853*3764Sdp78419 	}
8540Sstevel@tonic-gate 
8550Sstevel@tonic-gate 	return (0);
8560Sstevel@tonic-gate }
8570Sstevel@tonic-gate 
8580Sstevel@tonic-gate /*
8590Sstevel@tonic-gate  * Allocate hat structs from the nucleus data memory.
8600Sstevel@tonic-gate  */
8610Sstevel@tonic-gate int
8620Sstevel@tonic-gate ndata_alloc_hat(struct memlist *ndata, pgcnt_t npages, pgcnt_t kpm_npages)
8630Sstevel@tonic-gate {
8640Sstevel@tonic-gate 	size_t	mml_alloc_sz;
8650Sstevel@tonic-gate 	size_t	cb_alloc_sz;
8660Sstevel@tonic-gate 	int	max_nucuhme_buckets = MAX_NUCUHME_BUCKETS;
8670Sstevel@tonic-gate 	int	max_nuckhme_buckets = MAX_NUCKHME_BUCKETS;
8680Sstevel@tonic-gate 	ulong_t hme_buckets;
8690Sstevel@tonic-gate 
8700Sstevel@tonic-gate 	if (enable_bigktsb) {
8710Sstevel@tonic-gate 		ASSERT((max_nucuhme_buckets + max_nuckhme_buckets) *
8720Sstevel@tonic-gate 		    sizeof (struct hmehash_bucket) <=
8730Sstevel@tonic-gate 			TSB_BYTES(TSB_1M_SZCODE));
8740Sstevel@tonic-gate 
8750Sstevel@tonic-gate 		max_nucuhme_buckets *= 2;
8760Sstevel@tonic-gate 		max_nuckhme_buckets *= 2;
8770Sstevel@tonic-gate 	}
8780Sstevel@tonic-gate 
8790Sstevel@tonic-gate 	/*
8800Sstevel@tonic-gate 	 * The number of buckets in the hme hash tables
8810Sstevel@tonic-gate 	 * is a power of 2 such that the average hash chain length is
8820Sstevel@tonic-gate 	 * HMENT_HASHAVELEN.  The number of buckets for the user hash is
8830Sstevel@tonic-gate 	 * a function of physical memory and a predefined overmapping factor.
8840Sstevel@tonic-gate 	 * The number of buckets for the kernel hash is a function of
8850Sstevel@tonic-gate 	 * physical memory only.
8860Sstevel@tonic-gate 	 */
8870Sstevel@tonic-gate 	hme_buckets = (npages * HMEHASH_FACTOR) /
8880Sstevel@tonic-gate 		(HMENT_HASHAVELEN * (HMEBLK_SPAN(TTE8K) >> MMU_PAGESHIFT));
8890Sstevel@tonic-gate 
8900Sstevel@tonic-gate 	uhmehash_num = (int)MIN(hme_buckets, MAX_UHME_BUCKETS);
8910Sstevel@tonic-gate 
8920Sstevel@tonic-gate 	if (uhmehash_num > USER_BUCKETS_THRESHOLD) {
8930Sstevel@tonic-gate 		/*
8940Sstevel@tonic-gate 		 * if uhmehash_num is not power of 2 round it down to the
8950Sstevel@tonic-gate 		 *  next power of 2.
8960Sstevel@tonic-gate 		 */
8970Sstevel@tonic-gate 		uint_t align = 1 << (highbit(uhmehash_num - 1) - 1);
8980Sstevel@tonic-gate 		uhmehash_num = P2ALIGN(uhmehash_num, align);
8990Sstevel@tonic-gate 	} else
9000Sstevel@tonic-gate 		uhmehash_num = 1 << highbit(uhmehash_num - 1);
9010Sstevel@tonic-gate 
9020Sstevel@tonic-gate 	hme_buckets = npages / (HMEBLK_SPAN(TTE8K) >> MMU_PAGESHIFT);
9030Sstevel@tonic-gate 	khmehash_num = (int)MIN(hme_buckets, MAX_KHME_BUCKETS);
9040Sstevel@tonic-gate 	khmehash_num = 1 << highbit(khmehash_num - 1);
9050Sstevel@tonic-gate 	khmehash_num = MAX(khmehash_num, MIN_KHME_BUCKETS);
9060Sstevel@tonic-gate 
9070Sstevel@tonic-gate 	if ((khmehash_num > max_nuckhme_buckets) ||
9080Sstevel@tonic-gate 		(uhmehash_num > max_nucuhme_buckets)) {
9090Sstevel@tonic-gate 		khme_hash = NULL;
9100Sstevel@tonic-gate 		uhme_hash = NULL;
9110Sstevel@tonic-gate 	} else {
9120Sstevel@tonic-gate 		size_t hmehash_sz = (uhmehash_num + khmehash_num) *
9130Sstevel@tonic-gate 		    sizeof (struct hmehash_bucket);
9140Sstevel@tonic-gate 
9150Sstevel@tonic-gate 		if ((khme_hash = ndata_alloc(ndata, hmehash_sz,
9160Sstevel@tonic-gate 		    ecache_alignsize)) != NULL)
9170Sstevel@tonic-gate 			uhme_hash = &khme_hash[khmehash_num];
9180Sstevel@tonic-gate 		else
9190Sstevel@tonic-gate 			uhme_hash = NULL;
9200Sstevel@tonic-gate 
9210Sstevel@tonic-gate 		PRM_DEBUG(hmehash_sz);
9220Sstevel@tonic-gate 	}
9230Sstevel@tonic-gate 
9240Sstevel@tonic-gate 	PRM_DEBUG(khme_hash);
9250Sstevel@tonic-gate 	PRM_DEBUG(khmehash_num);
9260Sstevel@tonic-gate 	PRM_DEBUG(uhme_hash);
9270Sstevel@tonic-gate 	PRM_DEBUG(uhmehash_num);
9280Sstevel@tonic-gate 
9290Sstevel@tonic-gate 	/*
9300Sstevel@tonic-gate 	 * For the page mapping list mutex array we allocate one mutex
9310Sstevel@tonic-gate 	 * for every 128 pages (1 MB) with a minimum of 64 entries and
9320Sstevel@tonic-gate 	 * a maximum of 8K entries. For the initial computation npages
9330Sstevel@tonic-gate 	 * is rounded up (ie. 1 << highbit(npages * 1.5 / 128))
9340Sstevel@tonic-gate 	 *
9350Sstevel@tonic-gate 	 * mml_shift is roughly log2(mml_table_sz) + 3 for MLIST_HASH
9360Sstevel@tonic-gate 	 *
9370Sstevel@tonic-gate 	 * It is not required that this be allocated from the nucleus,
9380Sstevel@tonic-gate 	 * but it is desirable.  So we first allocate from the nucleus
9390Sstevel@tonic-gate 	 * everything that must be there.  Having done so, if mml_table
9400Sstevel@tonic-gate 	 * will fit within what remains of the nucleus then it will be
9410Sstevel@tonic-gate 	 * allocated here.  If not, set mml_table to NULL, which will cause
9420Sstevel@tonic-gate 	 * startup_memlist() to BOP_ALLOC() space for it after our return...
9430Sstevel@tonic-gate 	 */
9440Sstevel@tonic-gate 	mml_table_sz = 1 << highbit((npages * 3) / 256);
9450Sstevel@tonic-gate 	if (mml_table_sz < 64)
9460Sstevel@tonic-gate 		mml_table_sz = 64;
9470Sstevel@tonic-gate 	else if (mml_table_sz > 8192)
9480Sstevel@tonic-gate 		mml_table_sz = 8192;
9490Sstevel@tonic-gate 	mml_shift = highbit(mml_table_sz) + 3;
9500Sstevel@tonic-gate 
9510Sstevel@tonic-gate 	PRM_DEBUG(mml_table_sz);
9520Sstevel@tonic-gate 	PRM_DEBUG(mml_shift);
9530Sstevel@tonic-gate 
9540Sstevel@tonic-gate 	mml_alloc_sz = mml_table_sz * sizeof (kmutex_t);
9550Sstevel@tonic-gate 
9560Sstevel@tonic-gate 	mml_table = ndata_alloc(ndata, mml_alloc_sz, ecache_alignsize);
9570Sstevel@tonic-gate 
9580Sstevel@tonic-gate 	PRM_DEBUG(mml_table);
9590Sstevel@tonic-gate 
9600Sstevel@tonic-gate 	cb_alloc_sz = sfmmu_max_cb_id * sizeof (struct sfmmu_callback);
9610Sstevel@tonic-gate 	PRM_DEBUG(cb_alloc_sz);
9620Sstevel@tonic-gate 	sfmmu_cb_table = ndata_alloc(ndata, cb_alloc_sz, ecache_alignsize);
9630Sstevel@tonic-gate 	PRM_DEBUG(sfmmu_cb_table);
9640Sstevel@tonic-gate 
9650Sstevel@tonic-gate 	/*
9660Sstevel@tonic-gate 	 * For the kpm_page mutex array we allocate one mutex every 16
9670Sstevel@tonic-gate 	 * kpm pages (64MB). In smallpage mode we allocate one mutex
9680Sstevel@tonic-gate 	 * every 8K pages. The minimum is set to 64 entries and the
9690Sstevel@tonic-gate 	 * maximum to 8K entries.
9700Sstevel@tonic-gate 	 *
9710Sstevel@tonic-gate 	 * It is not required that this be allocated from the nucleus,
9720Sstevel@tonic-gate 	 * but it is desirable.  So we first allocate from the nucleus
9730Sstevel@tonic-gate 	 * everything that must be there.  Having done so, if kpmp_table
9740Sstevel@tonic-gate 	 * or kpmp_stable will fit within what remains of the nucleus
9750Sstevel@tonic-gate 	 * then it will be allocated here.  If not, startup_memlist()
9760Sstevel@tonic-gate 	 * will use BOP_ALLOC() space for it after our return...
9770Sstevel@tonic-gate 	 */
9780Sstevel@tonic-gate 	if (kpm_enable) {
9790Sstevel@tonic-gate 		size_t	kpmp_alloc_sz;
9800Sstevel@tonic-gate 
9810Sstevel@tonic-gate 		if (kpm_smallpages == 0) {
9820Sstevel@tonic-gate 			kpmp_shift = highbit(sizeof (kpm_page_t)) - 1;
9830Sstevel@tonic-gate 			kpmp_table_sz = 1 << highbit(kpm_npages / 16);
9840Sstevel@tonic-gate 			kpmp_table_sz = (kpmp_table_sz < 64) ? 64 :
9850Sstevel@tonic-gate 			    ((kpmp_table_sz > 8192) ? 8192 : kpmp_table_sz);
9860Sstevel@tonic-gate 			kpmp_alloc_sz = kpmp_table_sz * sizeof (kpm_hlk_t);
9870Sstevel@tonic-gate 
9880Sstevel@tonic-gate 			kpmp_table = ndata_alloc(ndata, kpmp_alloc_sz,
9890Sstevel@tonic-gate 			    ecache_alignsize);
9900Sstevel@tonic-gate 
9910Sstevel@tonic-gate 			PRM_DEBUG(kpmp_table);
9920Sstevel@tonic-gate 			PRM_DEBUG(kpmp_table_sz);
9930Sstevel@tonic-gate 
9940Sstevel@tonic-gate 			kpmp_stable_sz = 0;
9950Sstevel@tonic-gate 			kpmp_stable = NULL;
9960Sstevel@tonic-gate 		} else {
9970Sstevel@tonic-gate 			ASSERT(kpm_pgsz == PAGESIZE);
9980Sstevel@tonic-gate 			kpmp_shift = highbit(sizeof (kpm_shlk_t)) + 1;
9990Sstevel@tonic-gate 			kpmp_stable_sz = 1 << highbit(kpm_npages / 8192);
10000Sstevel@tonic-gate 			kpmp_stable_sz = (kpmp_stable_sz < 64) ? 64 :
10010Sstevel@tonic-gate 			    ((kpmp_stable_sz > 8192) ? 8192 : kpmp_stable_sz);
10020Sstevel@tonic-gate 			kpmp_alloc_sz = kpmp_stable_sz * sizeof (kpm_shlk_t);
10030Sstevel@tonic-gate 
10040Sstevel@tonic-gate 			kpmp_stable = ndata_alloc(ndata, kpmp_alloc_sz,
10050Sstevel@tonic-gate 			    ecache_alignsize);
10060Sstevel@tonic-gate 
10070Sstevel@tonic-gate 			PRM_DEBUG(kpmp_stable);
10080Sstevel@tonic-gate 			PRM_DEBUG(kpmp_stable_sz);
10090Sstevel@tonic-gate 
10100Sstevel@tonic-gate 			kpmp_table_sz = 0;
10110Sstevel@tonic-gate 			kpmp_table = NULL;
10120Sstevel@tonic-gate 		}
10130Sstevel@tonic-gate 		PRM_DEBUG(kpmp_shift);
10140Sstevel@tonic-gate 	}
10150Sstevel@tonic-gate 
10160Sstevel@tonic-gate 	return (0);
10170Sstevel@tonic-gate }
10180Sstevel@tonic-gate 
1019*3764Sdp78419 /*
1020*3764Sdp78419  * Allocate virtual addresses at base with given alignment.
1021*3764Sdp78419  * Note that there is no physical memory behind the address yet.
1022*3764Sdp78419  */
10230Sstevel@tonic-gate caddr_t
1024*3764Sdp78419 alloc_hme_buckets(caddr_t base, int alignsize)
10250Sstevel@tonic-gate {
10260Sstevel@tonic-gate 	size_t hmehash_sz = (uhmehash_num + khmehash_num) *
1027*3764Sdp78419 	    sizeof (struct hmehash_bucket);
10280Sstevel@tonic-gate 
10290Sstevel@tonic-gate 	ASSERT(khme_hash == NULL);
10300Sstevel@tonic-gate 	ASSERT(uhme_hash == NULL);
10310Sstevel@tonic-gate 
1032*3764Sdp78419 	base = (caddr_t)roundup((uintptr_t)base, alignsize);
1033*3764Sdp78419 	hmehash_sz = roundup(hmehash_sz, alignsize);
10340Sstevel@tonic-gate 
1035*3764Sdp78419 	khme_hash = (struct hmehash_bucket *)base;
10360Sstevel@tonic-gate 	uhme_hash = (struct hmehash_bucket *)((caddr_t)khme_hash +
1037*3764Sdp78419 	    khmehash_num * sizeof (struct hmehash_bucket));
10380Sstevel@tonic-gate 	base += hmehash_sz;
10390Sstevel@tonic-gate 	return (base);
10400Sstevel@tonic-gate }
10410Sstevel@tonic-gate 
10420Sstevel@tonic-gate /*
1043*3764Sdp78419  * This function bop allocs kernel TSBs.
10440Sstevel@tonic-gate  */
10450Sstevel@tonic-gate caddr_t
10460Sstevel@tonic-gate sfmmu_ktsb_alloc(caddr_t tsbbase)
10470Sstevel@tonic-gate {
10480Sstevel@tonic-gate 	caddr_t vaddr;
10490Sstevel@tonic-gate 
10500Sstevel@tonic-gate 	if (enable_bigktsb) {
10510Sstevel@tonic-gate 		ktsb_base = (caddr_t)roundup((uintptr_t)tsbbase, ktsb_sz);
10520Sstevel@tonic-gate 		vaddr = (caddr_t)BOP_ALLOC(bootops, ktsb_base, ktsb_sz,
10530Sstevel@tonic-gate 		    ktsb_sz);
10540Sstevel@tonic-gate 		if (vaddr != ktsb_base)
10550Sstevel@tonic-gate 			cmn_err(CE_PANIC, "sfmmu_ktsb_alloc: can't alloc"
1056*3764Sdp78419 			    " 8K bigktsb");
10570Sstevel@tonic-gate 		ktsb_base = vaddr;
10580Sstevel@tonic-gate 		tsbbase = ktsb_base + ktsb_sz;
10590Sstevel@tonic-gate 		PRM_DEBUG(ktsb_base);
10600Sstevel@tonic-gate 		PRM_DEBUG(tsbbase);
10610Sstevel@tonic-gate 	}
1062*3764Sdp78419 
1063*3764Sdp78419 	if (ktsb4m_szcode > TSB_64K_SZCODE) {
1064*3764Sdp78419 		ASSERT(ktsb_phys && enable_bigktsb);
1065*3764Sdp78419 		ktsb4m_base = (caddr_t)roundup((uintptr_t)tsbbase, ktsb4m_sz);
1066*3764Sdp78419 		vaddr = (caddr_t)BOP_ALLOC(bootops, ktsb4m_base, ktsb4m_sz,
1067*3764Sdp78419 		    ktsb4m_sz);
1068*3764Sdp78419 		if (vaddr != ktsb4m_base)
1069*3764Sdp78419 			cmn_err(CE_PANIC, "sfmmu_ktsb_alloc: can't alloc"
1070*3764Sdp78419 			    " 4M bigktsb");
1071*3764Sdp78419 		ktsb4m_base = vaddr;
1072*3764Sdp78419 		tsbbase = ktsb4m_base + ktsb4m_sz;
1073*3764Sdp78419 		PRM_DEBUG(ktsb4m_base);
1074*3764Sdp78419 		PRM_DEBUG(tsbbase);
1075*3764Sdp78419 	}
10760Sstevel@tonic-gate 	return (tsbbase);
10770Sstevel@tonic-gate }
10780Sstevel@tonic-gate 
10790Sstevel@tonic-gate /*
10800Sstevel@tonic-gate  * Moves code assembled outside of the trap table into the trap
10810Sstevel@tonic-gate  * table taking care to relocate relative branches to code outside
10820Sstevel@tonic-gate  * of the trap handler.
10830Sstevel@tonic-gate  */
10840Sstevel@tonic-gate static void
10850Sstevel@tonic-gate sfmmu_reloc_trap_handler(void *tablep, void *start, size_t count)
10860Sstevel@tonic-gate {
10870Sstevel@tonic-gate 	size_t i;
10880Sstevel@tonic-gate 	uint32_t *src;
10890Sstevel@tonic-gate 	uint32_t *dst;
10900Sstevel@tonic-gate 	uint32_t inst;
10910Sstevel@tonic-gate 	int op, op2;
10920Sstevel@tonic-gate 	int32_t offset;
10930Sstevel@tonic-gate 	int disp;
10940Sstevel@tonic-gate 
10950Sstevel@tonic-gate 	src = start;
10960Sstevel@tonic-gate 	dst = tablep;
10970Sstevel@tonic-gate 	offset = src - dst;
10980Sstevel@tonic-gate 	for (src = start, i = 0; i < count; i++, src++, dst++) {
10990Sstevel@tonic-gate 		inst = *dst = *src;
11000Sstevel@tonic-gate 		op = (inst >> 30) & 0x2;
11010Sstevel@tonic-gate 		if (op == 1) {
11020Sstevel@tonic-gate 			/* call */
11030Sstevel@tonic-gate 			disp = ((int32_t)inst << 2) >> 2; /* sign-extend */
11040Sstevel@tonic-gate 			if (disp + i >= 0 && disp + i < count)
11050Sstevel@tonic-gate 				continue;
11060Sstevel@tonic-gate 			disp += offset;
11070Sstevel@tonic-gate 			inst = 0x40000000u | (disp & 0x3fffffffu);
11080Sstevel@tonic-gate 			*dst = inst;
11090Sstevel@tonic-gate 		} else if (op == 0) {
11100Sstevel@tonic-gate 			/* branch or sethi */
11110Sstevel@tonic-gate 			op2 = (inst >> 22) & 0x7;
11120Sstevel@tonic-gate 
11130Sstevel@tonic-gate 			switch (op2) {
11140Sstevel@tonic-gate 			case 0x3: /* BPr */
11150Sstevel@tonic-gate 				disp = (((inst >> 20) & 0x3) << 14) |
11160Sstevel@tonic-gate 				    (inst & 0x3fff);
11170Sstevel@tonic-gate 				disp = (disp << 16) >> 16; /* sign-extend */
11180Sstevel@tonic-gate 				if (disp + i >= 0 && disp + i < count)
11190Sstevel@tonic-gate 					continue;
11200Sstevel@tonic-gate 				disp += offset;
11210Sstevel@tonic-gate 				if (((disp << 16) >> 16) != disp)
11220Sstevel@tonic-gate 					cmn_err(CE_PANIC, "bad reloc");
11230Sstevel@tonic-gate 				inst &= ~0x303fff;
11240Sstevel@tonic-gate 				inst |= (disp & 0x3fff);
11250Sstevel@tonic-gate 				inst |= (disp & 0xc000) << 6;
11260Sstevel@tonic-gate 				break;
11270Sstevel@tonic-gate 
11280Sstevel@tonic-gate 			case 0x2: /* Bicc */
11290Sstevel@tonic-gate 				disp = ((int32_t)inst << 10) >> 10;
11300Sstevel@tonic-gate 				if (disp + i >= 0 && disp + i < count)
11310Sstevel@tonic-gate 					continue;
11320Sstevel@tonic-gate 				disp += offset;
11330Sstevel@tonic-gate 				if (((disp << 10) >> 10) != disp)
11340Sstevel@tonic-gate 					cmn_err(CE_PANIC, "bad reloc");
11350Sstevel@tonic-gate 				inst &= ~0x3fffff;
11360Sstevel@tonic-gate 				inst |= (disp & 0x3fffff);
11370Sstevel@tonic-gate 				break;
11380Sstevel@tonic-gate 
11390Sstevel@tonic-gate 			case 0x1: /* Bpcc */
11400Sstevel@tonic-gate 				disp = ((int32_t)inst << 13) >> 13;
11410Sstevel@tonic-gate 				if (disp + i >= 0 && disp + i < count)
11420Sstevel@tonic-gate 					continue;
11430Sstevel@tonic-gate 				disp += offset;
11440Sstevel@tonic-gate 				if (((disp << 13) >> 13) != disp)
11450Sstevel@tonic-gate 					cmn_err(CE_PANIC, "bad reloc");
11460Sstevel@tonic-gate 				inst &= ~0x7ffff;
11470Sstevel@tonic-gate 				inst |= (disp & 0x7ffffu);
11480Sstevel@tonic-gate 				break;
11490Sstevel@tonic-gate 			}
11500Sstevel@tonic-gate 			*dst = inst;
11510Sstevel@tonic-gate 		}
11520Sstevel@tonic-gate 	}
11530Sstevel@tonic-gate 	flush_instr_mem(tablep, count * sizeof (uint32_t));
11540Sstevel@tonic-gate }
11550Sstevel@tonic-gate 
11560Sstevel@tonic-gate /*
11570Sstevel@tonic-gate  * Routine to allocate a large page to use in the TSB caches.
11580Sstevel@tonic-gate  */
11590Sstevel@tonic-gate /*ARGSUSED*/
11600Sstevel@tonic-gate static page_t *
11610Sstevel@tonic-gate sfmmu_tsb_page_create(void *addr, size_t size, int vmflag, void *arg)
11620Sstevel@tonic-gate {
11630Sstevel@tonic-gate 	int pgflags;
11640Sstevel@tonic-gate 
11650Sstevel@tonic-gate 	pgflags = PG_EXCL;
11660Sstevel@tonic-gate 	if ((vmflag & VM_NOSLEEP) == 0)
11670Sstevel@tonic-gate 		pgflags |= PG_WAIT;
11680Sstevel@tonic-gate 	if (vmflag & VM_PANIC)
11690Sstevel@tonic-gate 		pgflags |= PG_PANIC;
11700Sstevel@tonic-gate 	if (vmflag & VM_PUSHPAGE)
11710Sstevel@tonic-gate 		pgflags |= PG_PUSHPAGE;
11720Sstevel@tonic-gate 
11730Sstevel@tonic-gate 	return (page_create_va_large(&kvp, (u_offset_t)(uintptr_t)addr, size,
11740Sstevel@tonic-gate 	    pgflags, &kvseg, addr, arg));
11750Sstevel@tonic-gate }
11760Sstevel@tonic-gate 
11770Sstevel@tonic-gate /*
11780Sstevel@tonic-gate  * Allocate a large page to back the virtual address range
11790Sstevel@tonic-gate  * [addr, addr + size).  If addr is NULL, allocate the virtual address
11800Sstevel@tonic-gate  * space as well.
11810Sstevel@tonic-gate  */
11820Sstevel@tonic-gate static void *
11830Sstevel@tonic-gate sfmmu_tsb_xalloc(vmem_t *vmp, void *inaddr, size_t size, int vmflag,
11840Sstevel@tonic-gate     uint_t attr, page_t *(*page_create_func)(void *, size_t, int, void *),
11850Sstevel@tonic-gate     void *pcarg)
11860Sstevel@tonic-gate {
11870Sstevel@tonic-gate 	page_t *ppl;
11880Sstevel@tonic-gate 	page_t *rootpp;
11890Sstevel@tonic-gate 	caddr_t addr = inaddr;
11900Sstevel@tonic-gate 	pgcnt_t npages = btopr(size);
11910Sstevel@tonic-gate 	page_t **ppa;
11920Sstevel@tonic-gate 	int i = 0;
11930Sstevel@tonic-gate 
11940Sstevel@tonic-gate 	/*
11950Sstevel@tonic-gate 	 * Assuming that only TSBs will call this with size > PAGESIZE
11960Sstevel@tonic-gate 	 * There is no reason why this couldn't be expanded to 8k pages as
11970Sstevel@tonic-gate 	 * well, or other page sizes in the future .... but for now, we
11980Sstevel@tonic-gate 	 * only support fixed sized page requests.
11990Sstevel@tonic-gate 	 */
12000Sstevel@tonic-gate 	if ((inaddr == NULL) && ((addr = vmem_xalloc(vmp, size, size, 0, 0,
12010Sstevel@tonic-gate 	    NULL, NULL, vmflag)) == NULL))
12020Sstevel@tonic-gate 		return (NULL);
12030Sstevel@tonic-gate 
12040Sstevel@tonic-gate 	/* If we ever don't want TSB slab-sized pages, this will panic */
12050Sstevel@tonic-gate 	ASSERT(((uintptr_t)addr & (tsb_slab_size - 1)) == 0);
12060Sstevel@tonic-gate 
12070Sstevel@tonic-gate 	if (page_resv(npages, vmflag & VM_KMFLAGS) == 0) {
12080Sstevel@tonic-gate 		if (inaddr == NULL)
12090Sstevel@tonic-gate 			vmem_xfree(vmp, addr, size);
12100Sstevel@tonic-gate 		return (NULL);
12110Sstevel@tonic-gate 	}
12120Sstevel@tonic-gate 
12130Sstevel@tonic-gate 	ppl = page_create_func(addr, size, vmflag, pcarg);
12140Sstevel@tonic-gate 	if (ppl == NULL) {
12150Sstevel@tonic-gate 		if (inaddr == NULL)
12160Sstevel@tonic-gate 			vmem_xfree(vmp, addr, size);
12170Sstevel@tonic-gate 		page_unresv(npages);
12180Sstevel@tonic-gate 		return (NULL);
12190Sstevel@tonic-gate 	}
12200Sstevel@tonic-gate 
12210Sstevel@tonic-gate 	rootpp = ppl;
12220Sstevel@tonic-gate 	ppa = kmem_zalloc(npages * sizeof (page_t *), KM_SLEEP);
12230Sstevel@tonic-gate 	while (ppl != NULL) {
12240Sstevel@tonic-gate 		page_t *pp = ppl;
12250Sstevel@tonic-gate 		ppa[i++] = pp;
12260Sstevel@tonic-gate 		page_sub(&ppl, pp);
12270Sstevel@tonic-gate 		ASSERT(page_iolock_assert(pp));
12280Sstevel@tonic-gate 		page_io_unlock(pp);
12290Sstevel@tonic-gate 	}
12300Sstevel@tonic-gate 
12310Sstevel@tonic-gate 	/*
12320Sstevel@tonic-gate 	 * Load the locked entry.  It's OK to preload the entry into
12330Sstevel@tonic-gate 	 * the TSB since we now support large mappings in the kernel TSB.
12340Sstevel@tonic-gate 	 */
12350Sstevel@tonic-gate 	hat_memload_array(kas.a_hat, (caddr_t)rootpp->p_offset, size,
12360Sstevel@tonic-gate 	    ppa, (PROT_ALL & ~PROT_USER) | HAT_NOSYNC | attr, HAT_LOAD_LOCK);
12370Sstevel@tonic-gate 
12380Sstevel@tonic-gate 	for (--i; i >= 0; --i) {
12390Sstevel@tonic-gate 		(void) page_pp_lock(ppa[i], 0, 1);
12400Sstevel@tonic-gate 		page_unlock(ppa[i]);
12410Sstevel@tonic-gate 	}
12420Sstevel@tonic-gate 
12430Sstevel@tonic-gate 	kmem_free(ppa, npages * sizeof (page_t *));
12440Sstevel@tonic-gate 	return (addr);
12450Sstevel@tonic-gate }
12460Sstevel@tonic-gate 
12470Sstevel@tonic-gate /* Called to import new spans into the TSB vmem arenas */
12480Sstevel@tonic-gate void *
12490Sstevel@tonic-gate sfmmu_tsb_segkmem_alloc(vmem_t *vmp, size_t size, int vmflag)
12500Sstevel@tonic-gate {
12510Sstevel@tonic-gate 	lgrp_id_t lgrpid = LGRP_NONE;
12520Sstevel@tonic-gate 
12530Sstevel@tonic-gate 	if (tsb_lgrp_affinity) {
12540Sstevel@tonic-gate 		/*
12550Sstevel@tonic-gate 		 * Search for the vmp->lgrpid mapping by brute force;
12560Sstevel@tonic-gate 		 * some day vmp will have an lgrp, until then we have
12570Sstevel@tonic-gate 		 * to do this the hard way.
12580Sstevel@tonic-gate 		 */
12590Sstevel@tonic-gate 		for (lgrpid = 0; lgrpid < NLGRPS_MAX &&
12600Sstevel@tonic-gate 		    vmp != kmem_tsb_default_arena[lgrpid]; lgrpid++);
12610Sstevel@tonic-gate 		if (lgrpid == NLGRPS_MAX)
12620Sstevel@tonic-gate 			lgrpid = LGRP_NONE;
12630Sstevel@tonic-gate 	}
12640Sstevel@tonic-gate 
12650Sstevel@tonic-gate 	return (sfmmu_tsb_xalloc(vmp, NULL, size, vmflag, 0,
12660Sstevel@tonic-gate 	    sfmmu_tsb_page_create, lgrpid != LGRP_NONE? &lgrpid : NULL));
12670Sstevel@tonic-gate }
12680Sstevel@tonic-gate 
12690Sstevel@tonic-gate /* Called to free spans from the TSB vmem arenas */
12700Sstevel@tonic-gate void
12710Sstevel@tonic-gate sfmmu_tsb_segkmem_free(vmem_t *vmp, void *inaddr, size_t size)
12720Sstevel@tonic-gate {
12730Sstevel@tonic-gate 	page_t *pp;
12740Sstevel@tonic-gate 	caddr_t addr = inaddr;
12750Sstevel@tonic-gate 	caddr_t eaddr;
12760Sstevel@tonic-gate 	pgcnt_t npages = btopr(size);
12770Sstevel@tonic-gate 	pgcnt_t pgs_left = npages;
12780Sstevel@tonic-gate 	page_t *rootpp = NULL;
12790Sstevel@tonic-gate 
12800Sstevel@tonic-gate 	ASSERT(((uintptr_t)addr & (tsb_slab_size - 1)) == 0);
12810Sstevel@tonic-gate 
12820Sstevel@tonic-gate 	hat_unload(kas.a_hat, addr, size, HAT_UNLOAD_UNLOCK);
12830Sstevel@tonic-gate 
12840Sstevel@tonic-gate 	for (eaddr = addr + size; addr < eaddr; addr += PAGESIZE) {
12850Sstevel@tonic-gate 		pp = page_lookup(&kvp, (u_offset_t)(uintptr_t)addr, SE_EXCL);
12860Sstevel@tonic-gate 		if (pp == NULL)
12870Sstevel@tonic-gate 			panic("sfmmu_tsb_segkmem_free: page not found");
12880Sstevel@tonic-gate 
12890Sstevel@tonic-gate 		ASSERT(PAGE_EXCL(pp));
12900Sstevel@tonic-gate 		page_pp_unlock(pp, 0, 1);
12910Sstevel@tonic-gate 
12920Sstevel@tonic-gate 		if (rootpp == NULL)
12930Sstevel@tonic-gate 			rootpp = pp;
12940Sstevel@tonic-gate 		if (--pgs_left == 0) {
12950Sstevel@tonic-gate 			/*
12960Sstevel@tonic-gate 			 * similar logic to segspt_free_pages, but we know we
12970Sstevel@tonic-gate 			 * have one large page.
12980Sstevel@tonic-gate 			 */
12990Sstevel@tonic-gate 			page_destroy_pages(rootpp);
13000Sstevel@tonic-gate 		}
13010Sstevel@tonic-gate 	}
13020Sstevel@tonic-gate 	page_unresv(npages);
13030Sstevel@tonic-gate 
13040Sstevel@tonic-gate 	if (vmp != NULL)
13050Sstevel@tonic-gate 		vmem_xfree(vmp, inaddr, size);
13060Sstevel@tonic-gate }
1307