10Sstevel@tonic-gate /* 23446Smrj * CDDL HEADER START 33446Smrj * 43446Smrj * The contents of this file are subject to the terms of the 53446Smrj * Common Development and Distribution License (the "License"). 63446Smrj * You may not use this file except in compliance with the License. 73446Smrj * 83446Smrj * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 93446Smrj * or http://www.opensolaris.org/os/licensing. 103446Smrj * See the License for the specific language governing permissions 113446Smrj * and limitations under the License. 123446Smrj * 133446Smrj * When distributing Covered Code, include this CDDL HEADER in each 143446Smrj * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 153446Smrj * If applicable, add the following below this CDDL HEADER, with the 163446Smrj * fields enclosed by brackets "[]" replaced with your own identifying 173446Smrj * information: Portions Copyright [yyyy] [name of copyright owner] 183446Smrj * 193446Smrj * CDDL HEADER END 203446Smrj */ 213446Smrj 223446Smrj /* 23*12613SSurya.Prakki@Sun.COM * Copyright (c) 2004, 2010, Oracle and/or its affiliates. All rights reserved. 240Sstevel@tonic-gate */ 250Sstevel@tonic-gate 260Sstevel@tonic-gate /* 270Sstevel@tonic-gate * Copyright (c) 1992 Terrence R. Lambert. 280Sstevel@tonic-gate * Copyright (c) 1990 The Regents of the University of California. 290Sstevel@tonic-gate * All rights reserved. 300Sstevel@tonic-gate * 310Sstevel@tonic-gate * This code is derived from software contributed to Berkeley by 320Sstevel@tonic-gate * William Jolitz. 330Sstevel@tonic-gate * 340Sstevel@tonic-gate * Redistribution and use in source and binary forms, with or without 350Sstevel@tonic-gate * modification, are permitted provided that the following conditions 360Sstevel@tonic-gate * are met: 370Sstevel@tonic-gate * 1. Redistributions of source code must retain the above copyright 380Sstevel@tonic-gate * notice, this list of conditions and the following disclaimer. 390Sstevel@tonic-gate * 2. Redistributions in binary form must reproduce the above copyright 400Sstevel@tonic-gate * notice, this list of conditions and the following disclaimer in the 410Sstevel@tonic-gate * documentation and/or other materials provided with the distribution. 420Sstevel@tonic-gate * 3. All advertising materials mentioning features or use of this software 430Sstevel@tonic-gate * must display the following acknowledgement: 440Sstevel@tonic-gate * This product includes software developed by the University of 450Sstevel@tonic-gate * California, Berkeley and its contributors. 460Sstevel@tonic-gate * 4. Neither the name of the University nor the names of its contributors 470Sstevel@tonic-gate * may be used to endorse or promote products derived from this software 480Sstevel@tonic-gate * without specific prior written permission. 490Sstevel@tonic-gate * 500Sstevel@tonic-gate * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND 510Sstevel@tonic-gate * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 520Sstevel@tonic-gate * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 530Sstevel@tonic-gate * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE 540Sstevel@tonic-gate * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 550Sstevel@tonic-gate * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 560Sstevel@tonic-gate * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 570Sstevel@tonic-gate * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 580Sstevel@tonic-gate * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 590Sstevel@tonic-gate * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 600Sstevel@tonic-gate * SUCH DAMAGE. 610Sstevel@tonic-gate * 620Sstevel@tonic-gate * from: @(#)machdep.c 7.4 (Berkeley) 6/3/91 630Sstevel@tonic-gate */ 640Sstevel@tonic-gate 650Sstevel@tonic-gate #include <sys/types.h> 663446Smrj #include <sys/sysmacros.h> 670Sstevel@tonic-gate #include <sys/tss.h> 680Sstevel@tonic-gate #include <sys/segments.h> 690Sstevel@tonic-gate #include <sys/trap.h> 700Sstevel@tonic-gate #include <sys/cpuvar.h> 713446Smrj #include <sys/bootconf.h> 720Sstevel@tonic-gate #include <sys/x86_archext.h> 733446Smrj #include <sys/controlregs.h> 740Sstevel@tonic-gate #include <sys/archsystm.h> 750Sstevel@tonic-gate #include <sys/machsystm.h> 760Sstevel@tonic-gate #include <sys/kobj.h> 770Sstevel@tonic-gate #include <sys/cmn_err.h> 780Sstevel@tonic-gate #include <sys/reboot.h> 790Sstevel@tonic-gate #include <sys/kdi.h> 803446Smrj #include <sys/mach_mmu.h> 811217Srab #include <sys/systm.h> 825084Sjohnlev 835084Sjohnlev #ifdef __xpv 845084Sjohnlev #include <sys/hypervisor.h> 855084Sjohnlev #include <vm/as.h> 865084Sjohnlev #endif 875084Sjohnlev 883446Smrj #include <sys/promif.h> 893446Smrj #include <sys/bootinfo.h> 903446Smrj #include <vm/kboot_mmu.h> 915084Sjohnlev #include <vm/hat_pte.h> 920Sstevel@tonic-gate 930Sstevel@tonic-gate /* 940Sstevel@tonic-gate * cpu0 and default tables and structures. 950Sstevel@tonic-gate */ 963446Smrj user_desc_t *gdt0; 975084Sjohnlev #if !defined(__xpv) 980Sstevel@tonic-gate desctbr_t gdt0_default_r; 995084Sjohnlev #endif 1000Sstevel@tonic-gate 1015460Sjosephb gate_desc_t *idt0; /* interrupt descriptor table */ 1023446Smrj #if defined(__i386) 1030Sstevel@tonic-gate desctbr_t idt0_default_r; /* describes idt0 in IDTR format */ 1043446Smrj #endif 1050Sstevel@tonic-gate 1065460Sjosephb struct tss *ktss0; /* kernel task state structure */ 1070Sstevel@tonic-gate 1080Sstevel@tonic-gate #if defined(__i386) 1095460Sjosephb struct tss *dftss0; /* #DF double-fault exception */ 1100Sstevel@tonic-gate #endif /* __i386 */ 1110Sstevel@tonic-gate 1120Sstevel@tonic-gate user_desc_t zero_udesc; /* base zero user desc native procs */ 1135084Sjohnlev user_desc_t null_udesc; /* null user descriptor */ 1145084Sjohnlev system_desc_t null_sdesc; /* null system descriptor */ 1150Sstevel@tonic-gate 1160Sstevel@tonic-gate #if defined(__amd64) 1170Sstevel@tonic-gate user_desc_t zero_u32desc; /* 32-bit compatibility procs */ 1180Sstevel@tonic-gate #endif /* __amd64 */ 1190Sstevel@tonic-gate 1205084Sjohnlev #if defined(__amd64) 1215084Sjohnlev user_desc_t ucs_on; 1225084Sjohnlev user_desc_t ucs_off; 1235084Sjohnlev user_desc_t ucs32_on; 1245084Sjohnlev user_desc_t ucs32_off; 1255084Sjohnlev #endif /* __amd64 */ 1265084Sjohnlev 1270Sstevel@tonic-gate #pragma align 16(dblfault_stack0) 1280Sstevel@tonic-gate char dblfault_stack0[DEFAULTSTKSZ]; 1290Sstevel@tonic-gate 1300Sstevel@tonic-gate extern void fast_null(void); 1310Sstevel@tonic-gate extern hrtime_t get_hrtime(void); 1320Sstevel@tonic-gate extern hrtime_t gethrvtime(void); 1330Sstevel@tonic-gate extern hrtime_t get_hrestime(void); 1340Sstevel@tonic-gate extern uint64_t getlgrp(void); 1350Sstevel@tonic-gate 1360Sstevel@tonic-gate void (*(fasttable[]))(void) = { 1370Sstevel@tonic-gate fast_null, /* T_FNULL routine */ 1380Sstevel@tonic-gate fast_null, /* T_FGETFP routine (initially null) */ 1390Sstevel@tonic-gate fast_null, /* T_FSETFP routine (initially null) */ 1400Sstevel@tonic-gate (void (*)())get_hrtime, /* T_GETHRTIME */ 1410Sstevel@tonic-gate (void (*)())gethrvtime, /* T_GETHRVTIME */ 1420Sstevel@tonic-gate (void (*)())get_hrestime, /* T_GETHRESTIME */ 1430Sstevel@tonic-gate (void (*)())getlgrp /* T_GETLGRP */ 1440Sstevel@tonic-gate }; 1450Sstevel@tonic-gate 1460Sstevel@tonic-gate /* 1472712Snn35248 * Structure containing pre-computed descriptors to allow us to temporarily 1482712Snn35248 * interpose on a standard handler. 1492712Snn35248 */ 1502712Snn35248 struct interposing_handler { 1512712Snn35248 int ih_inum; 1522712Snn35248 gate_desc_t ih_interp_desc; 1532712Snn35248 gate_desc_t ih_default_desc; 1542712Snn35248 }; 1552712Snn35248 1562712Snn35248 /* 1572712Snn35248 * The brand infrastructure interposes on two handlers, and we use one as a 1582712Snn35248 * NULL signpost. 1592712Snn35248 */ 160*12613SSurya.Prakki@Sun.COM static struct interposing_handler brand_tbl[2]; 1612712Snn35248 1622712Snn35248 /* 1630Sstevel@tonic-gate * software prototypes for default local descriptor table 1640Sstevel@tonic-gate */ 1650Sstevel@tonic-gate 1660Sstevel@tonic-gate /* 1670Sstevel@tonic-gate * Routines for loading segment descriptors in format the hardware 1680Sstevel@tonic-gate * can understand. 1690Sstevel@tonic-gate */ 1700Sstevel@tonic-gate 1710Sstevel@tonic-gate #if defined(__amd64) 1720Sstevel@tonic-gate 1730Sstevel@tonic-gate /* 1740Sstevel@tonic-gate * In long mode we have the new L or long mode attribute bit 1750Sstevel@tonic-gate * for code segments. Only the conforming bit in type is used along 1760Sstevel@tonic-gate * with descriptor priority and present bits. Default operand size must 1770Sstevel@tonic-gate * be zero when in long mode. In 32-bit compatibility mode all fields 1780Sstevel@tonic-gate * are treated as in legacy mode. For data segments while in long mode 1790Sstevel@tonic-gate * only the present bit is loaded. 1800Sstevel@tonic-gate */ 1810Sstevel@tonic-gate void 1820Sstevel@tonic-gate set_usegd(user_desc_t *dp, uint_t lmode, void *base, size_t size, 1830Sstevel@tonic-gate uint_t type, uint_t dpl, uint_t gran, uint_t defopsz) 1840Sstevel@tonic-gate { 1850Sstevel@tonic-gate ASSERT(lmode == SDP_SHORT || lmode == SDP_LONG); 1860Sstevel@tonic-gate 1870Sstevel@tonic-gate /* 1880Sstevel@tonic-gate * 64-bit long mode. 1890Sstevel@tonic-gate */ 1900Sstevel@tonic-gate if (lmode == SDP_LONG) 1910Sstevel@tonic-gate dp->usd_def32 = 0; /* 32-bit operands only */ 1920Sstevel@tonic-gate else 1930Sstevel@tonic-gate /* 1940Sstevel@tonic-gate * 32-bit compatibility mode. 1950Sstevel@tonic-gate */ 1960Sstevel@tonic-gate dp->usd_def32 = defopsz; /* 0 = 16, 1 = 32-bit ops */ 1970Sstevel@tonic-gate 1980Sstevel@tonic-gate dp->usd_long = lmode; /* 64-bit mode */ 1990Sstevel@tonic-gate dp->usd_type = type; 2000Sstevel@tonic-gate dp->usd_dpl = dpl; 2010Sstevel@tonic-gate dp->usd_p = 1; 2020Sstevel@tonic-gate dp->usd_gran = gran; /* 0 = bytes, 1 = pages */ 2030Sstevel@tonic-gate 2040Sstevel@tonic-gate dp->usd_lobase = (uintptr_t)base; 2050Sstevel@tonic-gate dp->usd_midbase = (uintptr_t)base >> 16; 2060Sstevel@tonic-gate dp->usd_hibase = (uintptr_t)base >> (16 + 8); 2070Sstevel@tonic-gate dp->usd_lolimit = size; 2080Sstevel@tonic-gate dp->usd_hilimit = (uintptr_t)size >> 16; 2090Sstevel@tonic-gate } 2100Sstevel@tonic-gate 2110Sstevel@tonic-gate #elif defined(__i386) 2120Sstevel@tonic-gate 2130Sstevel@tonic-gate /* 2140Sstevel@tonic-gate * Install user segment descriptor for code and data. 2150Sstevel@tonic-gate */ 2160Sstevel@tonic-gate void 2170Sstevel@tonic-gate set_usegd(user_desc_t *dp, void *base, size_t size, uint_t type, 2180Sstevel@tonic-gate uint_t dpl, uint_t gran, uint_t defopsz) 2190Sstevel@tonic-gate { 2200Sstevel@tonic-gate dp->usd_lolimit = size; 2210Sstevel@tonic-gate dp->usd_hilimit = (uintptr_t)size >> 16; 2220Sstevel@tonic-gate 2230Sstevel@tonic-gate dp->usd_lobase = (uintptr_t)base; 2240Sstevel@tonic-gate dp->usd_midbase = (uintptr_t)base >> 16; 2250Sstevel@tonic-gate dp->usd_hibase = (uintptr_t)base >> (16 + 8); 2260Sstevel@tonic-gate 2270Sstevel@tonic-gate dp->usd_type = type; 2280Sstevel@tonic-gate dp->usd_dpl = dpl; 2290Sstevel@tonic-gate dp->usd_p = 1; 2300Sstevel@tonic-gate dp->usd_def32 = defopsz; /* 0 = 16, 1 = 32 bit operands */ 2310Sstevel@tonic-gate dp->usd_gran = gran; /* 0 = bytes, 1 = pages */ 2320Sstevel@tonic-gate } 2330Sstevel@tonic-gate 2340Sstevel@tonic-gate #endif /* __i386 */ 2350Sstevel@tonic-gate 2360Sstevel@tonic-gate /* 2370Sstevel@tonic-gate * Install system segment descriptor for LDT and TSS segments. 2380Sstevel@tonic-gate */ 2390Sstevel@tonic-gate 2400Sstevel@tonic-gate #if defined(__amd64) 2410Sstevel@tonic-gate 2420Sstevel@tonic-gate void 2430Sstevel@tonic-gate set_syssegd(system_desc_t *dp, void *base, size_t size, uint_t type, 2440Sstevel@tonic-gate uint_t dpl) 2450Sstevel@tonic-gate { 2460Sstevel@tonic-gate dp->ssd_lolimit = size; 2470Sstevel@tonic-gate dp->ssd_hilimit = (uintptr_t)size >> 16; 2480Sstevel@tonic-gate 2490Sstevel@tonic-gate dp->ssd_lobase = (uintptr_t)base; 2500Sstevel@tonic-gate dp->ssd_midbase = (uintptr_t)base >> 16; 2510Sstevel@tonic-gate dp->ssd_hibase = (uintptr_t)base >> (16 + 8); 2520Sstevel@tonic-gate dp->ssd_hi64base = (uintptr_t)base >> (16 + 8 + 8); 2530Sstevel@tonic-gate 2540Sstevel@tonic-gate dp->ssd_type = type; 2550Sstevel@tonic-gate dp->ssd_zero1 = 0; /* must be zero */ 2560Sstevel@tonic-gate dp->ssd_zero2 = 0; 2570Sstevel@tonic-gate dp->ssd_dpl = dpl; 2580Sstevel@tonic-gate dp->ssd_p = 1; 2590Sstevel@tonic-gate dp->ssd_gran = 0; /* force byte units */ 2600Sstevel@tonic-gate } 2610Sstevel@tonic-gate 2625084Sjohnlev void * 2635084Sjohnlev get_ssd_base(system_desc_t *dp) 2645084Sjohnlev { 2655084Sjohnlev uintptr_t base; 2665084Sjohnlev 2675084Sjohnlev base = (uintptr_t)dp->ssd_lobase | 2685084Sjohnlev (uintptr_t)dp->ssd_midbase << 16 | 2695084Sjohnlev (uintptr_t)dp->ssd_hibase << (16 + 8) | 2705084Sjohnlev (uintptr_t)dp->ssd_hi64base << (16 + 8 + 8); 2715084Sjohnlev return ((void *)base); 2725084Sjohnlev } 2735084Sjohnlev 2740Sstevel@tonic-gate #elif defined(__i386) 2750Sstevel@tonic-gate 2760Sstevel@tonic-gate void 2770Sstevel@tonic-gate set_syssegd(system_desc_t *dp, void *base, size_t size, uint_t type, 2780Sstevel@tonic-gate uint_t dpl) 2790Sstevel@tonic-gate { 2800Sstevel@tonic-gate dp->ssd_lolimit = size; 2810Sstevel@tonic-gate dp->ssd_hilimit = (uintptr_t)size >> 16; 2820Sstevel@tonic-gate 2830Sstevel@tonic-gate dp->ssd_lobase = (uintptr_t)base; 2840Sstevel@tonic-gate dp->ssd_midbase = (uintptr_t)base >> 16; 2850Sstevel@tonic-gate dp->ssd_hibase = (uintptr_t)base >> (16 + 8); 2860Sstevel@tonic-gate 2870Sstevel@tonic-gate dp->ssd_type = type; 2880Sstevel@tonic-gate dp->ssd_zero = 0; /* must be zero */ 2890Sstevel@tonic-gate dp->ssd_dpl = dpl; 2900Sstevel@tonic-gate dp->ssd_p = 1; 2910Sstevel@tonic-gate dp->ssd_gran = 0; /* force byte units */ 2920Sstevel@tonic-gate } 2930Sstevel@tonic-gate 2945084Sjohnlev void * 2955084Sjohnlev get_ssd_base(system_desc_t *dp) 2965084Sjohnlev { 2975084Sjohnlev uintptr_t base; 2985084Sjohnlev 2995084Sjohnlev base = (uintptr_t)dp->ssd_lobase | 3005084Sjohnlev (uintptr_t)dp->ssd_midbase << 16 | 3015084Sjohnlev (uintptr_t)dp->ssd_hibase << (16 + 8); 3025084Sjohnlev return ((void *)base); 3035084Sjohnlev } 3045084Sjohnlev 3050Sstevel@tonic-gate #endif /* __i386 */ 3060Sstevel@tonic-gate 3070Sstevel@tonic-gate /* 3080Sstevel@tonic-gate * Install gate segment descriptor for interrupt, trap, call and task gates. 3090Sstevel@tonic-gate */ 3100Sstevel@tonic-gate 3110Sstevel@tonic-gate #if defined(__amd64) 3120Sstevel@tonic-gate 3138679SSeth.Goldberg@Sun.COM /*ARGSUSED*/ 3140Sstevel@tonic-gate void 3153446Smrj set_gatesegd(gate_desc_t *dp, void (*func)(void), selector_t sel, 3168679SSeth.Goldberg@Sun.COM uint_t type, uint_t dpl, uint_t vector) 3170Sstevel@tonic-gate { 3180Sstevel@tonic-gate dp->sgd_looffset = (uintptr_t)func; 3190Sstevel@tonic-gate dp->sgd_hioffset = (uintptr_t)func >> 16; 3200Sstevel@tonic-gate dp->sgd_hi64offset = (uintptr_t)func >> (16 + 16); 3210Sstevel@tonic-gate 3220Sstevel@tonic-gate dp->sgd_selector = (uint16_t)sel; 3233446Smrj 3243446Smrj /* 3253446Smrj * For 64 bit native we use the IST stack mechanism 3263446Smrj * for double faults. All other traps use the CPL = 0 3273446Smrj * (tss_rsp0) stack. 3283446Smrj */ 3295084Sjohnlev #if !defined(__xpv) 3308679SSeth.Goldberg@Sun.COM if (vector == T_DBLFLT) 3313446Smrj dp->sgd_ist = 1; 3323446Smrj else 3335084Sjohnlev #endif 3343446Smrj dp->sgd_ist = 0; 3353446Smrj 3360Sstevel@tonic-gate dp->sgd_type = type; 3370Sstevel@tonic-gate dp->sgd_dpl = dpl; 3380Sstevel@tonic-gate dp->sgd_p = 1; 3390Sstevel@tonic-gate } 3400Sstevel@tonic-gate 3410Sstevel@tonic-gate #elif defined(__i386) 3420Sstevel@tonic-gate 3438679SSeth.Goldberg@Sun.COM /*ARGSUSED*/ 3440Sstevel@tonic-gate void 3450Sstevel@tonic-gate set_gatesegd(gate_desc_t *dp, void (*func)(void), selector_t sel, 3468679SSeth.Goldberg@Sun.COM uint_t type, uint_t dpl, uint_t unused) 3470Sstevel@tonic-gate { 3480Sstevel@tonic-gate dp->sgd_looffset = (uintptr_t)func; 3490Sstevel@tonic-gate dp->sgd_hioffset = (uintptr_t)func >> 16; 3500Sstevel@tonic-gate 3510Sstevel@tonic-gate dp->sgd_selector = (uint16_t)sel; 3523446Smrj dp->sgd_stkcpy = 0; /* always zero bytes */ 3530Sstevel@tonic-gate dp->sgd_type = type; 3540Sstevel@tonic-gate dp->sgd_dpl = dpl; 3550Sstevel@tonic-gate dp->sgd_p = 1; 3560Sstevel@tonic-gate } 3570Sstevel@tonic-gate 3583446Smrj #endif /* __i386 */ 3593446Smrj 3605084Sjohnlev /* 3615084Sjohnlev * Updates a single user descriptor in the the GDT of the current cpu. 3625084Sjohnlev * Caller is responsible for preventing cpu migration. 3635084Sjohnlev */ 3645084Sjohnlev 3655084Sjohnlev void 3665084Sjohnlev gdt_update_usegd(uint_t sidx, user_desc_t *udp) 3675084Sjohnlev { 3685084Sjohnlev #if defined(__xpv) 3695084Sjohnlev 3705084Sjohnlev uint64_t dpa = CPU->cpu_m.mcpu_gdtpa + sizeof (*udp) * sidx; 3715084Sjohnlev 3725084Sjohnlev if (HYPERVISOR_update_descriptor(pa_to_ma(dpa), *(uint64_t *)udp)) 3735084Sjohnlev panic("gdt_update_usegd: HYPERVISOR_update_descriptor"); 3745084Sjohnlev 3755084Sjohnlev #else /* __xpv */ 3765084Sjohnlev 3775084Sjohnlev CPU->cpu_gdt[sidx] = *udp; 3785084Sjohnlev 3795084Sjohnlev #endif /* __xpv */ 3805084Sjohnlev } 3815084Sjohnlev 3825084Sjohnlev /* 3835084Sjohnlev * Writes single descriptor pointed to by udp into a processes 3845084Sjohnlev * LDT entry pointed to by ldp. 3855084Sjohnlev */ 3865084Sjohnlev int 3875084Sjohnlev ldt_update_segd(user_desc_t *ldp, user_desc_t *udp) 3885084Sjohnlev { 3895084Sjohnlev #if defined(__xpv) 3905084Sjohnlev 3915084Sjohnlev uint64_t dpa; 3925084Sjohnlev 3935084Sjohnlev dpa = mmu_ptob(hat_getpfnum(kas.a_hat, (caddr_t)ldp)) | 3945084Sjohnlev ((uintptr_t)ldp & PAGEOFFSET); 3955084Sjohnlev 3965084Sjohnlev /* 3975084Sjohnlev * The hypervisor is a little more restrictive about what it 3985084Sjohnlev * supports in the LDT. 3995084Sjohnlev */ 4005084Sjohnlev if (HYPERVISOR_update_descriptor(pa_to_ma(dpa), *(uint64_t *)udp) != 0) 4015084Sjohnlev return (EINVAL); 4025084Sjohnlev 4035084Sjohnlev #else /* __xpv */ 4045084Sjohnlev 4055084Sjohnlev *ldp = *udp; 4065084Sjohnlev 4075084Sjohnlev #endif /* __xpv */ 4085084Sjohnlev return (0); 4095084Sjohnlev } 4105084Sjohnlev 4115084Sjohnlev #if defined(__xpv) 4125084Sjohnlev 4135084Sjohnlev /* 4145084Sjohnlev * Converts hw format gate descriptor into pseudo-IDT format for the hypervisor. 4155084Sjohnlev * Returns true if a valid entry was written. 4165084Sjohnlev */ 4175084Sjohnlev int 4185084Sjohnlev xen_idt_to_trap_info(uint_t vec, gate_desc_t *sgd, void *ti_arg) 4195084Sjohnlev { 4205084Sjohnlev trap_info_t *ti = ti_arg; /* XXPV Aargh - segments.h comment */ 4215084Sjohnlev 4225084Sjohnlev /* 4235084Sjohnlev * skip holes in the IDT 4245084Sjohnlev */ 4255084Sjohnlev if (GATESEG_GETOFFSET(sgd) == 0) 4265084Sjohnlev return (0); 4275084Sjohnlev 4285084Sjohnlev ASSERT(sgd->sgd_type == SDT_SYSIGT); 4295084Sjohnlev ti->vector = vec; 4305084Sjohnlev TI_SET_DPL(ti, sgd->sgd_dpl); 4315084Sjohnlev 4325084Sjohnlev /* 4335084Sjohnlev * Is this an interrupt gate? 4345084Sjohnlev */ 4355084Sjohnlev if (sgd->sgd_type == SDT_SYSIGT) { 4365084Sjohnlev /* LINTED */ 4375084Sjohnlev TI_SET_IF(ti, 1); 4385084Sjohnlev } 4395084Sjohnlev ti->cs = sgd->sgd_selector; 4405084Sjohnlev #if defined(__amd64) 4415084Sjohnlev ti->cs |= SEL_KPL; /* force into ring 3. see KCS_SEL */ 4425084Sjohnlev #endif 4435084Sjohnlev ti->address = GATESEG_GETOFFSET(sgd); 4445084Sjohnlev return (1); 4455084Sjohnlev } 4465084Sjohnlev 4475084Sjohnlev /* 4485084Sjohnlev * Convert a single hw format gate descriptor and write it into our virtual IDT. 4495084Sjohnlev */ 4505084Sjohnlev void 4515084Sjohnlev xen_idt_write(gate_desc_t *sgd, uint_t vec) 4525084Sjohnlev { 4535084Sjohnlev trap_info_t trapinfo[2]; 4545084Sjohnlev 4555084Sjohnlev bzero(trapinfo, sizeof (trapinfo)); 4565084Sjohnlev if (xen_idt_to_trap_info(vec, sgd, &trapinfo[0]) == 0) 4575084Sjohnlev return; 4585084Sjohnlev if (xen_set_trap_table(trapinfo) != 0) 4595084Sjohnlev panic("xen_idt_write: xen_set_trap_table() failed"); 4605084Sjohnlev } 4615084Sjohnlev 4625084Sjohnlev #endif /* __xpv */ 4635084Sjohnlev 4643446Smrj #if defined(__amd64) 4650Sstevel@tonic-gate 4660Sstevel@tonic-gate /* 4670Sstevel@tonic-gate * Build kernel GDT. 4680Sstevel@tonic-gate */ 4690Sstevel@tonic-gate 4700Sstevel@tonic-gate static void 4713446Smrj init_gdt_common(user_desc_t *gdt) 4720Sstevel@tonic-gate { 4733446Smrj int i; 4740Sstevel@tonic-gate 4750Sstevel@tonic-gate /* 4760Sstevel@tonic-gate * 64-bit kernel code segment. 4770Sstevel@tonic-gate */ 4783446Smrj set_usegd(&gdt[GDT_KCODE], SDP_LONG, NULL, 0, SDT_MEMERA, SEL_KPL, 4790Sstevel@tonic-gate SDP_PAGES, SDP_OP32); 4800Sstevel@tonic-gate 4810Sstevel@tonic-gate /* 4820Sstevel@tonic-gate * 64-bit kernel data segment. The limit attribute is ignored in 64-bit 4830Sstevel@tonic-gate * mode, but we set it here to 0xFFFF so that we can use the SYSRET 4840Sstevel@tonic-gate * instruction to return from system calls back to 32-bit applications. 4850Sstevel@tonic-gate * SYSRET doesn't update the base, limit, or attributes of %ss or %ds 4860Sstevel@tonic-gate * descriptors. We therefore must ensure that the kernel uses something, 4870Sstevel@tonic-gate * though it will be ignored by hardware, that is compatible with 32-bit 4880Sstevel@tonic-gate * apps. For the same reason we must set the default op size of this 4890Sstevel@tonic-gate * descriptor to 32-bit operands. 4900Sstevel@tonic-gate */ 4913446Smrj set_usegd(&gdt[GDT_KDATA], SDP_LONG, NULL, -1, SDT_MEMRWA, 4920Sstevel@tonic-gate SEL_KPL, SDP_PAGES, SDP_OP32); 4933446Smrj gdt[GDT_KDATA].usd_def32 = 1; 4940Sstevel@tonic-gate 4950Sstevel@tonic-gate /* 4960Sstevel@tonic-gate * 64-bit user code segment. 4970Sstevel@tonic-gate */ 4983446Smrj set_usegd(&gdt[GDT_UCODE], SDP_LONG, NULL, 0, SDT_MEMERA, SEL_UPL, 4990Sstevel@tonic-gate SDP_PAGES, SDP_OP32); 5000Sstevel@tonic-gate 5010Sstevel@tonic-gate /* 5020Sstevel@tonic-gate * 32-bit user code segment. 5030Sstevel@tonic-gate */ 5043446Smrj set_usegd(&gdt[GDT_U32CODE], SDP_SHORT, NULL, -1, SDT_MEMERA, 5050Sstevel@tonic-gate SEL_UPL, SDP_PAGES, SDP_OP32); 5060Sstevel@tonic-gate 5070Sstevel@tonic-gate /* 5085084Sjohnlev * See gdt_ucode32() and gdt_ucode_native(). 5095084Sjohnlev */ 5105084Sjohnlev ucs_on = ucs_off = gdt[GDT_UCODE]; 5115084Sjohnlev ucs_off.usd_p = 0; /* forces #np fault */ 5125084Sjohnlev 5135084Sjohnlev ucs32_on = ucs32_off = gdt[GDT_U32CODE]; 5145084Sjohnlev ucs32_off.usd_p = 0; /* forces #np fault */ 5155084Sjohnlev 5165084Sjohnlev /* 5170Sstevel@tonic-gate * 32 and 64 bit data segments can actually share the same descriptor. 5180Sstevel@tonic-gate * In long mode only the present bit is checked but all other fields 5190Sstevel@tonic-gate * are loaded. But in compatibility mode all fields are interpreted 5200Sstevel@tonic-gate * as in legacy mode so they must be set correctly for a 32-bit data 5210Sstevel@tonic-gate * segment. 5220Sstevel@tonic-gate */ 5233446Smrj set_usegd(&gdt[GDT_UDATA], SDP_SHORT, NULL, -1, SDT_MEMRWA, SEL_UPL, 5240Sstevel@tonic-gate SDP_PAGES, SDP_OP32); 5250Sstevel@tonic-gate 5265084Sjohnlev #if !defined(__xpv) 5275084Sjohnlev 5280Sstevel@tonic-gate /* 5291217Srab * The 64-bit kernel has no default LDT. By default, the LDT descriptor 5301217Srab * in the GDT is 0. 5310Sstevel@tonic-gate */ 5320Sstevel@tonic-gate 5330Sstevel@tonic-gate /* 5340Sstevel@tonic-gate * Kernel TSS 5350Sstevel@tonic-gate */ 5365460Sjosephb set_syssegd((system_desc_t *)&gdt[GDT_KTSS], ktss0, 5375460Sjosephb sizeof (*ktss0) - 1, SDT_SYSTSS, SEL_KPL); 5380Sstevel@tonic-gate 5395084Sjohnlev #endif /* !__xpv */ 5405084Sjohnlev 5410Sstevel@tonic-gate /* 5420Sstevel@tonic-gate * Initialize fs and gs descriptors for 32 bit processes. 5430Sstevel@tonic-gate * Only attributes and limits are initialized, the effective 5440Sstevel@tonic-gate * base address is programmed via fsbase/gsbase. 5450Sstevel@tonic-gate */ 5463446Smrj set_usegd(&gdt[GDT_LWPFS], SDP_SHORT, NULL, -1, SDT_MEMRWA, 5470Sstevel@tonic-gate SEL_UPL, SDP_PAGES, SDP_OP32); 5483446Smrj set_usegd(&gdt[GDT_LWPGS], SDP_SHORT, NULL, -1, SDT_MEMRWA, 5490Sstevel@tonic-gate SEL_UPL, SDP_PAGES, SDP_OP32); 5500Sstevel@tonic-gate 5510Sstevel@tonic-gate /* 5522712Snn35248 * Initialize the descriptors set aside for brand usage. 5532712Snn35248 * Only attributes and limits are initialized. 5542712Snn35248 */ 5552712Snn35248 for (i = GDT_BRANDMIN; i <= GDT_BRANDMAX; i++) 5563446Smrj set_usegd(&gdt0[i], SDP_SHORT, NULL, -1, SDT_MEMRWA, 5572712Snn35248 SEL_UPL, SDP_PAGES, SDP_OP32); 5582712Snn35248 5592712Snn35248 /* 5600Sstevel@tonic-gate * Initialize convenient zero base user descriptors for clearing 5610Sstevel@tonic-gate * lwp private %fs and %gs descriptors in GDT. See setregs() for 5620Sstevel@tonic-gate * an example. 5630Sstevel@tonic-gate */ 5640Sstevel@tonic-gate set_usegd(&zero_udesc, SDP_LONG, 0, 0, SDT_MEMRWA, SEL_UPL, 5650Sstevel@tonic-gate SDP_BYTES, SDP_OP32); 5660Sstevel@tonic-gate set_usegd(&zero_u32desc, SDP_SHORT, 0, -1, SDT_MEMRWA, SEL_UPL, 5670Sstevel@tonic-gate SDP_PAGES, SDP_OP32); 5680Sstevel@tonic-gate } 5690Sstevel@tonic-gate 5705084Sjohnlev #if defined(__xpv) 5715084Sjohnlev 5725084Sjohnlev static user_desc_t * 5735084Sjohnlev init_gdt(void) 5745084Sjohnlev { 5755084Sjohnlev uint64_t gdtpa; 5765084Sjohnlev ulong_t ma[1]; /* XXPV should be a memory_t */ 5775084Sjohnlev ulong_t addr; 5785084Sjohnlev 5795084Sjohnlev #if !defined(__lint) 5805084Sjohnlev /* 5815084Sjohnlev * Our gdt is never larger than a single page. 5825084Sjohnlev */ 5835084Sjohnlev ASSERT((sizeof (*gdt0) * NGDT) <= PAGESIZE); 5845084Sjohnlev #endif 5855084Sjohnlev gdt0 = (user_desc_t *)BOP_ALLOC(bootops, (caddr_t)GDT_VA, 5865084Sjohnlev PAGESIZE, PAGESIZE); 5875084Sjohnlev bzero(gdt0, PAGESIZE); 5885084Sjohnlev 5895084Sjohnlev init_gdt_common(gdt0); 5905084Sjohnlev 5915084Sjohnlev /* 5925084Sjohnlev * XXX Since we never invoke kmdb until after the kernel takes 5935084Sjohnlev * over the descriptor tables why not have it use the kernel's 5945084Sjohnlev * selectors? 5955084Sjohnlev */ 5965084Sjohnlev if (boothowto & RB_DEBUG) { 5975084Sjohnlev set_usegd(&gdt0[GDT_B32DATA], SDP_LONG, NULL, -1, SDT_MEMRWA, 5985084Sjohnlev SEL_KPL, SDP_PAGES, SDP_OP32); 5995084Sjohnlev set_usegd(&gdt0[GDT_B64CODE], SDP_LONG, NULL, -1, SDT_MEMERA, 6005084Sjohnlev SEL_KPL, SDP_PAGES, SDP_OP32); 6015084Sjohnlev } 6025084Sjohnlev 6035084Sjohnlev /* 6045084Sjohnlev * Clear write permission for page containing the gdt and install it. 6055084Sjohnlev */ 6065084Sjohnlev gdtpa = pfn_to_pa(va_to_pfn(gdt0)); 6075084Sjohnlev ma[0] = (ulong_t)(pa_to_ma(gdtpa) >> PAGESHIFT); 6085084Sjohnlev kbm_read_only((uintptr_t)gdt0, gdtpa); 6095084Sjohnlev xen_set_gdt(ma, NGDT); 6105084Sjohnlev 6115084Sjohnlev /* 6125084Sjohnlev * Reload the segment registers to use the new GDT. 6135084Sjohnlev * On 64-bit, fixup KCS_SEL to be in ring 3. 6145084Sjohnlev * See KCS_SEL in segments.h. 6155084Sjohnlev */ 6165084Sjohnlev load_segment_registers((KCS_SEL | SEL_KPL), KFS_SEL, KGS_SEL, KDS_SEL); 6175084Sjohnlev 6185084Sjohnlev /* 6195084Sjohnlev * setup %gs for kernel 6205084Sjohnlev */ 6215084Sjohnlev xen_set_segment_base(SEGBASE_GS_KERNEL, (ulong_t)&cpus[0]); 6225084Sjohnlev 6235084Sjohnlev /* 6245084Sjohnlev * XX64 We should never dereference off "other gsbase" or 6255084Sjohnlev * "fsbase". So, we should arrange to point FSBASE and 6265084Sjohnlev * KGSBASE somewhere truly awful e.g. point it at the last 6275084Sjohnlev * valid address below the hole so that any attempts to index 6285084Sjohnlev * off them cause an exception. 6295084Sjohnlev * 6305084Sjohnlev * For now, point it at 8G -- at least it should be unmapped 6315084Sjohnlev * until some 64-bit processes run. 6325084Sjohnlev */ 6335084Sjohnlev addr = 0x200000000ul; 6345084Sjohnlev xen_set_segment_base(SEGBASE_FS, addr); 6355084Sjohnlev xen_set_segment_base(SEGBASE_GS_USER, addr); 6365084Sjohnlev xen_set_segment_base(SEGBASE_GS_USER_SEL, 0); 6375084Sjohnlev 6385084Sjohnlev return (gdt0); 6395084Sjohnlev } 6405084Sjohnlev 6415084Sjohnlev #else /* __xpv */ 6425084Sjohnlev 6433446Smrj static user_desc_t * 6440Sstevel@tonic-gate init_gdt(void) 6450Sstevel@tonic-gate { 6460Sstevel@tonic-gate desctbr_t r_bgdt, r_gdt; 6470Sstevel@tonic-gate user_desc_t *bgdt; 6483446Smrj 6493446Smrj #if !defined(__lint) 6503446Smrj /* 6513446Smrj * Our gdt is never larger than a single page. 6523446Smrj */ 6533446Smrj ASSERT((sizeof (*gdt0) * NGDT) <= PAGESIZE); 6543446Smrj #endif 6553446Smrj gdt0 = (user_desc_t *)BOP_ALLOC(bootops, (caddr_t)GDT_VA, 6563446Smrj PAGESIZE, PAGESIZE); 6573446Smrj bzero(gdt0, PAGESIZE); 6583446Smrj 6593446Smrj init_gdt_common(gdt0); 6600Sstevel@tonic-gate 6610Sstevel@tonic-gate /* 6623446Smrj * Copy in from boot's gdt to our gdt. 6633446Smrj * Entry 0 is the null descriptor by definition. 6640Sstevel@tonic-gate */ 6650Sstevel@tonic-gate rd_gdtr(&r_bgdt); 6660Sstevel@tonic-gate bgdt = (user_desc_t *)r_bgdt.dtr_base; 6670Sstevel@tonic-gate if (bgdt == NULL) 6680Sstevel@tonic-gate panic("null boot gdt"); 6690Sstevel@tonic-gate 6703446Smrj gdt0[GDT_B32DATA] = bgdt[GDT_B32DATA]; 6713446Smrj gdt0[GDT_B32CODE] = bgdt[GDT_B32CODE]; 6723446Smrj gdt0[GDT_B16CODE] = bgdt[GDT_B16CODE]; 6733446Smrj gdt0[GDT_B16DATA] = bgdt[GDT_B16DATA]; 6743446Smrj gdt0[GDT_B64CODE] = bgdt[GDT_B64CODE]; 6753446Smrj 6763446Smrj /* 6773446Smrj * Install our new GDT 6783446Smrj */ 6793446Smrj r_gdt.dtr_limit = (sizeof (*gdt0) * NGDT) - 1; 6803446Smrj r_gdt.dtr_base = (uintptr_t)gdt0; 6813446Smrj wr_gdtr(&r_gdt); 6823446Smrj 6833446Smrj /* 6843446Smrj * Reload the segment registers to use the new GDT 6853446Smrj */ 6863446Smrj load_segment_registers(KCS_SEL, KFS_SEL, KGS_SEL, KDS_SEL); 6873446Smrj 6883446Smrj /* 6893446Smrj * setup %gs for kernel 6903446Smrj */ 6913446Smrj wrmsr(MSR_AMD_GSBASE, (uint64_t)&cpus[0]); 6923446Smrj 6933446Smrj /* 6943446Smrj * XX64 We should never dereference off "other gsbase" or 6953446Smrj * "fsbase". So, we should arrange to point FSBASE and 6963446Smrj * KGSBASE somewhere truly awful e.g. point it at the last 6973446Smrj * valid address below the hole so that any attempts to index 6983446Smrj * off them cause an exception. 6993446Smrj * 7003446Smrj * For now, point it at 8G -- at least it should be unmapped 7013446Smrj * until some 64-bit processes run. 7023446Smrj */ 7033446Smrj wrmsr(MSR_AMD_FSBASE, 0x200000000ul); 7043446Smrj wrmsr(MSR_AMD_KGSBASE, 0x200000000ul); 7053446Smrj return (gdt0); 7063446Smrj } 7073446Smrj 7085084Sjohnlev #endif /* __xpv */ 7095084Sjohnlev 7103446Smrj #elif defined(__i386) 7113446Smrj 7123446Smrj static void 7133446Smrj init_gdt_common(user_desc_t *gdt) 7143446Smrj { 7153446Smrj int i; 7160Sstevel@tonic-gate 7170Sstevel@tonic-gate /* 7180Sstevel@tonic-gate * Text and data for both kernel and user span entire 32 bit 7190Sstevel@tonic-gate * address space. 7200Sstevel@tonic-gate */ 7210Sstevel@tonic-gate 7220Sstevel@tonic-gate /* 7230Sstevel@tonic-gate * kernel code segment. 7240Sstevel@tonic-gate */ 7253446Smrj set_usegd(&gdt[GDT_KCODE], NULL, -1, SDT_MEMERA, SEL_KPL, SDP_PAGES, 7260Sstevel@tonic-gate SDP_OP32); 7270Sstevel@tonic-gate 7280Sstevel@tonic-gate /* 7290Sstevel@tonic-gate * kernel data segment. 7300Sstevel@tonic-gate */ 7313446Smrj set_usegd(&gdt[GDT_KDATA], NULL, -1, SDT_MEMRWA, SEL_KPL, SDP_PAGES, 7320Sstevel@tonic-gate SDP_OP32); 7330Sstevel@tonic-gate 7340Sstevel@tonic-gate /* 7350Sstevel@tonic-gate * user code segment. 7360Sstevel@tonic-gate */ 7373446Smrj set_usegd(&gdt[GDT_UCODE], NULL, -1, SDT_MEMERA, SEL_UPL, SDP_PAGES, 7380Sstevel@tonic-gate SDP_OP32); 7390Sstevel@tonic-gate 7400Sstevel@tonic-gate /* 7410Sstevel@tonic-gate * user data segment. 7420Sstevel@tonic-gate */ 7433446Smrj set_usegd(&gdt[GDT_UDATA], NULL, -1, SDT_MEMRWA, SEL_UPL, SDP_PAGES, 7440Sstevel@tonic-gate SDP_OP32); 7450Sstevel@tonic-gate 7465084Sjohnlev #if !defined(__xpv) 7475084Sjohnlev 7480Sstevel@tonic-gate /* 7490Sstevel@tonic-gate * TSS for T_DBLFLT (double fault) handler 7500Sstevel@tonic-gate */ 7515460Sjosephb set_syssegd((system_desc_t *)&gdt[GDT_DBFLT], dftss0, 7525460Sjosephb sizeof (*dftss0) - 1, SDT_SYSTSS, SEL_KPL); 7530Sstevel@tonic-gate 7540Sstevel@tonic-gate /* 7550Sstevel@tonic-gate * TSS for kernel 7560Sstevel@tonic-gate */ 7575460Sjosephb set_syssegd((system_desc_t *)&gdt[GDT_KTSS], ktss0, 7585460Sjosephb sizeof (*ktss0) - 1, SDT_SYSTSS, SEL_KPL); 7590Sstevel@tonic-gate 7605084Sjohnlev #endif /* !__xpv */ 7615084Sjohnlev 7620Sstevel@tonic-gate /* 7630Sstevel@tonic-gate * %gs selector for kernel 7640Sstevel@tonic-gate */ 7653446Smrj set_usegd(&gdt[GDT_GS], &cpus[0], sizeof (struct cpu) -1, SDT_MEMRWA, 7660Sstevel@tonic-gate SEL_KPL, SDP_BYTES, SDP_OP32); 7670Sstevel@tonic-gate 7680Sstevel@tonic-gate /* 7690Sstevel@tonic-gate * Initialize lwp private descriptors. 7700Sstevel@tonic-gate * Only attributes and limits are initialized, the effective 7710Sstevel@tonic-gate * base address is programmed via fsbase/gsbase. 7720Sstevel@tonic-gate */ 7733446Smrj set_usegd(&gdt[GDT_LWPFS], NULL, (size_t)-1, SDT_MEMRWA, SEL_UPL, 7740Sstevel@tonic-gate SDP_PAGES, SDP_OP32); 7753446Smrj set_usegd(&gdt[GDT_LWPGS], NULL, (size_t)-1, SDT_MEMRWA, SEL_UPL, 7760Sstevel@tonic-gate SDP_PAGES, SDP_OP32); 7770Sstevel@tonic-gate 7780Sstevel@tonic-gate /* 7792712Snn35248 * Initialize the descriptors set aside for brand usage. 7802712Snn35248 * Only attributes and limits are initialized. 7812712Snn35248 */ 7822712Snn35248 for (i = GDT_BRANDMIN; i <= GDT_BRANDMAX; i++) 7832712Snn35248 set_usegd(&gdt0[i], NULL, (size_t)-1, SDT_MEMRWA, SEL_UPL, 7842712Snn35248 SDP_PAGES, SDP_OP32); 7853446Smrj /* 7863446Smrj * Initialize convenient zero base user descriptor for clearing 7873446Smrj * lwp private %fs and %gs descriptors in GDT. See setregs() for 7883446Smrj * an example. 7893446Smrj */ 7903446Smrj set_usegd(&zero_udesc, NULL, -1, SDT_MEMRWA, SEL_UPL, 7913446Smrj SDP_BYTES, SDP_OP32); 7923446Smrj } 7933446Smrj 7945084Sjohnlev #if defined(__xpv) 7955084Sjohnlev 7965084Sjohnlev static user_desc_t * 7975084Sjohnlev init_gdt(void) 7985084Sjohnlev { 7995084Sjohnlev uint64_t gdtpa; 8005084Sjohnlev ulong_t ma[1]; /* XXPV should be a memory_t */ 8015084Sjohnlev 8025084Sjohnlev #if !defined(__lint) 8035084Sjohnlev /* 8045084Sjohnlev * Our gdt is never larger than a single page. 8055084Sjohnlev */ 8065084Sjohnlev ASSERT((sizeof (*gdt0) * NGDT) <= PAGESIZE); 8075084Sjohnlev #endif 8085084Sjohnlev gdt0 = (user_desc_t *)BOP_ALLOC(bootops, (caddr_t)GDT_VA, 8095084Sjohnlev PAGESIZE, PAGESIZE); 8105084Sjohnlev bzero(gdt0, PAGESIZE); 8115084Sjohnlev 8125084Sjohnlev init_gdt_common(gdt0); 8135084Sjohnlev gdtpa = pfn_to_pa(va_to_pfn(gdt0)); 8145084Sjohnlev 8155084Sjohnlev /* 8165084Sjohnlev * XXX Since we never invoke kmdb until after the kernel takes 8175084Sjohnlev * over the descriptor tables why not have it use the kernel's 8185084Sjohnlev * selectors? 8195084Sjohnlev */ 8205084Sjohnlev if (boothowto & RB_DEBUG) { 8215084Sjohnlev set_usegd(&gdt0[GDT_B32DATA], NULL, -1, SDT_MEMRWA, SEL_KPL, 8225084Sjohnlev SDP_PAGES, SDP_OP32); 8235084Sjohnlev set_usegd(&gdt0[GDT_B32CODE], NULL, -1, SDT_MEMERA, SEL_KPL, 8245084Sjohnlev SDP_PAGES, SDP_OP32); 8255084Sjohnlev } 8265084Sjohnlev 8275084Sjohnlev /* 8285084Sjohnlev * Clear write permission for page containing the gdt and install it. 8295084Sjohnlev */ 8305084Sjohnlev ma[0] = (ulong_t)(pa_to_ma(gdtpa) >> PAGESHIFT); 8315084Sjohnlev kbm_read_only((uintptr_t)gdt0, gdtpa); 8325084Sjohnlev xen_set_gdt(ma, NGDT); 8335084Sjohnlev 8345084Sjohnlev /* 8355084Sjohnlev * Reload the segment registers to use the new GDT 8365084Sjohnlev */ 8375084Sjohnlev load_segment_registers( 8385084Sjohnlev KCS_SEL, KDS_SEL, KDS_SEL, KFS_SEL, KGS_SEL, KDS_SEL); 8395084Sjohnlev 8405084Sjohnlev return (gdt0); 8415084Sjohnlev } 8425084Sjohnlev 8435084Sjohnlev #else /* __xpv */ 8445084Sjohnlev 8453446Smrj static user_desc_t * 8463446Smrj init_gdt(void) 8473446Smrj { 8483446Smrj desctbr_t r_bgdt, r_gdt; 8493446Smrj user_desc_t *bgdt; 8503446Smrj 8513446Smrj #if !defined(__lint) 8523446Smrj /* 8533446Smrj * Our gdt is never larger than a single page. 8543446Smrj */ 8553446Smrj ASSERT((sizeof (*gdt0) * NGDT) <= PAGESIZE); 8563446Smrj #endif 8573446Smrj /* 8583446Smrj * XXX this allocation belongs in our caller, not here. 8593446Smrj */ 8603446Smrj gdt0 = (user_desc_t *)BOP_ALLOC(bootops, (caddr_t)GDT_VA, 8613446Smrj PAGESIZE, PAGESIZE); 8623446Smrj bzero(gdt0, PAGESIZE); 8633446Smrj 8643446Smrj init_gdt_common(gdt0); 8653446Smrj 8663446Smrj /* 8673446Smrj * Copy in from boot's gdt to our gdt entries. 8683446Smrj * Entry 0 is null descriptor by definition. 8693446Smrj */ 8703446Smrj rd_gdtr(&r_bgdt); 8713446Smrj bgdt = (user_desc_t *)r_bgdt.dtr_base; 8723446Smrj if (bgdt == NULL) 8733446Smrj panic("null boot gdt"); 8743446Smrj 8753446Smrj gdt0[GDT_B32DATA] = bgdt[GDT_B32DATA]; 8763446Smrj gdt0[GDT_B32CODE] = bgdt[GDT_B32CODE]; 8773446Smrj gdt0[GDT_B16CODE] = bgdt[GDT_B16CODE]; 8783446Smrj gdt0[GDT_B16DATA] = bgdt[GDT_B16DATA]; 8792712Snn35248 8802712Snn35248 /* 8810Sstevel@tonic-gate * Install our new GDT 8820Sstevel@tonic-gate */ 8833446Smrj r_gdt.dtr_limit = (sizeof (*gdt0) * NGDT) - 1; 8840Sstevel@tonic-gate r_gdt.dtr_base = (uintptr_t)gdt0; 8850Sstevel@tonic-gate wr_gdtr(&r_gdt); 8860Sstevel@tonic-gate 8870Sstevel@tonic-gate /* 8883446Smrj * Reload the segment registers to use the new GDT 8890Sstevel@tonic-gate */ 8903446Smrj load_segment_registers( 8913446Smrj KCS_SEL, KDS_SEL, KDS_SEL, KFS_SEL, KGS_SEL, KDS_SEL); 8923446Smrj 8933446Smrj return (gdt0); 8940Sstevel@tonic-gate } 8950Sstevel@tonic-gate 8965084Sjohnlev #endif /* __xpv */ 8970Sstevel@tonic-gate #endif /* __i386 */ 8980Sstevel@tonic-gate 8990Sstevel@tonic-gate /* 9000Sstevel@tonic-gate * Build kernel IDT. 9010Sstevel@tonic-gate * 9023446Smrj * Note that for amd64 we pretty much require every gate to be an interrupt 9033446Smrj * gate which blocks interrupts atomically on entry; that's because of our 9043446Smrj * dependency on using 'swapgs' every time we come into the kernel to find 9053446Smrj * the cpu structure. If we get interrupted just before doing that, %cs could 9063446Smrj * be in kernel mode (so that the trap prolog doesn't do a swapgs), but 9073446Smrj * %gsbase is really still pointing at something in userland. Bad things will 9083446Smrj * ensue. We also use interrupt gates for i386 as well even though this is not 9093446Smrj * required for some traps. 9100Sstevel@tonic-gate * 9110Sstevel@tonic-gate * Perhaps they should have invented a trap gate that does an atomic swapgs? 9120Sstevel@tonic-gate */ 9130Sstevel@tonic-gate static void 9143446Smrj init_idt_common(gate_desc_t *idt) 9153446Smrj { 9168679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_ZERODIV], &div0trap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 9178679SSeth.Goldberg@Sun.COM 0); 9188679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_SGLSTP], &dbgtrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 9198679SSeth.Goldberg@Sun.COM 0); 9208679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_NMIFLT], &nmiint, KCS_SEL, SDT_SYSIGT, TRP_KPL, 9218679SSeth.Goldberg@Sun.COM 0); 9228679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_BPTFLT], &brktrap, KCS_SEL, SDT_SYSIGT, TRP_UPL, 9238679SSeth.Goldberg@Sun.COM 0); 9248679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_OVFLW], &ovflotrap, KCS_SEL, SDT_SYSIGT, TRP_UPL, 9258679SSeth.Goldberg@Sun.COM 0); 9263446Smrj set_gatesegd(&idt[T_BOUNDFLT], &boundstrap, KCS_SEL, SDT_SYSIGT, 9278679SSeth.Goldberg@Sun.COM TRP_KPL, 0); 9288679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_ILLINST], &invoptrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 9298679SSeth.Goldberg@Sun.COM 0); 9308679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_NOEXTFLT], &ndptrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 9318679SSeth.Goldberg@Sun.COM 0); 9323446Smrj 9333446Smrj /* 9343446Smrj * double fault handler. 9355084Sjohnlev * 9365084Sjohnlev * Note that on the hypervisor a guest does not receive #df faults. 9375084Sjohnlev * Instead a failsafe event is injected into the guest if its selectors 9385084Sjohnlev * and/or stack is in a broken state. See xen_failsafe_callback. 9393446Smrj */ 9405084Sjohnlev #if !defined(__xpv) 9413446Smrj #if defined(__amd64) 9425084Sjohnlev 9438679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_DBLFLT], &syserrtrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 9448679SSeth.Goldberg@Sun.COM T_DBLFLT); 9455084Sjohnlev 9463446Smrj #elif defined(__i386) 9475084Sjohnlev 9483446Smrj /* 9493446Smrj * task gate required. 9503446Smrj */ 9518679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_DBLFLT], NULL, DFTSS_SEL, SDT_SYSTASKGT, TRP_KPL, 9528679SSeth.Goldberg@Sun.COM 0); 9533446Smrj 9543446Smrj #endif /* __i386 */ 9555084Sjohnlev #endif /* !__xpv */ 9563446Smrj 9573446Smrj /* 9583446Smrj * T_EXTOVRFLT coprocessor-segment-overrun not supported. 9593446Smrj */ 9603446Smrj 9618679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_TSSFLT], &invtsstrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 9628679SSeth.Goldberg@Sun.COM 0); 9638679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_SEGFLT], &segnptrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 9648679SSeth.Goldberg@Sun.COM 0); 9658679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_STKFLT], &stktrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 0); 9668679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_GPFLT], &gptrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 0); 9678679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_PGFLT], &pftrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 0); 9688679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_EXTERRFLT], &ndperr, KCS_SEL, SDT_SYSIGT, TRP_KPL, 9698679SSeth.Goldberg@Sun.COM 0); 9703446Smrj set_gatesegd(&idt[T_ALIGNMENT], &achktrap, KCS_SEL, SDT_SYSIGT, 9718679SSeth.Goldberg@Sun.COM TRP_KPL, 0); 9728679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_MCE], &mcetrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 0); 9738679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_SIMDFPE], &xmtrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 0); 9743446Smrj 9753446Smrj /* 9763446Smrj * install fast trap handler at 210. 9773446Smrj */ 9788679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_FASTTRAP], &fasttrap, KCS_SEL, SDT_SYSIGT, TRP_UPL, 9798679SSeth.Goldberg@Sun.COM 0); 9803446Smrj 9813446Smrj /* 9823446Smrj * System call handler. 9833446Smrj */ 9843446Smrj #if defined(__amd64) 9853446Smrj set_gatesegd(&idt[T_SYSCALLINT], &sys_syscall_int, KCS_SEL, SDT_SYSIGT, 9868679SSeth.Goldberg@Sun.COM TRP_UPL, 0); 9873446Smrj 9883446Smrj #elif defined(__i386) 9893446Smrj set_gatesegd(&idt[T_SYSCALLINT], &sys_call, KCS_SEL, SDT_SYSIGT, 9908679SSeth.Goldberg@Sun.COM TRP_UPL, 0); 9913446Smrj #endif /* __i386 */ 9923446Smrj 9933446Smrj /* 9943446Smrj * Install the DTrace interrupt handler for the pid provider. 9953446Smrj */ 9963446Smrj set_gatesegd(&idt[T_DTRACE_RET], &dtrace_ret, KCS_SEL, 9978679SSeth.Goldberg@Sun.COM SDT_SYSIGT, TRP_UPL, 0); 9983446Smrj 9993446Smrj /* 1000*12613SSurya.Prakki@Sun.COM * Prepare interposing descriptor for the syscall handler 1001*12613SSurya.Prakki@Sun.COM * and cache copy of the default descriptor. 10023446Smrj */ 1003*12613SSurya.Prakki@Sun.COM brand_tbl[0].ih_inum = T_SYSCALLINT; 1004*12613SSurya.Prakki@Sun.COM brand_tbl[0].ih_default_desc = idt0[T_SYSCALLINT]; 10053446Smrj 10063446Smrj #if defined(__amd64) 1007*12613SSurya.Prakki@Sun.COM set_gatesegd(&(brand_tbl[0].ih_interp_desc), &brand_sys_syscall_int, 10088679SSeth.Goldberg@Sun.COM KCS_SEL, SDT_SYSIGT, TRP_UPL, 0); 10093446Smrj #elif defined(__i386) 1010*12613SSurya.Prakki@Sun.COM set_gatesegd(&(brand_tbl[0].ih_interp_desc), &brand_sys_call, 10118679SSeth.Goldberg@Sun.COM KCS_SEL, SDT_SYSIGT, TRP_UPL, 0); 10123446Smrj #endif /* __i386 */ 10133446Smrj 1014*12613SSurya.Prakki@Sun.COM brand_tbl[1].ih_inum = 0; 10153446Smrj } 10163446Smrj 10175084Sjohnlev #if defined(__xpv) 10185084Sjohnlev 10195084Sjohnlev static void 10205084Sjohnlev init_idt(gate_desc_t *idt) 10215084Sjohnlev { 10225084Sjohnlev init_idt_common(idt); 10235084Sjohnlev } 10245084Sjohnlev 10255084Sjohnlev #else /* __xpv */ 10265084Sjohnlev 10273446Smrj static void 10283446Smrj init_idt(gate_desc_t *idt) 10290Sstevel@tonic-gate { 10300Sstevel@tonic-gate char ivctname[80]; 10310Sstevel@tonic-gate void (*ivctptr)(void); 10320Sstevel@tonic-gate int i; 10330Sstevel@tonic-gate 10340Sstevel@tonic-gate /* 10350Sstevel@tonic-gate * Initialize entire table with 'reserved' trap and then overwrite 10360Sstevel@tonic-gate * specific entries. T_EXTOVRFLT (9) is unsupported and reserved 10370Sstevel@tonic-gate * since it can only be generated on a 386 processor. 15 is also 10380Sstevel@tonic-gate * unsupported and reserved. 10390Sstevel@tonic-gate */ 10400Sstevel@tonic-gate for (i = 0; i < NIDT; i++) 10418679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[i], &resvtrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 10428679SSeth.Goldberg@Sun.COM 0); 10430Sstevel@tonic-gate 10440Sstevel@tonic-gate /* 10450Sstevel@tonic-gate * 20-31 reserved 10460Sstevel@tonic-gate */ 10470Sstevel@tonic-gate for (i = 20; i < 32; i++) 10488679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[i], &invaltrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 10498679SSeth.Goldberg@Sun.COM 0); 10500Sstevel@tonic-gate 10510Sstevel@tonic-gate /* 10520Sstevel@tonic-gate * interrupts 32 - 255 10530Sstevel@tonic-gate */ 10540Sstevel@tonic-gate for (i = 32; i < 256; i++) { 10550Sstevel@tonic-gate (void) snprintf(ivctname, sizeof (ivctname), "ivct%d", i); 10560Sstevel@tonic-gate ivctptr = (void (*)(void))kobj_getsymvalue(ivctname, 0); 10570Sstevel@tonic-gate if (ivctptr == NULL) 10580Sstevel@tonic-gate panic("kobj_getsymvalue(%s) failed", ivctname); 10590Sstevel@tonic-gate 10608679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[i], ivctptr, KCS_SEL, SDT_SYSIGT, TRP_KPL, 0); 10610Sstevel@tonic-gate } 10620Sstevel@tonic-gate 10630Sstevel@tonic-gate /* 10643446Smrj * Now install the common ones. Note that it will overlay some 10653446Smrj * entries installed above like T_SYSCALLINT, T_FASTTRAP etc. 10660Sstevel@tonic-gate */ 10673446Smrj init_idt_common(idt); 10680Sstevel@tonic-gate } 10690Sstevel@tonic-gate 10705084Sjohnlev #endif /* __xpv */ 10715084Sjohnlev 10720Sstevel@tonic-gate /* 10731217Srab * The kernel does not deal with LDTs unless a user explicitly creates 10741217Srab * one. Under normal circumstances, the LDTR contains 0. Any process attempting 10751217Srab * to reference the LDT will therefore cause a #gp. System calls made via the 10761217Srab * obsolete lcall mechanism are emulated by the #gp fault handler. 10770Sstevel@tonic-gate */ 10780Sstevel@tonic-gate static void 10790Sstevel@tonic-gate init_ldt(void) 10800Sstevel@tonic-gate { 10815084Sjohnlev #if defined(__xpv) 10825084Sjohnlev xen_set_ldt(NULL, 0); 10835084Sjohnlev #else 10841217Srab wr_ldtr(0); 10855084Sjohnlev #endif 10860Sstevel@tonic-gate } 10870Sstevel@tonic-gate 10885084Sjohnlev #if !defined(__xpv) 10890Sstevel@tonic-gate #if defined(__amd64) 10900Sstevel@tonic-gate 10910Sstevel@tonic-gate static void 10920Sstevel@tonic-gate init_tss(void) 10930Sstevel@tonic-gate { 10940Sstevel@tonic-gate /* 10950Sstevel@tonic-gate * tss_rsp0 is dynamically filled in by resume() on each context switch. 10960Sstevel@tonic-gate * All exceptions but #DF will run on the thread stack. 10970Sstevel@tonic-gate * Set up the double fault stack here. 10980Sstevel@tonic-gate */ 10995460Sjosephb ktss0->tss_ist1 = 11000Sstevel@tonic-gate (uint64_t)&dblfault_stack0[sizeof (dblfault_stack0)]; 11010Sstevel@tonic-gate 11020Sstevel@tonic-gate /* 11030Sstevel@tonic-gate * Set I/O bit map offset equal to size of TSS segment limit 11040Sstevel@tonic-gate * for no I/O permission map. This will force all user I/O 11050Sstevel@tonic-gate * instructions to generate #gp fault. 11060Sstevel@tonic-gate */ 11075460Sjosephb ktss0->tss_bitmapbase = sizeof (*ktss0); 11080Sstevel@tonic-gate 11090Sstevel@tonic-gate /* 11100Sstevel@tonic-gate * Point %tr to descriptor for ktss0 in gdt. 11110Sstevel@tonic-gate */ 11120Sstevel@tonic-gate wr_tsr(KTSS_SEL); 11130Sstevel@tonic-gate } 11140Sstevel@tonic-gate 11150Sstevel@tonic-gate #elif defined(__i386) 11160Sstevel@tonic-gate 11170Sstevel@tonic-gate static void 11180Sstevel@tonic-gate init_tss(void) 11190Sstevel@tonic-gate { 11200Sstevel@tonic-gate /* 11215460Sjosephb * ktss0->tss_esp dynamically filled in by resume() on each 11220Sstevel@tonic-gate * context switch. 11230Sstevel@tonic-gate */ 11245460Sjosephb ktss0->tss_ss0 = KDS_SEL; 11255460Sjosephb ktss0->tss_eip = (uint32_t)_start; 11265460Sjosephb ktss0->tss_ds = ktss0->tss_es = ktss0->tss_ss = KDS_SEL; 11275460Sjosephb ktss0->tss_cs = KCS_SEL; 11285460Sjosephb ktss0->tss_fs = KFS_SEL; 11295460Sjosephb ktss0->tss_gs = KGS_SEL; 11305460Sjosephb ktss0->tss_ldt = ULDT_SEL; 11310Sstevel@tonic-gate 11320Sstevel@tonic-gate /* 11330Sstevel@tonic-gate * Initialize double fault tss. 11340Sstevel@tonic-gate */ 11355460Sjosephb dftss0->tss_esp0 = (uint32_t)&dblfault_stack0[sizeof (dblfault_stack0)]; 11365460Sjosephb dftss0->tss_ss0 = KDS_SEL; 11370Sstevel@tonic-gate 11380Sstevel@tonic-gate /* 11390Sstevel@tonic-gate * tss_cr3 will get initialized in hat_kern_setup() once our page 11400Sstevel@tonic-gate * tables have been setup. 11410Sstevel@tonic-gate */ 11425460Sjosephb dftss0->tss_eip = (uint32_t)syserrtrap; 11435460Sjosephb dftss0->tss_esp = (uint32_t)&dblfault_stack0[sizeof (dblfault_stack0)]; 11445460Sjosephb dftss0->tss_cs = KCS_SEL; 11455460Sjosephb dftss0->tss_ds = KDS_SEL; 11465460Sjosephb dftss0->tss_es = KDS_SEL; 11475460Sjosephb dftss0->tss_ss = KDS_SEL; 11485460Sjosephb dftss0->tss_fs = KFS_SEL; 11495460Sjosephb dftss0->tss_gs = KGS_SEL; 11500Sstevel@tonic-gate 11510Sstevel@tonic-gate /* 11520Sstevel@tonic-gate * Set I/O bit map offset equal to size of TSS segment limit 11530Sstevel@tonic-gate * for no I/O permission map. This will force all user I/O 11540Sstevel@tonic-gate * instructions to generate #gp fault. 11550Sstevel@tonic-gate */ 11565460Sjosephb ktss0->tss_bitmapbase = sizeof (*ktss0); 11570Sstevel@tonic-gate 11580Sstevel@tonic-gate /* 11590Sstevel@tonic-gate * Point %tr to descriptor for ktss0 in gdt. 11600Sstevel@tonic-gate */ 11610Sstevel@tonic-gate wr_tsr(KTSS_SEL); 11620Sstevel@tonic-gate } 11630Sstevel@tonic-gate 11640Sstevel@tonic-gate #endif /* __i386 */ 11655084Sjohnlev #endif /* !__xpv */ 11665084Sjohnlev 11675084Sjohnlev #if defined(__xpv) 11685084Sjohnlev 11695084Sjohnlev void 11705084Sjohnlev init_desctbls(void) 11715084Sjohnlev { 11725084Sjohnlev uint_t vec; 11735084Sjohnlev user_desc_t *gdt; 11745084Sjohnlev 11755084Sjohnlev /* 11765084Sjohnlev * Setup and install our GDT. 11775084Sjohnlev */ 11785084Sjohnlev gdt = init_gdt(); 11795084Sjohnlev 11805084Sjohnlev /* 11815084Sjohnlev * Store static pa of gdt to speed up pa_to_ma() translations 11825084Sjohnlev * on lwp context switches. 11835084Sjohnlev */ 11845084Sjohnlev ASSERT(IS_P2ALIGNED((uintptr_t)gdt, PAGESIZE)); 11855460Sjosephb CPU->cpu_gdt = gdt; 11865084Sjohnlev CPU->cpu_m.mcpu_gdtpa = pfn_to_pa(va_to_pfn(gdt)); 11875084Sjohnlev 11885084Sjohnlev /* 11895084Sjohnlev * Setup and install our IDT. 11905084Sjohnlev */ 11915460Sjosephb #if !defined(__lint) 11925460Sjosephb ASSERT(NIDT * sizeof (*idt0) <= PAGESIZE); 11935460Sjosephb #endif 11945460Sjosephb idt0 = (gate_desc_t *)BOP_ALLOC(bootops, (caddr_t)IDT_VA, 11955460Sjosephb PAGESIZE, PAGESIZE); 11968679SSeth.Goldberg@Sun.COM bzero(idt0, PAGESIZE); 11975460Sjosephb init_idt(idt0); 11985084Sjohnlev for (vec = 0; vec < NIDT; vec++) 11995084Sjohnlev xen_idt_write(&idt0[vec], vec); 12005084Sjohnlev 12015460Sjosephb CPU->cpu_idt = idt0; 12025084Sjohnlev 12035084Sjohnlev /* 12045084Sjohnlev * set default kernel stack 12055084Sjohnlev */ 12065084Sjohnlev xen_stack_switch(KDS_SEL, 12075084Sjohnlev (ulong_t)&dblfault_stack0[sizeof (dblfault_stack0)]); 12085084Sjohnlev 12095084Sjohnlev xen_init_callbacks(); 12105084Sjohnlev 12115084Sjohnlev init_ldt(); 12125084Sjohnlev } 12135084Sjohnlev 12145084Sjohnlev #else /* __xpv */ 12150Sstevel@tonic-gate 12160Sstevel@tonic-gate void 12173446Smrj init_desctbls(void) 12180Sstevel@tonic-gate { 12193446Smrj user_desc_t *gdt; 12203446Smrj desctbr_t idtr; 12213446Smrj 12223446Smrj /* 12235460Sjosephb * Allocate IDT and TSS structures on unique pages for better 12245460Sjosephb * performance in virtual machines. 12255460Sjosephb */ 12265460Sjosephb #if !defined(__lint) 12275460Sjosephb ASSERT(NIDT * sizeof (*idt0) <= PAGESIZE); 12285460Sjosephb #endif 12295460Sjosephb idt0 = (gate_desc_t *)BOP_ALLOC(bootops, (caddr_t)IDT_VA, 12305460Sjosephb PAGESIZE, PAGESIZE); 12318679SSeth.Goldberg@Sun.COM bzero(idt0, PAGESIZE); 12325460Sjosephb #if !defined(__lint) 12335460Sjosephb ASSERT(sizeof (*ktss0) <= PAGESIZE); 12345460Sjosephb #endif 12355460Sjosephb ktss0 = (struct tss *)BOP_ALLOC(bootops, (caddr_t)KTSS_VA, 12365460Sjosephb PAGESIZE, PAGESIZE); 12378679SSeth.Goldberg@Sun.COM bzero(ktss0, PAGESIZE); 12385460Sjosephb 12395460Sjosephb #if defined(__i386) 12405460Sjosephb #if !defined(__lint) 12415460Sjosephb ASSERT(sizeof (*dftss0) <= PAGESIZE); 12425460Sjosephb #endif 12435460Sjosephb dftss0 = (struct tss *)BOP_ALLOC(bootops, (caddr_t)DFTSS_VA, 12445460Sjosephb PAGESIZE, PAGESIZE); 12458679SSeth.Goldberg@Sun.COM bzero(dftss0, PAGESIZE); 12465460Sjosephb #endif 12475460Sjosephb 12485460Sjosephb /* 12493446Smrj * Setup and install our GDT. 12503446Smrj */ 12513446Smrj gdt = init_gdt(); 12523446Smrj ASSERT(IS_P2ALIGNED((uintptr_t)gdt, PAGESIZE)); 12535460Sjosephb CPU->cpu_gdt = gdt; 12543446Smrj 12553446Smrj /* 12563446Smrj * Setup and install our IDT. 12573446Smrj */ 12585460Sjosephb init_idt(idt0); 12593446Smrj 12603446Smrj idtr.dtr_base = (uintptr_t)idt0; 12615460Sjosephb idtr.dtr_limit = (NIDT * sizeof (*idt0)) - 1; 12623446Smrj wr_idtr(&idtr); 12635460Sjosephb CPU->cpu_idt = idt0; 12643446Smrj 12653446Smrj #if defined(__i386) 12663446Smrj /* 12673446Smrj * We maintain a description of idt0 in convenient IDTR format 12683446Smrj * for #pf's on some older pentium processors. See pentium_pftrap(). 12693446Smrj */ 12703446Smrj idt0_default_r = idtr; 12713446Smrj #endif /* __i386 */ 12723446Smrj 12730Sstevel@tonic-gate init_tss(); 12745460Sjosephb CPU->cpu_tss = ktss0; 12750Sstevel@tonic-gate init_ldt(); 12760Sstevel@tonic-gate } 12772712Snn35248 12785084Sjohnlev #endif /* __xpv */ 12795084Sjohnlev 12802712Snn35248 /* 12813446Smrj * In the early kernel, we need to set up a simple GDT to run on. 12825084Sjohnlev * 12835084Sjohnlev * XXPV Can dboot use this too? See dboot_gdt.s 12843446Smrj */ 12853446Smrj void 12863446Smrj init_boot_gdt(user_desc_t *bgdt) 12873446Smrj { 12883446Smrj #if defined(__amd64) 12893446Smrj set_usegd(&bgdt[GDT_B32DATA], SDP_LONG, NULL, -1, SDT_MEMRWA, SEL_KPL, 12903446Smrj SDP_PAGES, SDP_OP32); 12913446Smrj set_usegd(&bgdt[GDT_B64CODE], SDP_LONG, NULL, -1, SDT_MEMERA, SEL_KPL, 12923446Smrj SDP_PAGES, SDP_OP32); 12933446Smrj #elif defined(__i386) 12943446Smrj set_usegd(&bgdt[GDT_B32DATA], NULL, -1, SDT_MEMRWA, SEL_KPL, 12953446Smrj SDP_PAGES, SDP_OP32); 12963446Smrj set_usegd(&bgdt[GDT_B32CODE], NULL, -1, SDT_MEMERA, SEL_KPL, 12973446Smrj SDP_PAGES, SDP_OP32); 12983446Smrj #endif /* __i386 */ 12993446Smrj } 13003446Smrj 13013446Smrj /* 13022712Snn35248 * Enable interpositioning on the system call path by rewriting the 13032712Snn35248 * sys{call|enter} MSRs and the syscall-related entries in the IDT to use 13042712Snn35248 * the branded entry points. 13052712Snn35248 */ 13062712Snn35248 void 13072712Snn35248 brand_interpositioning_enable(void) 13082712Snn35248 { 13095084Sjohnlev gate_desc_t *idt = CPU->cpu_idt; 13105084Sjohnlev int i; 13115084Sjohnlev 13125084Sjohnlev ASSERT(curthread->t_preempt != 0 || getpil() >= DISP_LEVEL); 13132712Snn35248 13145084Sjohnlev for (i = 0; brand_tbl[i].ih_inum; i++) { 13155084Sjohnlev idt[brand_tbl[i].ih_inum] = brand_tbl[i].ih_interp_desc; 13165084Sjohnlev #if defined(__xpv) 13175084Sjohnlev xen_idt_write(&idt[brand_tbl[i].ih_inum], 13185084Sjohnlev brand_tbl[i].ih_inum); 13195084Sjohnlev #endif 13205084Sjohnlev } 13212712Snn35248 13222712Snn35248 #if defined(__amd64) 13235084Sjohnlev #if defined(__xpv) 13245084Sjohnlev 13255084Sjohnlev /* 13265084Sjohnlev * Currently the hypervisor only supports 64-bit syscalls via 13275084Sjohnlev * syscall instruction. The 32-bit syscalls are handled by 13285084Sjohnlev * interrupt gate above. 13295084Sjohnlev */ 13305084Sjohnlev xen_set_callback(brand_sys_syscall, CALLBACKTYPE_syscall, 13315084Sjohnlev CALLBACKF_mask_events); 13325084Sjohnlev 13335084Sjohnlev #else 13345084Sjohnlev 13355084Sjohnlev if (x86_feature & X86_ASYSC) { 13365084Sjohnlev wrmsr(MSR_AMD_LSTAR, (uintptr_t)brand_sys_syscall); 13375084Sjohnlev wrmsr(MSR_AMD_CSTAR, (uintptr_t)brand_sys_syscall32); 13385084Sjohnlev } 13395084Sjohnlev 13402712Snn35248 #endif 13415084Sjohnlev #endif /* __amd64 */ 13422712Snn35248 13432712Snn35248 if (x86_feature & X86_SEP) 13442712Snn35248 wrmsr(MSR_INTC_SEP_EIP, (uintptr_t)brand_sys_sysenter); 13452712Snn35248 } 13462712Snn35248 13472712Snn35248 /* 13482712Snn35248 * Disable interpositioning on the system call path by rewriting the 13492712Snn35248 * sys{call|enter} MSRs and the syscall-related entries in the IDT to use 13502712Snn35248 * the standard entry points, which bypass the interpositioning hooks. 13512712Snn35248 */ 13522712Snn35248 void 13532712Snn35248 brand_interpositioning_disable(void) 13542712Snn35248 { 13555084Sjohnlev gate_desc_t *idt = CPU->cpu_idt; 13562712Snn35248 int i; 13572712Snn35248 13585084Sjohnlev ASSERT(curthread->t_preempt != 0 || getpil() >= DISP_LEVEL); 13595084Sjohnlev 13605084Sjohnlev for (i = 0; brand_tbl[i].ih_inum; i++) { 13615084Sjohnlev idt[brand_tbl[i].ih_inum] = brand_tbl[i].ih_default_desc; 13625084Sjohnlev #if defined(__xpv) 13635084Sjohnlev xen_idt_write(&idt[brand_tbl[i].ih_inum], 13645084Sjohnlev brand_tbl[i].ih_inum); 13655084Sjohnlev #endif 13665084Sjohnlev } 13672712Snn35248 13682712Snn35248 #if defined(__amd64) 13695084Sjohnlev #if defined(__xpv) 13705084Sjohnlev 13715084Sjohnlev /* 13725084Sjohnlev * See comment above in brand_interpositioning_enable. 13735084Sjohnlev */ 13745084Sjohnlev xen_set_callback(sys_syscall, CALLBACKTYPE_syscall, 13755084Sjohnlev CALLBACKF_mask_events); 13765084Sjohnlev 13775084Sjohnlev #else 13785084Sjohnlev 13795084Sjohnlev if (x86_feature & X86_ASYSC) { 13805084Sjohnlev wrmsr(MSR_AMD_LSTAR, (uintptr_t)sys_syscall); 13815084Sjohnlev wrmsr(MSR_AMD_CSTAR, (uintptr_t)sys_syscall32); 13825084Sjohnlev } 13835084Sjohnlev 13842712Snn35248 #endif 13855084Sjohnlev #endif /* __amd64 */ 13862712Snn35248 13872712Snn35248 if (x86_feature & X86_SEP) 13882712Snn35248 wrmsr(MSR_INTC_SEP_EIP, (uintptr_t)sys_sysenter); 13892712Snn35248 } 1390