10Sstevel@tonic-gate /*
23446Smrj * CDDL HEADER START
33446Smrj *
43446Smrj * The contents of this file are subject to the terms of the
53446Smrj * Common Development and Distribution License (the "License").
63446Smrj * You may not use this file except in compliance with the License.
73446Smrj *
83446Smrj * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
93446Smrj * or http://www.opensolaris.org/os/licensing.
103446Smrj * See the License for the specific language governing permissions
113446Smrj * and limitations under the License.
123446Smrj *
133446Smrj * When distributing Covered Code, include this CDDL HEADER in each
143446Smrj * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
153446Smrj * If applicable, add the following below this CDDL HEADER, with the
163446Smrj * fields enclosed by brackets "[]" replaced with your own identifying
173446Smrj * information: Portions Copyright [yyyy] [name of copyright owner]
183446Smrj *
193446Smrj * CDDL HEADER END
203446Smrj */
213446Smrj
223446Smrj /*
2312613SSurya.Prakki@Sun.COM * Copyright (c) 2004, 2010, Oracle and/or its affiliates. All rights reserved.
240Sstevel@tonic-gate */
250Sstevel@tonic-gate
260Sstevel@tonic-gate /*
270Sstevel@tonic-gate * Copyright (c) 1992 Terrence R. Lambert.
280Sstevel@tonic-gate * Copyright (c) 1990 The Regents of the University of California.
290Sstevel@tonic-gate * All rights reserved.
300Sstevel@tonic-gate *
310Sstevel@tonic-gate * This code is derived from software contributed to Berkeley by
320Sstevel@tonic-gate * William Jolitz.
330Sstevel@tonic-gate *
340Sstevel@tonic-gate * Redistribution and use in source and binary forms, with or without
350Sstevel@tonic-gate * modification, are permitted provided that the following conditions
360Sstevel@tonic-gate * are met:
370Sstevel@tonic-gate * 1. Redistributions of source code must retain the above copyright
380Sstevel@tonic-gate * notice, this list of conditions and the following disclaimer.
390Sstevel@tonic-gate * 2. Redistributions in binary form must reproduce the above copyright
400Sstevel@tonic-gate * notice, this list of conditions and the following disclaimer in the
410Sstevel@tonic-gate * documentation and/or other materials provided with the distribution.
420Sstevel@tonic-gate * 3. All advertising materials mentioning features or use of this software
430Sstevel@tonic-gate * must display the following acknowledgement:
440Sstevel@tonic-gate * This product includes software developed by the University of
450Sstevel@tonic-gate * California, Berkeley and its contributors.
460Sstevel@tonic-gate * 4. Neither the name of the University nor the names of its contributors
470Sstevel@tonic-gate * may be used to endorse or promote products derived from this software
480Sstevel@tonic-gate * without specific prior written permission.
490Sstevel@tonic-gate *
500Sstevel@tonic-gate * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
510Sstevel@tonic-gate * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
520Sstevel@tonic-gate * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
530Sstevel@tonic-gate * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
540Sstevel@tonic-gate * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
550Sstevel@tonic-gate * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
560Sstevel@tonic-gate * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
570Sstevel@tonic-gate * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
580Sstevel@tonic-gate * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
590Sstevel@tonic-gate * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
600Sstevel@tonic-gate * SUCH DAMAGE.
610Sstevel@tonic-gate *
620Sstevel@tonic-gate * from: @(#)machdep.c 7.4 (Berkeley) 6/3/91
630Sstevel@tonic-gate */
640Sstevel@tonic-gate
650Sstevel@tonic-gate #include <sys/types.h>
663446Smrj #include <sys/sysmacros.h>
670Sstevel@tonic-gate #include <sys/tss.h>
680Sstevel@tonic-gate #include <sys/segments.h>
690Sstevel@tonic-gate #include <sys/trap.h>
700Sstevel@tonic-gate #include <sys/cpuvar.h>
713446Smrj #include <sys/bootconf.h>
720Sstevel@tonic-gate #include <sys/x86_archext.h>
733446Smrj #include <sys/controlregs.h>
740Sstevel@tonic-gate #include <sys/archsystm.h>
750Sstevel@tonic-gate #include <sys/machsystm.h>
760Sstevel@tonic-gate #include <sys/kobj.h>
770Sstevel@tonic-gate #include <sys/cmn_err.h>
780Sstevel@tonic-gate #include <sys/reboot.h>
790Sstevel@tonic-gate #include <sys/kdi.h>
803446Smrj #include <sys/mach_mmu.h>
811217Srab #include <sys/systm.h>
825084Sjohnlev
835084Sjohnlev #ifdef __xpv
845084Sjohnlev #include <sys/hypervisor.h>
855084Sjohnlev #include <vm/as.h>
865084Sjohnlev #endif
875084Sjohnlev
883446Smrj #include <sys/promif.h>
893446Smrj #include <sys/bootinfo.h>
903446Smrj #include <vm/kboot_mmu.h>
915084Sjohnlev #include <vm/hat_pte.h>
920Sstevel@tonic-gate
930Sstevel@tonic-gate /*
940Sstevel@tonic-gate * cpu0 and default tables and structures.
950Sstevel@tonic-gate */
963446Smrj user_desc_t *gdt0;
975084Sjohnlev #if !defined(__xpv)
980Sstevel@tonic-gate desctbr_t gdt0_default_r;
995084Sjohnlev #endif
1000Sstevel@tonic-gate
1015460Sjosephb gate_desc_t *idt0; /* interrupt descriptor table */
1023446Smrj #if defined(__i386)
1030Sstevel@tonic-gate desctbr_t idt0_default_r; /* describes idt0 in IDTR format */
1043446Smrj #endif
1050Sstevel@tonic-gate
1065460Sjosephb struct tss *ktss0; /* kernel task state structure */
1070Sstevel@tonic-gate
1080Sstevel@tonic-gate #if defined(__i386)
1095460Sjosephb struct tss *dftss0; /* #DF double-fault exception */
1100Sstevel@tonic-gate #endif /* __i386 */
1110Sstevel@tonic-gate
1120Sstevel@tonic-gate user_desc_t zero_udesc; /* base zero user desc native procs */
1135084Sjohnlev user_desc_t null_udesc; /* null user descriptor */
1145084Sjohnlev system_desc_t null_sdesc; /* null system descriptor */
1150Sstevel@tonic-gate
1160Sstevel@tonic-gate #if defined(__amd64)
1170Sstevel@tonic-gate user_desc_t zero_u32desc; /* 32-bit compatibility procs */
1180Sstevel@tonic-gate #endif /* __amd64 */
1190Sstevel@tonic-gate
1205084Sjohnlev #if defined(__amd64)
1215084Sjohnlev user_desc_t ucs_on;
1225084Sjohnlev user_desc_t ucs_off;
1235084Sjohnlev user_desc_t ucs32_on;
1245084Sjohnlev user_desc_t ucs32_off;
1255084Sjohnlev #endif /* __amd64 */
1265084Sjohnlev
1270Sstevel@tonic-gate #pragma align 16(dblfault_stack0)
1280Sstevel@tonic-gate char dblfault_stack0[DEFAULTSTKSZ];
1290Sstevel@tonic-gate
1300Sstevel@tonic-gate extern void fast_null(void);
1310Sstevel@tonic-gate extern hrtime_t get_hrtime(void);
1320Sstevel@tonic-gate extern hrtime_t gethrvtime(void);
1330Sstevel@tonic-gate extern hrtime_t get_hrestime(void);
1340Sstevel@tonic-gate extern uint64_t getlgrp(void);
1350Sstevel@tonic-gate
1360Sstevel@tonic-gate void (*(fasttable[]))(void) = {
1370Sstevel@tonic-gate fast_null, /* T_FNULL routine */
1380Sstevel@tonic-gate fast_null, /* T_FGETFP routine (initially null) */
1390Sstevel@tonic-gate fast_null, /* T_FSETFP routine (initially null) */
1400Sstevel@tonic-gate (void (*)())get_hrtime, /* T_GETHRTIME */
1410Sstevel@tonic-gate (void (*)())gethrvtime, /* T_GETHRVTIME */
1420Sstevel@tonic-gate (void (*)())get_hrestime, /* T_GETHRESTIME */
1430Sstevel@tonic-gate (void (*)())getlgrp /* T_GETLGRP */
1440Sstevel@tonic-gate };
1450Sstevel@tonic-gate
1460Sstevel@tonic-gate /*
1472712Snn35248 * Structure containing pre-computed descriptors to allow us to temporarily
1482712Snn35248 * interpose on a standard handler.
1492712Snn35248 */
1502712Snn35248 struct interposing_handler {
1512712Snn35248 int ih_inum;
1522712Snn35248 gate_desc_t ih_interp_desc;
1532712Snn35248 gate_desc_t ih_default_desc;
1542712Snn35248 };
1552712Snn35248
1562712Snn35248 /*
1572712Snn35248 * The brand infrastructure interposes on two handlers, and we use one as a
1582712Snn35248 * NULL signpost.
1592712Snn35248 */
16012613SSurya.Prakki@Sun.COM static struct interposing_handler brand_tbl[2];
1612712Snn35248
1622712Snn35248 /*
1630Sstevel@tonic-gate * software prototypes for default local descriptor table
1640Sstevel@tonic-gate */
1650Sstevel@tonic-gate
1660Sstevel@tonic-gate /*
1670Sstevel@tonic-gate * Routines for loading segment descriptors in format the hardware
1680Sstevel@tonic-gate * can understand.
1690Sstevel@tonic-gate */
1700Sstevel@tonic-gate
1710Sstevel@tonic-gate #if defined(__amd64)
1720Sstevel@tonic-gate
1730Sstevel@tonic-gate /*
1740Sstevel@tonic-gate * In long mode we have the new L or long mode attribute bit
1750Sstevel@tonic-gate * for code segments. Only the conforming bit in type is used along
1760Sstevel@tonic-gate * with descriptor priority and present bits. Default operand size must
1770Sstevel@tonic-gate * be zero when in long mode. In 32-bit compatibility mode all fields
1780Sstevel@tonic-gate * are treated as in legacy mode. For data segments while in long mode
1790Sstevel@tonic-gate * only the present bit is loaded.
1800Sstevel@tonic-gate */
1810Sstevel@tonic-gate void
set_usegd(user_desc_t * dp,uint_t lmode,void * base,size_t size,uint_t type,uint_t dpl,uint_t gran,uint_t defopsz)1820Sstevel@tonic-gate set_usegd(user_desc_t *dp, uint_t lmode, void *base, size_t size,
1830Sstevel@tonic-gate uint_t type, uint_t dpl, uint_t gran, uint_t defopsz)
1840Sstevel@tonic-gate {
1850Sstevel@tonic-gate ASSERT(lmode == SDP_SHORT || lmode == SDP_LONG);
1860Sstevel@tonic-gate
1870Sstevel@tonic-gate /*
1880Sstevel@tonic-gate * 64-bit long mode.
1890Sstevel@tonic-gate */
1900Sstevel@tonic-gate if (lmode == SDP_LONG)
1910Sstevel@tonic-gate dp->usd_def32 = 0; /* 32-bit operands only */
1920Sstevel@tonic-gate else
1930Sstevel@tonic-gate /*
1940Sstevel@tonic-gate * 32-bit compatibility mode.
1950Sstevel@tonic-gate */
1960Sstevel@tonic-gate dp->usd_def32 = defopsz; /* 0 = 16, 1 = 32-bit ops */
1970Sstevel@tonic-gate
1980Sstevel@tonic-gate dp->usd_long = lmode; /* 64-bit mode */
1990Sstevel@tonic-gate dp->usd_type = type;
2000Sstevel@tonic-gate dp->usd_dpl = dpl;
2010Sstevel@tonic-gate dp->usd_p = 1;
2020Sstevel@tonic-gate dp->usd_gran = gran; /* 0 = bytes, 1 = pages */
2030Sstevel@tonic-gate
2040Sstevel@tonic-gate dp->usd_lobase = (uintptr_t)base;
2050Sstevel@tonic-gate dp->usd_midbase = (uintptr_t)base >> 16;
2060Sstevel@tonic-gate dp->usd_hibase = (uintptr_t)base >> (16 + 8);
2070Sstevel@tonic-gate dp->usd_lolimit = size;
2080Sstevel@tonic-gate dp->usd_hilimit = (uintptr_t)size >> 16;
2090Sstevel@tonic-gate }
2100Sstevel@tonic-gate
2110Sstevel@tonic-gate #elif defined(__i386)
2120Sstevel@tonic-gate
2130Sstevel@tonic-gate /*
2140Sstevel@tonic-gate * Install user segment descriptor for code and data.
2150Sstevel@tonic-gate */
2160Sstevel@tonic-gate void
set_usegd(user_desc_t * dp,void * base,size_t size,uint_t type,uint_t dpl,uint_t gran,uint_t defopsz)2170Sstevel@tonic-gate set_usegd(user_desc_t *dp, void *base, size_t size, uint_t type,
2180Sstevel@tonic-gate uint_t dpl, uint_t gran, uint_t defopsz)
2190Sstevel@tonic-gate {
2200Sstevel@tonic-gate dp->usd_lolimit = size;
2210Sstevel@tonic-gate dp->usd_hilimit = (uintptr_t)size >> 16;
2220Sstevel@tonic-gate
2230Sstevel@tonic-gate dp->usd_lobase = (uintptr_t)base;
2240Sstevel@tonic-gate dp->usd_midbase = (uintptr_t)base >> 16;
2250Sstevel@tonic-gate dp->usd_hibase = (uintptr_t)base >> (16 + 8);
2260Sstevel@tonic-gate
2270Sstevel@tonic-gate dp->usd_type = type;
2280Sstevel@tonic-gate dp->usd_dpl = dpl;
2290Sstevel@tonic-gate dp->usd_p = 1;
2300Sstevel@tonic-gate dp->usd_def32 = defopsz; /* 0 = 16, 1 = 32 bit operands */
2310Sstevel@tonic-gate dp->usd_gran = gran; /* 0 = bytes, 1 = pages */
2320Sstevel@tonic-gate }
2330Sstevel@tonic-gate
2340Sstevel@tonic-gate #endif /* __i386 */
2350Sstevel@tonic-gate
2360Sstevel@tonic-gate /*
2370Sstevel@tonic-gate * Install system segment descriptor for LDT and TSS segments.
2380Sstevel@tonic-gate */
2390Sstevel@tonic-gate
2400Sstevel@tonic-gate #if defined(__amd64)
2410Sstevel@tonic-gate
2420Sstevel@tonic-gate void
set_syssegd(system_desc_t * dp,void * base,size_t size,uint_t type,uint_t dpl)2430Sstevel@tonic-gate set_syssegd(system_desc_t *dp, void *base, size_t size, uint_t type,
2440Sstevel@tonic-gate uint_t dpl)
2450Sstevel@tonic-gate {
2460Sstevel@tonic-gate dp->ssd_lolimit = size;
2470Sstevel@tonic-gate dp->ssd_hilimit = (uintptr_t)size >> 16;
2480Sstevel@tonic-gate
2490Sstevel@tonic-gate dp->ssd_lobase = (uintptr_t)base;
2500Sstevel@tonic-gate dp->ssd_midbase = (uintptr_t)base >> 16;
2510Sstevel@tonic-gate dp->ssd_hibase = (uintptr_t)base >> (16 + 8);
2520Sstevel@tonic-gate dp->ssd_hi64base = (uintptr_t)base >> (16 + 8 + 8);
2530Sstevel@tonic-gate
2540Sstevel@tonic-gate dp->ssd_type = type;
2550Sstevel@tonic-gate dp->ssd_zero1 = 0; /* must be zero */
2560Sstevel@tonic-gate dp->ssd_zero2 = 0;
2570Sstevel@tonic-gate dp->ssd_dpl = dpl;
2580Sstevel@tonic-gate dp->ssd_p = 1;
2590Sstevel@tonic-gate dp->ssd_gran = 0; /* force byte units */
2600Sstevel@tonic-gate }
2610Sstevel@tonic-gate
2625084Sjohnlev void *
get_ssd_base(system_desc_t * dp)2635084Sjohnlev get_ssd_base(system_desc_t *dp)
2645084Sjohnlev {
2655084Sjohnlev uintptr_t base;
2665084Sjohnlev
2675084Sjohnlev base = (uintptr_t)dp->ssd_lobase |
2685084Sjohnlev (uintptr_t)dp->ssd_midbase << 16 |
2695084Sjohnlev (uintptr_t)dp->ssd_hibase << (16 + 8) |
2705084Sjohnlev (uintptr_t)dp->ssd_hi64base << (16 + 8 + 8);
2715084Sjohnlev return ((void *)base);
2725084Sjohnlev }
2735084Sjohnlev
2740Sstevel@tonic-gate #elif defined(__i386)
2750Sstevel@tonic-gate
2760Sstevel@tonic-gate void
set_syssegd(system_desc_t * dp,void * base,size_t size,uint_t type,uint_t dpl)2770Sstevel@tonic-gate set_syssegd(system_desc_t *dp, void *base, size_t size, uint_t type,
2780Sstevel@tonic-gate uint_t dpl)
2790Sstevel@tonic-gate {
2800Sstevel@tonic-gate dp->ssd_lolimit = size;
2810Sstevel@tonic-gate dp->ssd_hilimit = (uintptr_t)size >> 16;
2820Sstevel@tonic-gate
2830Sstevel@tonic-gate dp->ssd_lobase = (uintptr_t)base;
2840Sstevel@tonic-gate dp->ssd_midbase = (uintptr_t)base >> 16;
2850Sstevel@tonic-gate dp->ssd_hibase = (uintptr_t)base >> (16 + 8);
2860Sstevel@tonic-gate
2870Sstevel@tonic-gate dp->ssd_type = type;
2880Sstevel@tonic-gate dp->ssd_zero = 0; /* must be zero */
2890Sstevel@tonic-gate dp->ssd_dpl = dpl;
2900Sstevel@tonic-gate dp->ssd_p = 1;
2910Sstevel@tonic-gate dp->ssd_gran = 0; /* force byte units */
2920Sstevel@tonic-gate }
2930Sstevel@tonic-gate
2945084Sjohnlev void *
get_ssd_base(system_desc_t * dp)2955084Sjohnlev get_ssd_base(system_desc_t *dp)
2965084Sjohnlev {
2975084Sjohnlev uintptr_t base;
2985084Sjohnlev
2995084Sjohnlev base = (uintptr_t)dp->ssd_lobase |
3005084Sjohnlev (uintptr_t)dp->ssd_midbase << 16 |
3015084Sjohnlev (uintptr_t)dp->ssd_hibase << (16 + 8);
3025084Sjohnlev return ((void *)base);
3035084Sjohnlev }
3045084Sjohnlev
3050Sstevel@tonic-gate #endif /* __i386 */
3060Sstevel@tonic-gate
3070Sstevel@tonic-gate /*
3080Sstevel@tonic-gate * Install gate segment descriptor for interrupt, trap, call and task gates.
3090Sstevel@tonic-gate */
3100Sstevel@tonic-gate
3110Sstevel@tonic-gate #if defined(__amd64)
3120Sstevel@tonic-gate
3138679SSeth.Goldberg@Sun.COM /*ARGSUSED*/
3140Sstevel@tonic-gate void
set_gatesegd(gate_desc_t * dp,void (* func)(void),selector_t sel,uint_t type,uint_t dpl,uint_t vector)3153446Smrj set_gatesegd(gate_desc_t *dp, void (*func)(void), selector_t sel,
3168679SSeth.Goldberg@Sun.COM uint_t type, uint_t dpl, uint_t vector)
3170Sstevel@tonic-gate {
3180Sstevel@tonic-gate dp->sgd_looffset = (uintptr_t)func;
3190Sstevel@tonic-gate dp->sgd_hioffset = (uintptr_t)func >> 16;
3200Sstevel@tonic-gate dp->sgd_hi64offset = (uintptr_t)func >> (16 + 16);
3210Sstevel@tonic-gate
3220Sstevel@tonic-gate dp->sgd_selector = (uint16_t)sel;
3233446Smrj
3243446Smrj /*
3253446Smrj * For 64 bit native we use the IST stack mechanism
3263446Smrj * for double faults. All other traps use the CPL = 0
3273446Smrj * (tss_rsp0) stack.
3283446Smrj */
3295084Sjohnlev #if !defined(__xpv)
3308679SSeth.Goldberg@Sun.COM if (vector == T_DBLFLT)
3313446Smrj dp->sgd_ist = 1;
3323446Smrj else
3335084Sjohnlev #endif
3343446Smrj dp->sgd_ist = 0;
3353446Smrj
3360Sstevel@tonic-gate dp->sgd_type = type;
3370Sstevel@tonic-gate dp->sgd_dpl = dpl;
3380Sstevel@tonic-gate dp->sgd_p = 1;
3390Sstevel@tonic-gate }
3400Sstevel@tonic-gate
3410Sstevel@tonic-gate #elif defined(__i386)
3420Sstevel@tonic-gate
3438679SSeth.Goldberg@Sun.COM /*ARGSUSED*/
3440Sstevel@tonic-gate void
set_gatesegd(gate_desc_t * dp,void (* func)(void),selector_t sel,uint_t type,uint_t dpl,uint_t unused)3450Sstevel@tonic-gate set_gatesegd(gate_desc_t *dp, void (*func)(void), selector_t sel,
3468679SSeth.Goldberg@Sun.COM uint_t type, uint_t dpl, uint_t unused)
3470Sstevel@tonic-gate {
3480Sstevel@tonic-gate dp->sgd_looffset = (uintptr_t)func;
3490Sstevel@tonic-gate dp->sgd_hioffset = (uintptr_t)func >> 16;
3500Sstevel@tonic-gate
3510Sstevel@tonic-gate dp->sgd_selector = (uint16_t)sel;
3523446Smrj dp->sgd_stkcpy = 0; /* always zero bytes */
3530Sstevel@tonic-gate dp->sgd_type = type;
3540Sstevel@tonic-gate dp->sgd_dpl = dpl;
3550Sstevel@tonic-gate dp->sgd_p = 1;
3560Sstevel@tonic-gate }
3570Sstevel@tonic-gate
3583446Smrj #endif /* __i386 */
3593446Smrj
3605084Sjohnlev /*
3615084Sjohnlev * Updates a single user descriptor in the the GDT of the current cpu.
3625084Sjohnlev * Caller is responsible for preventing cpu migration.
3635084Sjohnlev */
3645084Sjohnlev
3655084Sjohnlev void
gdt_update_usegd(uint_t sidx,user_desc_t * udp)3665084Sjohnlev gdt_update_usegd(uint_t sidx, user_desc_t *udp)
3675084Sjohnlev {
3685084Sjohnlev #if defined(__xpv)
3695084Sjohnlev
3705084Sjohnlev uint64_t dpa = CPU->cpu_m.mcpu_gdtpa + sizeof (*udp) * sidx;
3715084Sjohnlev
3725084Sjohnlev if (HYPERVISOR_update_descriptor(pa_to_ma(dpa), *(uint64_t *)udp))
3735084Sjohnlev panic("gdt_update_usegd: HYPERVISOR_update_descriptor");
3745084Sjohnlev
3755084Sjohnlev #else /* __xpv */
3765084Sjohnlev
3775084Sjohnlev CPU->cpu_gdt[sidx] = *udp;
3785084Sjohnlev
3795084Sjohnlev #endif /* __xpv */
3805084Sjohnlev }
3815084Sjohnlev
3825084Sjohnlev /*
3835084Sjohnlev * Writes single descriptor pointed to by udp into a processes
3845084Sjohnlev * LDT entry pointed to by ldp.
3855084Sjohnlev */
3865084Sjohnlev int
ldt_update_segd(user_desc_t * ldp,user_desc_t * udp)3875084Sjohnlev ldt_update_segd(user_desc_t *ldp, user_desc_t *udp)
3885084Sjohnlev {
3895084Sjohnlev #if defined(__xpv)
3905084Sjohnlev
3915084Sjohnlev uint64_t dpa;
3925084Sjohnlev
3935084Sjohnlev dpa = mmu_ptob(hat_getpfnum(kas.a_hat, (caddr_t)ldp)) |
3945084Sjohnlev ((uintptr_t)ldp & PAGEOFFSET);
3955084Sjohnlev
3965084Sjohnlev /*
3975084Sjohnlev * The hypervisor is a little more restrictive about what it
3985084Sjohnlev * supports in the LDT.
3995084Sjohnlev */
4005084Sjohnlev if (HYPERVISOR_update_descriptor(pa_to_ma(dpa), *(uint64_t *)udp) != 0)
4015084Sjohnlev return (EINVAL);
4025084Sjohnlev
4035084Sjohnlev #else /* __xpv */
4045084Sjohnlev
4055084Sjohnlev *ldp = *udp;
4065084Sjohnlev
4075084Sjohnlev #endif /* __xpv */
4085084Sjohnlev return (0);
4095084Sjohnlev }
4105084Sjohnlev
4115084Sjohnlev #if defined(__xpv)
4125084Sjohnlev
4135084Sjohnlev /*
4145084Sjohnlev * Converts hw format gate descriptor into pseudo-IDT format for the hypervisor.
4155084Sjohnlev * Returns true if a valid entry was written.
4165084Sjohnlev */
4175084Sjohnlev int
xen_idt_to_trap_info(uint_t vec,gate_desc_t * sgd,void * ti_arg)4185084Sjohnlev xen_idt_to_trap_info(uint_t vec, gate_desc_t *sgd, void *ti_arg)
4195084Sjohnlev {
4205084Sjohnlev trap_info_t *ti = ti_arg; /* XXPV Aargh - segments.h comment */
4215084Sjohnlev
4225084Sjohnlev /*
4235084Sjohnlev * skip holes in the IDT
4245084Sjohnlev */
4255084Sjohnlev if (GATESEG_GETOFFSET(sgd) == 0)
4265084Sjohnlev return (0);
4275084Sjohnlev
4285084Sjohnlev ASSERT(sgd->sgd_type == SDT_SYSIGT);
4295084Sjohnlev ti->vector = vec;
4305084Sjohnlev TI_SET_DPL(ti, sgd->sgd_dpl);
4315084Sjohnlev
4325084Sjohnlev /*
4335084Sjohnlev * Is this an interrupt gate?
4345084Sjohnlev */
4355084Sjohnlev if (sgd->sgd_type == SDT_SYSIGT) {
4365084Sjohnlev /* LINTED */
4375084Sjohnlev TI_SET_IF(ti, 1);
4385084Sjohnlev }
4395084Sjohnlev ti->cs = sgd->sgd_selector;
4405084Sjohnlev #if defined(__amd64)
4415084Sjohnlev ti->cs |= SEL_KPL; /* force into ring 3. see KCS_SEL */
4425084Sjohnlev #endif
4435084Sjohnlev ti->address = GATESEG_GETOFFSET(sgd);
4445084Sjohnlev return (1);
4455084Sjohnlev }
4465084Sjohnlev
4475084Sjohnlev /*
4485084Sjohnlev * Convert a single hw format gate descriptor and write it into our virtual IDT.
4495084Sjohnlev */
4505084Sjohnlev void
xen_idt_write(gate_desc_t * sgd,uint_t vec)4515084Sjohnlev xen_idt_write(gate_desc_t *sgd, uint_t vec)
4525084Sjohnlev {
4535084Sjohnlev trap_info_t trapinfo[2];
4545084Sjohnlev
4555084Sjohnlev bzero(trapinfo, sizeof (trapinfo));
4565084Sjohnlev if (xen_idt_to_trap_info(vec, sgd, &trapinfo[0]) == 0)
4575084Sjohnlev return;
4585084Sjohnlev if (xen_set_trap_table(trapinfo) != 0)
4595084Sjohnlev panic("xen_idt_write: xen_set_trap_table() failed");
4605084Sjohnlev }
4615084Sjohnlev
4625084Sjohnlev #endif /* __xpv */
4635084Sjohnlev
4643446Smrj #if defined(__amd64)
4650Sstevel@tonic-gate
4660Sstevel@tonic-gate /*
4670Sstevel@tonic-gate * Build kernel GDT.
4680Sstevel@tonic-gate */
4690Sstevel@tonic-gate
4700Sstevel@tonic-gate static void
init_gdt_common(user_desc_t * gdt)4713446Smrj init_gdt_common(user_desc_t *gdt)
4720Sstevel@tonic-gate {
4733446Smrj int i;
4740Sstevel@tonic-gate
4750Sstevel@tonic-gate /*
4760Sstevel@tonic-gate * 64-bit kernel code segment.
4770Sstevel@tonic-gate */
4783446Smrj set_usegd(&gdt[GDT_KCODE], SDP_LONG, NULL, 0, SDT_MEMERA, SEL_KPL,
4790Sstevel@tonic-gate SDP_PAGES, SDP_OP32);
4800Sstevel@tonic-gate
4810Sstevel@tonic-gate /*
4820Sstevel@tonic-gate * 64-bit kernel data segment. The limit attribute is ignored in 64-bit
4830Sstevel@tonic-gate * mode, but we set it here to 0xFFFF so that we can use the SYSRET
4840Sstevel@tonic-gate * instruction to return from system calls back to 32-bit applications.
4850Sstevel@tonic-gate * SYSRET doesn't update the base, limit, or attributes of %ss or %ds
4860Sstevel@tonic-gate * descriptors. We therefore must ensure that the kernel uses something,
4870Sstevel@tonic-gate * though it will be ignored by hardware, that is compatible with 32-bit
4880Sstevel@tonic-gate * apps. For the same reason we must set the default op size of this
4890Sstevel@tonic-gate * descriptor to 32-bit operands.
4900Sstevel@tonic-gate */
4913446Smrj set_usegd(&gdt[GDT_KDATA], SDP_LONG, NULL, -1, SDT_MEMRWA,
4920Sstevel@tonic-gate SEL_KPL, SDP_PAGES, SDP_OP32);
4933446Smrj gdt[GDT_KDATA].usd_def32 = 1;
4940Sstevel@tonic-gate
4950Sstevel@tonic-gate /*
4960Sstevel@tonic-gate * 64-bit user code segment.
4970Sstevel@tonic-gate */
4983446Smrj set_usegd(&gdt[GDT_UCODE], SDP_LONG, NULL, 0, SDT_MEMERA, SEL_UPL,
4990Sstevel@tonic-gate SDP_PAGES, SDP_OP32);
5000Sstevel@tonic-gate
5010Sstevel@tonic-gate /*
5020Sstevel@tonic-gate * 32-bit user code segment.
5030Sstevel@tonic-gate */
5043446Smrj set_usegd(&gdt[GDT_U32CODE], SDP_SHORT, NULL, -1, SDT_MEMERA,
5050Sstevel@tonic-gate SEL_UPL, SDP_PAGES, SDP_OP32);
5060Sstevel@tonic-gate
5070Sstevel@tonic-gate /*
5085084Sjohnlev * See gdt_ucode32() and gdt_ucode_native().
5095084Sjohnlev */
5105084Sjohnlev ucs_on = ucs_off = gdt[GDT_UCODE];
5115084Sjohnlev ucs_off.usd_p = 0; /* forces #np fault */
5125084Sjohnlev
5135084Sjohnlev ucs32_on = ucs32_off = gdt[GDT_U32CODE];
5145084Sjohnlev ucs32_off.usd_p = 0; /* forces #np fault */
5155084Sjohnlev
5165084Sjohnlev /*
5170Sstevel@tonic-gate * 32 and 64 bit data segments can actually share the same descriptor.
5180Sstevel@tonic-gate * In long mode only the present bit is checked but all other fields
5190Sstevel@tonic-gate * are loaded. But in compatibility mode all fields are interpreted
5200Sstevel@tonic-gate * as in legacy mode so they must be set correctly for a 32-bit data
5210Sstevel@tonic-gate * segment.
5220Sstevel@tonic-gate */
5233446Smrj set_usegd(&gdt[GDT_UDATA], SDP_SHORT, NULL, -1, SDT_MEMRWA, SEL_UPL,
5240Sstevel@tonic-gate SDP_PAGES, SDP_OP32);
5250Sstevel@tonic-gate
5265084Sjohnlev #if !defined(__xpv)
5275084Sjohnlev
5280Sstevel@tonic-gate /*
5291217Srab * The 64-bit kernel has no default LDT. By default, the LDT descriptor
5301217Srab * in the GDT is 0.
5310Sstevel@tonic-gate */
5320Sstevel@tonic-gate
5330Sstevel@tonic-gate /*
5340Sstevel@tonic-gate * Kernel TSS
5350Sstevel@tonic-gate */
5365460Sjosephb set_syssegd((system_desc_t *)&gdt[GDT_KTSS], ktss0,
5375460Sjosephb sizeof (*ktss0) - 1, SDT_SYSTSS, SEL_KPL);
5380Sstevel@tonic-gate
5395084Sjohnlev #endif /* !__xpv */
5405084Sjohnlev
5410Sstevel@tonic-gate /*
5420Sstevel@tonic-gate * Initialize fs and gs descriptors for 32 bit processes.
5430Sstevel@tonic-gate * Only attributes and limits are initialized, the effective
5440Sstevel@tonic-gate * base address is programmed via fsbase/gsbase.
5450Sstevel@tonic-gate */
5463446Smrj set_usegd(&gdt[GDT_LWPFS], SDP_SHORT, NULL, -1, SDT_MEMRWA,
5470Sstevel@tonic-gate SEL_UPL, SDP_PAGES, SDP_OP32);
5483446Smrj set_usegd(&gdt[GDT_LWPGS], SDP_SHORT, NULL, -1, SDT_MEMRWA,
5490Sstevel@tonic-gate SEL_UPL, SDP_PAGES, SDP_OP32);
5500Sstevel@tonic-gate
5510Sstevel@tonic-gate /*
5522712Snn35248 * Initialize the descriptors set aside for brand usage.
5532712Snn35248 * Only attributes and limits are initialized.
5542712Snn35248 */
5552712Snn35248 for (i = GDT_BRANDMIN; i <= GDT_BRANDMAX; i++)
5563446Smrj set_usegd(&gdt0[i], SDP_SHORT, NULL, -1, SDT_MEMRWA,
5572712Snn35248 SEL_UPL, SDP_PAGES, SDP_OP32);
5582712Snn35248
5592712Snn35248 /*
5600Sstevel@tonic-gate * Initialize convenient zero base user descriptors for clearing
5610Sstevel@tonic-gate * lwp private %fs and %gs descriptors in GDT. See setregs() for
5620Sstevel@tonic-gate * an example.
5630Sstevel@tonic-gate */
5640Sstevel@tonic-gate set_usegd(&zero_udesc, SDP_LONG, 0, 0, SDT_MEMRWA, SEL_UPL,
5650Sstevel@tonic-gate SDP_BYTES, SDP_OP32);
5660Sstevel@tonic-gate set_usegd(&zero_u32desc, SDP_SHORT, 0, -1, SDT_MEMRWA, SEL_UPL,
5670Sstevel@tonic-gate SDP_PAGES, SDP_OP32);
5680Sstevel@tonic-gate }
5690Sstevel@tonic-gate
5705084Sjohnlev #if defined(__xpv)
5715084Sjohnlev
5725084Sjohnlev static user_desc_t *
init_gdt(void)5735084Sjohnlev init_gdt(void)
5745084Sjohnlev {
5755084Sjohnlev uint64_t gdtpa;
5765084Sjohnlev ulong_t ma[1]; /* XXPV should be a memory_t */
5775084Sjohnlev ulong_t addr;
5785084Sjohnlev
5795084Sjohnlev #if !defined(__lint)
5805084Sjohnlev /*
5815084Sjohnlev * Our gdt is never larger than a single page.
5825084Sjohnlev */
5835084Sjohnlev ASSERT((sizeof (*gdt0) * NGDT) <= PAGESIZE);
5845084Sjohnlev #endif
5855084Sjohnlev gdt0 = (user_desc_t *)BOP_ALLOC(bootops, (caddr_t)GDT_VA,
5865084Sjohnlev PAGESIZE, PAGESIZE);
5875084Sjohnlev bzero(gdt0, PAGESIZE);
5885084Sjohnlev
5895084Sjohnlev init_gdt_common(gdt0);
5905084Sjohnlev
5915084Sjohnlev /*
5925084Sjohnlev * XXX Since we never invoke kmdb until after the kernel takes
5935084Sjohnlev * over the descriptor tables why not have it use the kernel's
5945084Sjohnlev * selectors?
5955084Sjohnlev */
5965084Sjohnlev if (boothowto & RB_DEBUG) {
5975084Sjohnlev set_usegd(&gdt0[GDT_B32DATA], SDP_LONG, NULL, -1, SDT_MEMRWA,
5985084Sjohnlev SEL_KPL, SDP_PAGES, SDP_OP32);
5995084Sjohnlev set_usegd(&gdt0[GDT_B64CODE], SDP_LONG, NULL, -1, SDT_MEMERA,
6005084Sjohnlev SEL_KPL, SDP_PAGES, SDP_OP32);
6015084Sjohnlev }
6025084Sjohnlev
6035084Sjohnlev /*
6045084Sjohnlev * Clear write permission for page containing the gdt and install it.
6055084Sjohnlev */
6065084Sjohnlev gdtpa = pfn_to_pa(va_to_pfn(gdt0));
6075084Sjohnlev ma[0] = (ulong_t)(pa_to_ma(gdtpa) >> PAGESHIFT);
6085084Sjohnlev kbm_read_only((uintptr_t)gdt0, gdtpa);
6095084Sjohnlev xen_set_gdt(ma, NGDT);
6105084Sjohnlev
6115084Sjohnlev /*
6125084Sjohnlev * Reload the segment registers to use the new GDT.
6135084Sjohnlev * On 64-bit, fixup KCS_SEL to be in ring 3.
6145084Sjohnlev * See KCS_SEL in segments.h.
6155084Sjohnlev */
6165084Sjohnlev load_segment_registers((KCS_SEL | SEL_KPL), KFS_SEL, KGS_SEL, KDS_SEL);
6175084Sjohnlev
6185084Sjohnlev /*
6195084Sjohnlev * setup %gs for kernel
6205084Sjohnlev */
6215084Sjohnlev xen_set_segment_base(SEGBASE_GS_KERNEL, (ulong_t)&cpus[0]);
6225084Sjohnlev
6235084Sjohnlev /*
6245084Sjohnlev * XX64 We should never dereference off "other gsbase" or
6255084Sjohnlev * "fsbase". So, we should arrange to point FSBASE and
6265084Sjohnlev * KGSBASE somewhere truly awful e.g. point it at the last
6275084Sjohnlev * valid address below the hole so that any attempts to index
6285084Sjohnlev * off them cause an exception.
6295084Sjohnlev *
6305084Sjohnlev * For now, point it at 8G -- at least it should be unmapped
6315084Sjohnlev * until some 64-bit processes run.
6325084Sjohnlev */
6335084Sjohnlev addr = 0x200000000ul;
6345084Sjohnlev xen_set_segment_base(SEGBASE_FS, addr);
6355084Sjohnlev xen_set_segment_base(SEGBASE_GS_USER, addr);
6365084Sjohnlev xen_set_segment_base(SEGBASE_GS_USER_SEL, 0);
6375084Sjohnlev
6385084Sjohnlev return (gdt0);
6395084Sjohnlev }
6405084Sjohnlev
6415084Sjohnlev #else /* __xpv */
6425084Sjohnlev
6433446Smrj static user_desc_t *
init_gdt(void)6440Sstevel@tonic-gate init_gdt(void)
6450Sstevel@tonic-gate {
6460Sstevel@tonic-gate desctbr_t r_bgdt, r_gdt;
6470Sstevel@tonic-gate user_desc_t *bgdt;
6483446Smrj
6493446Smrj #if !defined(__lint)
6503446Smrj /*
6513446Smrj * Our gdt is never larger than a single page.
6523446Smrj */
6533446Smrj ASSERT((sizeof (*gdt0) * NGDT) <= PAGESIZE);
6543446Smrj #endif
6553446Smrj gdt0 = (user_desc_t *)BOP_ALLOC(bootops, (caddr_t)GDT_VA,
6563446Smrj PAGESIZE, PAGESIZE);
6573446Smrj bzero(gdt0, PAGESIZE);
6583446Smrj
6593446Smrj init_gdt_common(gdt0);
6600Sstevel@tonic-gate
6610Sstevel@tonic-gate /*
6623446Smrj * Copy in from boot's gdt to our gdt.
6633446Smrj * Entry 0 is the null descriptor by definition.
6640Sstevel@tonic-gate */
6650Sstevel@tonic-gate rd_gdtr(&r_bgdt);
6660Sstevel@tonic-gate bgdt = (user_desc_t *)r_bgdt.dtr_base;
6670Sstevel@tonic-gate if (bgdt == NULL)
6680Sstevel@tonic-gate panic("null boot gdt");
6690Sstevel@tonic-gate
6703446Smrj gdt0[GDT_B32DATA] = bgdt[GDT_B32DATA];
6713446Smrj gdt0[GDT_B32CODE] = bgdt[GDT_B32CODE];
6723446Smrj gdt0[GDT_B16CODE] = bgdt[GDT_B16CODE];
6733446Smrj gdt0[GDT_B16DATA] = bgdt[GDT_B16DATA];
6743446Smrj gdt0[GDT_B64CODE] = bgdt[GDT_B64CODE];
6753446Smrj
6763446Smrj /*
6773446Smrj * Install our new GDT
6783446Smrj */
6793446Smrj r_gdt.dtr_limit = (sizeof (*gdt0) * NGDT) - 1;
6803446Smrj r_gdt.dtr_base = (uintptr_t)gdt0;
6813446Smrj wr_gdtr(&r_gdt);
6823446Smrj
6833446Smrj /*
6843446Smrj * Reload the segment registers to use the new GDT
6853446Smrj */
6863446Smrj load_segment_registers(KCS_SEL, KFS_SEL, KGS_SEL, KDS_SEL);
6873446Smrj
6883446Smrj /*
6893446Smrj * setup %gs for kernel
6903446Smrj */
6913446Smrj wrmsr(MSR_AMD_GSBASE, (uint64_t)&cpus[0]);
6923446Smrj
6933446Smrj /*
6943446Smrj * XX64 We should never dereference off "other gsbase" or
6953446Smrj * "fsbase". So, we should arrange to point FSBASE and
6963446Smrj * KGSBASE somewhere truly awful e.g. point it at the last
6973446Smrj * valid address below the hole so that any attempts to index
6983446Smrj * off them cause an exception.
6993446Smrj *
7003446Smrj * For now, point it at 8G -- at least it should be unmapped
7013446Smrj * until some 64-bit processes run.
7023446Smrj */
7033446Smrj wrmsr(MSR_AMD_FSBASE, 0x200000000ul);
7043446Smrj wrmsr(MSR_AMD_KGSBASE, 0x200000000ul);
7053446Smrj return (gdt0);
7063446Smrj }
7073446Smrj
7085084Sjohnlev #endif /* __xpv */
7095084Sjohnlev
7103446Smrj #elif defined(__i386)
7113446Smrj
7123446Smrj static void
init_gdt_common(user_desc_t * gdt)7133446Smrj init_gdt_common(user_desc_t *gdt)
7143446Smrj {
7153446Smrj int i;
7160Sstevel@tonic-gate
7170Sstevel@tonic-gate /*
7180Sstevel@tonic-gate * Text and data for both kernel and user span entire 32 bit
7190Sstevel@tonic-gate * address space.
7200Sstevel@tonic-gate */
7210Sstevel@tonic-gate
7220Sstevel@tonic-gate /*
7230Sstevel@tonic-gate * kernel code segment.
7240Sstevel@tonic-gate */
7253446Smrj set_usegd(&gdt[GDT_KCODE], NULL, -1, SDT_MEMERA, SEL_KPL, SDP_PAGES,
7260Sstevel@tonic-gate SDP_OP32);
7270Sstevel@tonic-gate
7280Sstevel@tonic-gate /*
7290Sstevel@tonic-gate * kernel data segment.
7300Sstevel@tonic-gate */
7313446Smrj set_usegd(&gdt[GDT_KDATA], NULL, -1, SDT_MEMRWA, SEL_KPL, SDP_PAGES,
7320Sstevel@tonic-gate SDP_OP32);
7330Sstevel@tonic-gate
7340Sstevel@tonic-gate /*
7350Sstevel@tonic-gate * user code segment.
7360Sstevel@tonic-gate */
7373446Smrj set_usegd(&gdt[GDT_UCODE], NULL, -1, SDT_MEMERA, SEL_UPL, SDP_PAGES,
7380Sstevel@tonic-gate SDP_OP32);
7390Sstevel@tonic-gate
7400Sstevel@tonic-gate /*
7410Sstevel@tonic-gate * user data segment.
7420Sstevel@tonic-gate */
7433446Smrj set_usegd(&gdt[GDT_UDATA], NULL, -1, SDT_MEMRWA, SEL_UPL, SDP_PAGES,
7440Sstevel@tonic-gate SDP_OP32);
7450Sstevel@tonic-gate
7465084Sjohnlev #if !defined(__xpv)
7475084Sjohnlev
7480Sstevel@tonic-gate /*
7490Sstevel@tonic-gate * TSS for T_DBLFLT (double fault) handler
7500Sstevel@tonic-gate */
7515460Sjosephb set_syssegd((system_desc_t *)&gdt[GDT_DBFLT], dftss0,
7525460Sjosephb sizeof (*dftss0) - 1, SDT_SYSTSS, SEL_KPL);
7530Sstevel@tonic-gate
7540Sstevel@tonic-gate /*
7550Sstevel@tonic-gate * TSS for kernel
7560Sstevel@tonic-gate */
7575460Sjosephb set_syssegd((system_desc_t *)&gdt[GDT_KTSS], ktss0,
7585460Sjosephb sizeof (*ktss0) - 1, SDT_SYSTSS, SEL_KPL);
7590Sstevel@tonic-gate
7605084Sjohnlev #endif /* !__xpv */
7615084Sjohnlev
7620Sstevel@tonic-gate /*
7630Sstevel@tonic-gate * %gs selector for kernel
7640Sstevel@tonic-gate */
7653446Smrj set_usegd(&gdt[GDT_GS], &cpus[0], sizeof (struct cpu) -1, SDT_MEMRWA,
7660Sstevel@tonic-gate SEL_KPL, SDP_BYTES, SDP_OP32);
7670Sstevel@tonic-gate
7680Sstevel@tonic-gate /*
7690Sstevel@tonic-gate * Initialize lwp private descriptors.
7700Sstevel@tonic-gate * Only attributes and limits are initialized, the effective
7710Sstevel@tonic-gate * base address is programmed via fsbase/gsbase.
7720Sstevel@tonic-gate */
7733446Smrj set_usegd(&gdt[GDT_LWPFS], NULL, (size_t)-1, SDT_MEMRWA, SEL_UPL,
7740Sstevel@tonic-gate SDP_PAGES, SDP_OP32);
7753446Smrj set_usegd(&gdt[GDT_LWPGS], NULL, (size_t)-1, SDT_MEMRWA, SEL_UPL,
7760Sstevel@tonic-gate SDP_PAGES, SDP_OP32);
7770Sstevel@tonic-gate
7780Sstevel@tonic-gate /*
7792712Snn35248 * Initialize the descriptors set aside for brand usage.
7802712Snn35248 * Only attributes and limits are initialized.
7812712Snn35248 */
7822712Snn35248 for (i = GDT_BRANDMIN; i <= GDT_BRANDMAX; i++)
7832712Snn35248 set_usegd(&gdt0[i], NULL, (size_t)-1, SDT_MEMRWA, SEL_UPL,
7842712Snn35248 SDP_PAGES, SDP_OP32);
7853446Smrj /*
7863446Smrj * Initialize convenient zero base user descriptor for clearing
7873446Smrj * lwp private %fs and %gs descriptors in GDT. See setregs() for
7883446Smrj * an example.
7893446Smrj */
7903446Smrj set_usegd(&zero_udesc, NULL, -1, SDT_MEMRWA, SEL_UPL,
7913446Smrj SDP_BYTES, SDP_OP32);
7923446Smrj }
7933446Smrj
7945084Sjohnlev #if defined(__xpv)
7955084Sjohnlev
7965084Sjohnlev static user_desc_t *
init_gdt(void)7975084Sjohnlev init_gdt(void)
7985084Sjohnlev {
7995084Sjohnlev uint64_t gdtpa;
8005084Sjohnlev ulong_t ma[1]; /* XXPV should be a memory_t */
8015084Sjohnlev
8025084Sjohnlev #if !defined(__lint)
8035084Sjohnlev /*
8045084Sjohnlev * Our gdt is never larger than a single page.
8055084Sjohnlev */
8065084Sjohnlev ASSERT((sizeof (*gdt0) * NGDT) <= PAGESIZE);
8075084Sjohnlev #endif
8085084Sjohnlev gdt0 = (user_desc_t *)BOP_ALLOC(bootops, (caddr_t)GDT_VA,
8095084Sjohnlev PAGESIZE, PAGESIZE);
8105084Sjohnlev bzero(gdt0, PAGESIZE);
8115084Sjohnlev
8125084Sjohnlev init_gdt_common(gdt0);
8135084Sjohnlev gdtpa = pfn_to_pa(va_to_pfn(gdt0));
8145084Sjohnlev
8155084Sjohnlev /*
8165084Sjohnlev * XXX Since we never invoke kmdb until after the kernel takes
8175084Sjohnlev * over the descriptor tables why not have it use the kernel's
8185084Sjohnlev * selectors?
8195084Sjohnlev */
8205084Sjohnlev if (boothowto & RB_DEBUG) {
8215084Sjohnlev set_usegd(&gdt0[GDT_B32DATA], NULL, -1, SDT_MEMRWA, SEL_KPL,
8225084Sjohnlev SDP_PAGES, SDP_OP32);
8235084Sjohnlev set_usegd(&gdt0[GDT_B32CODE], NULL, -1, SDT_MEMERA, SEL_KPL,
8245084Sjohnlev SDP_PAGES, SDP_OP32);
8255084Sjohnlev }
8265084Sjohnlev
8275084Sjohnlev /*
8285084Sjohnlev * Clear write permission for page containing the gdt and install it.
8295084Sjohnlev */
8305084Sjohnlev ma[0] = (ulong_t)(pa_to_ma(gdtpa) >> PAGESHIFT);
8315084Sjohnlev kbm_read_only((uintptr_t)gdt0, gdtpa);
8325084Sjohnlev xen_set_gdt(ma, NGDT);
8335084Sjohnlev
8345084Sjohnlev /*
8355084Sjohnlev * Reload the segment registers to use the new GDT
8365084Sjohnlev */
8375084Sjohnlev load_segment_registers(
8385084Sjohnlev KCS_SEL, KDS_SEL, KDS_SEL, KFS_SEL, KGS_SEL, KDS_SEL);
8395084Sjohnlev
8405084Sjohnlev return (gdt0);
8415084Sjohnlev }
8425084Sjohnlev
8435084Sjohnlev #else /* __xpv */
8445084Sjohnlev
8453446Smrj static user_desc_t *
init_gdt(void)8463446Smrj init_gdt(void)
8473446Smrj {
8483446Smrj desctbr_t r_bgdt, r_gdt;
8493446Smrj user_desc_t *bgdt;
8503446Smrj
8513446Smrj #if !defined(__lint)
8523446Smrj /*
8533446Smrj * Our gdt is never larger than a single page.
8543446Smrj */
8553446Smrj ASSERT((sizeof (*gdt0) * NGDT) <= PAGESIZE);
8563446Smrj #endif
8573446Smrj /*
8583446Smrj * XXX this allocation belongs in our caller, not here.
8593446Smrj */
8603446Smrj gdt0 = (user_desc_t *)BOP_ALLOC(bootops, (caddr_t)GDT_VA,
8613446Smrj PAGESIZE, PAGESIZE);
8623446Smrj bzero(gdt0, PAGESIZE);
8633446Smrj
8643446Smrj init_gdt_common(gdt0);
8653446Smrj
8663446Smrj /*
8673446Smrj * Copy in from boot's gdt to our gdt entries.
8683446Smrj * Entry 0 is null descriptor by definition.
8693446Smrj */
8703446Smrj rd_gdtr(&r_bgdt);
8713446Smrj bgdt = (user_desc_t *)r_bgdt.dtr_base;
8723446Smrj if (bgdt == NULL)
8733446Smrj panic("null boot gdt");
8743446Smrj
8753446Smrj gdt0[GDT_B32DATA] = bgdt[GDT_B32DATA];
8763446Smrj gdt0[GDT_B32CODE] = bgdt[GDT_B32CODE];
8773446Smrj gdt0[GDT_B16CODE] = bgdt[GDT_B16CODE];
8783446Smrj gdt0[GDT_B16DATA] = bgdt[GDT_B16DATA];
8792712Snn35248
8802712Snn35248 /*
8810Sstevel@tonic-gate * Install our new GDT
8820Sstevel@tonic-gate */
8833446Smrj r_gdt.dtr_limit = (sizeof (*gdt0) * NGDT) - 1;
8840Sstevel@tonic-gate r_gdt.dtr_base = (uintptr_t)gdt0;
8850Sstevel@tonic-gate wr_gdtr(&r_gdt);
8860Sstevel@tonic-gate
8870Sstevel@tonic-gate /*
8883446Smrj * Reload the segment registers to use the new GDT
8890Sstevel@tonic-gate */
8903446Smrj load_segment_registers(
8913446Smrj KCS_SEL, KDS_SEL, KDS_SEL, KFS_SEL, KGS_SEL, KDS_SEL);
8923446Smrj
8933446Smrj return (gdt0);
8940Sstevel@tonic-gate }
8950Sstevel@tonic-gate
8965084Sjohnlev #endif /* __xpv */
8970Sstevel@tonic-gate #endif /* __i386 */
8980Sstevel@tonic-gate
8990Sstevel@tonic-gate /*
9000Sstevel@tonic-gate * Build kernel IDT.
9010Sstevel@tonic-gate *
9023446Smrj * Note that for amd64 we pretty much require every gate to be an interrupt
9033446Smrj * gate which blocks interrupts atomically on entry; that's because of our
9043446Smrj * dependency on using 'swapgs' every time we come into the kernel to find
9053446Smrj * the cpu structure. If we get interrupted just before doing that, %cs could
9063446Smrj * be in kernel mode (so that the trap prolog doesn't do a swapgs), but
9073446Smrj * %gsbase is really still pointing at something in userland. Bad things will
9083446Smrj * ensue. We also use interrupt gates for i386 as well even though this is not
9093446Smrj * required for some traps.
9100Sstevel@tonic-gate *
9110Sstevel@tonic-gate * Perhaps they should have invented a trap gate that does an atomic swapgs?
9120Sstevel@tonic-gate */
9130Sstevel@tonic-gate static void
init_idt_common(gate_desc_t * idt)9143446Smrj init_idt_common(gate_desc_t *idt)
9153446Smrj {
9168679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_ZERODIV], &div0trap, KCS_SEL, SDT_SYSIGT, TRP_KPL,
9178679SSeth.Goldberg@Sun.COM 0);
9188679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_SGLSTP], &dbgtrap, KCS_SEL, SDT_SYSIGT, TRP_KPL,
9198679SSeth.Goldberg@Sun.COM 0);
9208679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_NMIFLT], &nmiint, KCS_SEL, SDT_SYSIGT, TRP_KPL,
9218679SSeth.Goldberg@Sun.COM 0);
9228679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_BPTFLT], &brktrap, KCS_SEL, SDT_SYSIGT, TRP_UPL,
9238679SSeth.Goldberg@Sun.COM 0);
9248679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_OVFLW], &ovflotrap, KCS_SEL, SDT_SYSIGT, TRP_UPL,
9258679SSeth.Goldberg@Sun.COM 0);
9263446Smrj set_gatesegd(&idt[T_BOUNDFLT], &boundstrap, KCS_SEL, SDT_SYSIGT,
9278679SSeth.Goldberg@Sun.COM TRP_KPL, 0);
9288679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_ILLINST], &invoptrap, KCS_SEL, SDT_SYSIGT, TRP_KPL,
9298679SSeth.Goldberg@Sun.COM 0);
9308679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_NOEXTFLT], &ndptrap, KCS_SEL, SDT_SYSIGT, TRP_KPL,
9318679SSeth.Goldberg@Sun.COM 0);
9323446Smrj
9333446Smrj /*
9343446Smrj * double fault handler.
9355084Sjohnlev *
9365084Sjohnlev * Note that on the hypervisor a guest does not receive #df faults.
9375084Sjohnlev * Instead a failsafe event is injected into the guest if its selectors
9385084Sjohnlev * and/or stack is in a broken state. See xen_failsafe_callback.
9393446Smrj */
9405084Sjohnlev #if !defined(__xpv)
9413446Smrj #if defined(__amd64)
9425084Sjohnlev
9438679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_DBLFLT], &syserrtrap, KCS_SEL, SDT_SYSIGT, TRP_KPL,
9448679SSeth.Goldberg@Sun.COM T_DBLFLT);
9455084Sjohnlev
9463446Smrj #elif defined(__i386)
9475084Sjohnlev
9483446Smrj /*
9493446Smrj * task gate required.
9503446Smrj */
9518679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_DBLFLT], NULL, DFTSS_SEL, SDT_SYSTASKGT, TRP_KPL,
9528679SSeth.Goldberg@Sun.COM 0);
9533446Smrj
9543446Smrj #endif /* __i386 */
9555084Sjohnlev #endif /* !__xpv */
9563446Smrj
9573446Smrj /*
9583446Smrj * T_EXTOVRFLT coprocessor-segment-overrun not supported.
9593446Smrj */
9603446Smrj
9618679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_TSSFLT], &invtsstrap, KCS_SEL, SDT_SYSIGT, TRP_KPL,
9628679SSeth.Goldberg@Sun.COM 0);
9638679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_SEGFLT], &segnptrap, KCS_SEL, SDT_SYSIGT, TRP_KPL,
9648679SSeth.Goldberg@Sun.COM 0);
9658679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_STKFLT], &stktrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 0);
9668679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_GPFLT], &gptrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 0);
9678679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_PGFLT], &pftrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 0);
9688679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_EXTERRFLT], &ndperr, KCS_SEL, SDT_SYSIGT, TRP_KPL,
9698679SSeth.Goldberg@Sun.COM 0);
9703446Smrj set_gatesegd(&idt[T_ALIGNMENT], &achktrap, KCS_SEL, SDT_SYSIGT,
9718679SSeth.Goldberg@Sun.COM TRP_KPL, 0);
9728679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_MCE], &mcetrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 0);
9738679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_SIMDFPE], &xmtrap, KCS_SEL, SDT_SYSIGT, TRP_KPL, 0);
9743446Smrj
9753446Smrj /*
9763446Smrj * install fast trap handler at 210.
9773446Smrj */
9788679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[T_FASTTRAP], &fasttrap, KCS_SEL, SDT_SYSIGT, TRP_UPL,
9798679SSeth.Goldberg@Sun.COM 0);
9803446Smrj
9813446Smrj /*
9823446Smrj * System call handler.
9833446Smrj */
9843446Smrj #if defined(__amd64)
9853446Smrj set_gatesegd(&idt[T_SYSCALLINT], &sys_syscall_int, KCS_SEL, SDT_SYSIGT,
9868679SSeth.Goldberg@Sun.COM TRP_UPL, 0);
9873446Smrj
9883446Smrj #elif defined(__i386)
9893446Smrj set_gatesegd(&idt[T_SYSCALLINT], &sys_call, KCS_SEL, SDT_SYSIGT,
9908679SSeth.Goldberg@Sun.COM TRP_UPL, 0);
9913446Smrj #endif /* __i386 */
9923446Smrj
9933446Smrj /*
9943446Smrj * Install the DTrace interrupt handler for the pid provider.
9953446Smrj */
9963446Smrj set_gatesegd(&idt[T_DTRACE_RET], &dtrace_ret, KCS_SEL,
9978679SSeth.Goldberg@Sun.COM SDT_SYSIGT, TRP_UPL, 0);
9983446Smrj
9993446Smrj /*
100012613SSurya.Prakki@Sun.COM * Prepare interposing descriptor for the syscall handler
100112613SSurya.Prakki@Sun.COM * and cache copy of the default descriptor.
10023446Smrj */
100312613SSurya.Prakki@Sun.COM brand_tbl[0].ih_inum = T_SYSCALLINT;
100412613SSurya.Prakki@Sun.COM brand_tbl[0].ih_default_desc = idt0[T_SYSCALLINT];
10053446Smrj
10063446Smrj #if defined(__amd64)
100712613SSurya.Prakki@Sun.COM set_gatesegd(&(brand_tbl[0].ih_interp_desc), &brand_sys_syscall_int,
10088679SSeth.Goldberg@Sun.COM KCS_SEL, SDT_SYSIGT, TRP_UPL, 0);
10093446Smrj #elif defined(__i386)
101012613SSurya.Prakki@Sun.COM set_gatesegd(&(brand_tbl[0].ih_interp_desc), &brand_sys_call,
10118679SSeth.Goldberg@Sun.COM KCS_SEL, SDT_SYSIGT, TRP_UPL, 0);
10123446Smrj #endif /* __i386 */
10133446Smrj
101412613SSurya.Prakki@Sun.COM brand_tbl[1].ih_inum = 0;
10153446Smrj }
10163446Smrj
10175084Sjohnlev #if defined(__xpv)
10185084Sjohnlev
10195084Sjohnlev static void
init_idt(gate_desc_t * idt)10205084Sjohnlev init_idt(gate_desc_t *idt)
10215084Sjohnlev {
10225084Sjohnlev init_idt_common(idt);
10235084Sjohnlev }
10245084Sjohnlev
10255084Sjohnlev #else /* __xpv */
10265084Sjohnlev
10273446Smrj static void
init_idt(gate_desc_t * idt)10283446Smrj init_idt(gate_desc_t *idt)
10290Sstevel@tonic-gate {
10300Sstevel@tonic-gate char ivctname[80];
10310Sstevel@tonic-gate void (*ivctptr)(void);
10320Sstevel@tonic-gate int i;
10330Sstevel@tonic-gate
10340Sstevel@tonic-gate /*
10350Sstevel@tonic-gate * Initialize entire table with 'reserved' trap and then overwrite
10360Sstevel@tonic-gate * specific entries. T_EXTOVRFLT (9) is unsupported and reserved
10370Sstevel@tonic-gate * since it can only be generated on a 386 processor. 15 is also
10380Sstevel@tonic-gate * unsupported and reserved.
10390Sstevel@tonic-gate */
10400Sstevel@tonic-gate for (i = 0; i < NIDT; i++)
10418679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[i], &resvtrap, KCS_SEL, SDT_SYSIGT, TRP_KPL,
10428679SSeth.Goldberg@Sun.COM 0);
10430Sstevel@tonic-gate
10440Sstevel@tonic-gate /*
10450Sstevel@tonic-gate * 20-31 reserved
10460Sstevel@tonic-gate */
10470Sstevel@tonic-gate for (i = 20; i < 32; i++)
10488679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[i], &invaltrap, KCS_SEL, SDT_SYSIGT, TRP_KPL,
10498679SSeth.Goldberg@Sun.COM 0);
10500Sstevel@tonic-gate
10510Sstevel@tonic-gate /*
10520Sstevel@tonic-gate * interrupts 32 - 255
10530Sstevel@tonic-gate */
10540Sstevel@tonic-gate for (i = 32; i < 256; i++) {
10550Sstevel@tonic-gate (void) snprintf(ivctname, sizeof (ivctname), "ivct%d", i);
10560Sstevel@tonic-gate ivctptr = (void (*)(void))kobj_getsymvalue(ivctname, 0);
10570Sstevel@tonic-gate if (ivctptr == NULL)
10580Sstevel@tonic-gate panic("kobj_getsymvalue(%s) failed", ivctname);
10590Sstevel@tonic-gate
10608679SSeth.Goldberg@Sun.COM set_gatesegd(&idt[i], ivctptr, KCS_SEL, SDT_SYSIGT, TRP_KPL, 0);
10610Sstevel@tonic-gate }
10620Sstevel@tonic-gate
10630Sstevel@tonic-gate /*
10643446Smrj * Now install the common ones. Note that it will overlay some
10653446Smrj * entries installed above like T_SYSCALLINT, T_FASTTRAP etc.
10660Sstevel@tonic-gate */
10673446Smrj init_idt_common(idt);
10680Sstevel@tonic-gate }
10690Sstevel@tonic-gate
10705084Sjohnlev #endif /* __xpv */
10715084Sjohnlev
10720Sstevel@tonic-gate /*
10731217Srab * The kernel does not deal with LDTs unless a user explicitly creates
10741217Srab * one. Under normal circumstances, the LDTR contains 0. Any process attempting
10751217Srab * to reference the LDT will therefore cause a #gp. System calls made via the
10761217Srab * obsolete lcall mechanism are emulated by the #gp fault handler.
10770Sstevel@tonic-gate */
10780Sstevel@tonic-gate static void
init_ldt(void)10790Sstevel@tonic-gate init_ldt(void)
10800Sstevel@tonic-gate {
10815084Sjohnlev #if defined(__xpv)
10825084Sjohnlev xen_set_ldt(NULL, 0);
10835084Sjohnlev #else
10841217Srab wr_ldtr(0);
10855084Sjohnlev #endif
10860Sstevel@tonic-gate }
10870Sstevel@tonic-gate
10885084Sjohnlev #if !defined(__xpv)
10890Sstevel@tonic-gate #if defined(__amd64)
10900Sstevel@tonic-gate
10910Sstevel@tonic-gate static void
init_tss(void)10920Sstevel@tonic-gate init_tss(void)
10930Sstevel@tonic-gate {
10940Sstevel@tonic-gate /*
10950Sstevel@tonic-gate * tss_rsp0 is dynamically filled in by resume() on each context switch.
10960Sstevel@tonic-gate * All exceptions but #DF will run on the thread stack.
10970Sstevel@tonic-gate * Set up the double fault stack here.
10980Sstevel@tonic-gate */
10995460Sjosephb ktss0->tss_ist1 =
11000Sstevel@tonic-gate (uint64_t)&dblfault_stack0[sizeof (dblfault_stack0)];
11010Sstevel@tonic-gate
11020Sstevel@tonic-gate /*
11030Sstevel@tonic-gate * Set I/O bit map offset equal to size of TSS segment limit
11040Sstevel@tonic-gate * for no I/O permission map. This will force all user I/O
11050Sstevel@tonic-gate * instructions to generate #gp fault.
11060Sstevel@tonic-gate */
11075460Sjosephb ktss0->tss_bitmapbase = sizeof (*ktss0);
11080Sstevel@tonic-gate
11090Sstevel@tonic-gate /*
11100Sstevel@tonic-gate * Point %tr to descriptor for ktss0 in gdt.
11110Sstevel@tonic-gate */
11120Sstevel@tonic-gate wr_tsr(KTSS_SEL);
11130Sstevel@tonic-gate }
11140Sstevel@tonic-gate
11150Sstevel@tonic-gate #elif defined(__i386)
11160Sstevel@tonic-gate
11170Sstevel@tonic-gate static void
init_tss(void)11180Sstevel@tonic-gate init_tss(void)
11190Sstevel@tonic-gate {
11200Sstevel@tonic-gate /*
11215460Sjosephb * ktss0->tss_esp dynamically filled in by resume() on each
11220Sstevel@tonic-gate * context switch.
11230Sstevel@tonic-gate */
11245460Sjosephb ktss0->tss_ss0 = KDS_SEL;
11255460Sjosephb ktss0->tss_eip = (uint32_t)_start;
11265460Sjosephb ktss0->tss_ds = ktss0->tss_es = ktss0->tss_ss = KDS_SEL;
11275460Sjosephb ktss0->tss_cs = KCS_SEL;
11285460Sjosephb ktss0->tss_fs = KFS_SEL;
11295460Sjosephb ktss0->tss_gs = KGS_SEL;
11305460Sjosephb ktss0->tss_ldt = ULDT_SEL;
11310Sstevel@tonic-gate
11320Sstevel@tonic-gate /*
11330Sstevel@tonic-gate * Initialize double fault tss.
11340Sstevel@tonic-gate */
11355460Sjosephb dftss0->tss_esp0 = (uint32_t)&dblfault_stack0[sizeof (dblfault_stack0)];
11365460Sjosephb dftss0->tss_ss0 = KDS_SEL;
11370Sstevel@tonic-gate
11380Sstevel@tonic-gate /*
11390Sstevel@tonic-gate * tss_cr3 will get initialized in hat_kern_setup() once our page
11400Sstevel@tonic-gate * tables have been setup.
11410Sstevel@tonic-gate */
11425460Sjosephb dftss0->tss_eip = (uint32_t)syserrtrap;
11435460Sjosephb dftss0->tss_esp = (uint32_t)&dblfault_stack0[sizeof (dblfault_stack0)];
11445460Sjosephb dftss0->tss_cs = KCS_SEL;
11455460Sjosephb dftss0->tss_ds = KDS_SEL;
11465460Sjosephb dftss0->tss_es = KDS_SEL;
11475460Sjosephb dftss0->tss_ss = KDS_SEL;
11485460Sjosephb dftss0->tss_fs = KFS_SEL;
11495460Sjosephb dftss0->tss_gs = KGS_SEL;
11500Sstevel@tonic-gate
11510Sstevel@tonic-gate /*
11520Sstevel@tonic-gate * Set I/O bit map offset equal to size of TSS segment limit
11530Sstevel@tonic-gate * for no I/O permission map. This will force all user I/O
11540Sstevel@tonic-gate * instructions to generate #gp fault.
11550Sstevel@tonic-gate */
11565460Sjosephb ktss0->tss_bitmapbase = sizeof (*ktss0);
11570Sstevel@tonic-gate
11580Sstevel@tonic-gate /*
11590Sstevel@tonic-gate * Point %tr to descriptor for ktss0 in gdt.
11600Sstevel@tonic-gate */
11610Sstevel@tonic-gate wr_tsr(KTSS_SEL);
11620Sstevel@tonic-gate }
11630Sstevel@tonic-gate
11640Sstevel@tonic-gate #endif /* __i386 */
11655084Sjohnlev #endif /* !__xpv */
11665084Sjohnlev
11675084Sjohnlev #if defined(__xpv)
11685084Sjohnlev
11695084Sjohnlev void
init_desctbls(void)11705084Sjohnlev init_desctbls(void)
11715084Sjohnlev {
11725084Sjohnlev uint_t vec;
11735084Sjohnlev user_desc_t *gdt;
11745084Sjohnlev
11755084Sjohnlev /*
11765084Sjohnlev * Setup and install our GDT.
11775084Sjohnlev */
11785084Sjohnlev gdt = init_gdt();
11795084Sjohnlev
11805084Sjohnlev /*
11815084Sjohnlev * Store static pa of gdt to speed up pa_to_ma() translations
11825084Sjohnlev * on lwp context switches.
11835084Sjohnlev */
11845084Sjohnlev ASSERT(IS_P2ALIGNED((uintptr_t)gdt, PAGESIZE));
11855460Sjosephb CPU->cpu_gdt = gdt;
11865084Sjohnlev CPU->cpu_m.mcpu_gdtpa = pfn_to_pa(va_to_pfn(gdt));
11875084Sjohnlev
11885084Sjohnlev /*
11895084Sjohnlev * Setup and install our IDT.
11905084Sjohnlev */
11915460Sjosephb #if !defined(__lint)
11925460Sjosephb ASSERT(NIDT * sizeof (*idt0) <= PAGESIZE);
11935460Sjosephb #endif
11945460Sjosephb idt0 = (gate_desc_t *)BOP_ALLOC(bootops, (caddr_t)IDT_VA,
11955460Sjosephb PAGESIZE, PAGESIZE);
11968679SSeth.Goldberg@Sun.COM bzero(idt0, PAGESIZE);
11975460Sjosephb init_idt(idt0);
11985084Sjohnlev for (vec = 0; vec < NIDT; vec++)
11995084Sjohnlev xen_idt_write(&idt0[vec], vec);
12005084Sjohnlev
12015460Sjosephb CPU->cpu_idt = idt0;
12025084Sjohnlev
12035084Sjohnlev /*
12045084Sjohnlev * set default kernel stack
12055084Sjohnlev */
12065084Sjohnlev xen_stack_switch(KDS_SEL,
12075084Sjohnlev (ulong_t)&dblfault_stack0[sizeof (dblfault_stack0)]);
12085084Sjohnlev
12095084Sjohnlev xen_init_callbacks();
12105084Sjohnlev
12115084Sjohnlev init_ldt();
12125084Sjohnlev }
12135084Sjohnlev
12145084Sjohnlev #else /* __xpv */
12150Sstevel@tonic-gate
12160Sstevel@tonic-gate void
init_desctbls(void)12173446Smrj init_desctbls(void)
12180Sstevel@tonic-gate {
12193446Smrj user_desc_t *gdt;
12203446Smrj desctbr_t idtr;
12213446Smrj
12223446Smrj /*
12235460Sjosephb * Allocate IDT and TSS structures on unique pages for better
12245460Sjosephb * performance in virtual machines.
12255460Sjosephb */
12265460Sjosephb #if !defined(__lint)
12275460Sjosephb ASSERT(NIDT * sizeof (*idt0) <= PAGESIZE);
12285460Sjosephb #endif
12295460Sjosephb idt0 = (gate_desc_t *)BOP_ALLOC(bootops, (caddr_t)IDT_VA,
12305460Sjosephb PAGESIZE, PAGESIZE);
12318679SSeth.Goldberg@Sun.COM bzero(idt0, PAGESIZE);
12325460Sjosephb #if !defined(__lint)
12335460Sjosephb ASSERT(sizeof (*ktss0) <= PAGESIZE);
12345460Sjosephb #endif
12355460Sjosephb ktss0 = (struct tss *)BOP_ALLOC(bootops, (caddr_t)KTSS_VA,
12365460Sjosephb PAGESIZE, PAGESIZE);
12378679SSeth.Goldberg@Sun.COM bzero(ktss0, PAGESIZE);
12385460Sjosephb
12395460Sjosephb #if defined(__i386)
12405460Sjosephb #if !defined(__lint)
12415460Sjosephb ASSERT(sizeof (*dftss0) <= PAGESIZE);
12425460Sjosephb #endif
12435460Sjosephb dftss0 = (struct tss *)BOP_ALLOC(bootops, (caddr_t)DFTSS_VA,
12445460Sjosephb PAGESIZE, PAGESIZE);
12458679SSeth.Goldberg@Sun.COM bzero(dftss0, PAGESIZE);
12465460Sjosephb #endif
12475460Sjosephb
12485460Sjosephb /*
12493446Smrj * Setup and install our GDT.
12503446Smrj */
12513446Smrj gdt = init_gdt();
12523446Smrj ASSERT(IS_P2ALIGNED((uintptr_t)gdt, PAGESIZE));
12535460Sjosephb CPU->cpu_gdt = gdt;
12543446Smrj
12553446Smrj /*
12563446Smrj * Setup and install our IDT.
12573446Smrj */
12585460Sjosephb init_idt(idt0);
12593446Smrj
12603446Smrj idtr.dtr_base = (uintptr_t)idt0;
12615460Sjosephb idtr.dtr_limit = (NIDT * sizeof (*idt0)) - 1;
12623446Smrj wr_idtr(&idtr);
12635460Sjosephb CPU->cpu_idt = idt0;
12643446Smrj
12653446Smrj #if defined(__i386)
12663446Smrj /*
12673446Smrj * We maintain a description of idt0 in convenient IDTR format
12683446Smrj * for #pf's on some older pentium processors. See pentium_pftrap().
12693446Smrj */
12703446Smrj idt0_default_r = idtr;
12713446Smrj #endif /* __i386 */
12723446Smrj
12730Sstevel@tonic-gate init_tss();
12745460Sjosephb CPU->cpu_tss = ktss0;
12750Sstevel@tonic-gate init_ldt();
12760Sstevel@tonic-gate }
12772712Snn35248
12785084Sjohnlev #endif /* __xpv */
12795084Sjohnlev
12802712Snn35248 /*
12813446Smrj * In the early kernel, we need to set up a simple GDT to run on.
12825084Sjohnlev *
12835084Sjohnlev * XXPV Can dboot use this too? See dboot_gdt.s
12843446Smrj */
12853446Smrj void
init_boot_gdt(user_desc_t * bgdt)12863446Smrj init_boot_gdt(user_desc_t *bgdt)
12873446Smrj {
12883446Smrj #if defined(__amd64)
12893446Smrj set_usegd(&bgdt[GDT_B32DATA], SDP_LONG, NULL, -1, SDT_MEMRWA, SEL_KPL,
12903446Smrj SDP_PAGES, SDP_OP32);
12913446Smrj set_usegd(&bgdt[GDT_B64CODE], SDP_LONG, NULL, -1, SDT_MEMERA, SEL_KPL,
12923446Smrj SDP_PAGES, SDP_OP32);
12933446Smrj #elif defined(__i386)
12943446Smrj set_usegd(&bgdt[GDT_B32DATA], NULL, -1, SDT_MEMRWA, SEL_KPL,
12953446Smrj SDP_PAGES, SDP_OP32);
12963446Smrj set_usegd(&bgdt[GDT_B32CODE], NULL, -1, SDT_MEMERA, SEL_KPL,
12973446Smrj SDP_PAGES, SDP_OP32);
12983446Smrj #endif /* __i386 */
12993446Smrj }
13003446Smrj
13013446Smrj /*
13022712Snn35248 * Enable interpositioning on the system call path by rewriting the
13032712Snn35248 * sys{call|enter} MSRs and the syscall-related entries in the IDT to use
13042712Snn35248 * the branded entry points.
13052712Snn35248 */
13062712Snn35248 void
brand_interpositioning_enable(void)13072712Snn35248 brand_interpositioning_enable(void)
13082712Snn35248 {
13095084Sjohnlev gate_desc_t *idt = CPU->cpu_idt;
13105084Sjohnlev int i;
13115084Sjohnlev
13125084Sjohnlev ASSERT(curthread->t_preempt != 0 || getpil() >= DISP_LEVEL);
13132712Snn35248
13145084Sjohnlev for (i = 0; brand_tbl[i].ih_inum; i++) {
13155084Sjohnlev idt[brand_tbl[i].ih_inum] = brand_tbl[i].ih_interp_desc;
13165084Sjohnlev #if defined(__xpv)
13175084Sjohnlev xen_idt_write(&idt[brand_tbl[i].ih_inum],
13185084Sjohnlev brand_tbl[i].ih_inum);
13195084Sjohnlev #endif
13205084Sjohnlev }
13212712Snn35248
13222712Snn35248 #if defined(__amd64)
13235084Sjohnlev #if defined(__xpv)
13245084Sjohnlev
13255084Sjohnlev /*
13265084Sjohnlev * Currently the hypervisor only supports 64-bit syscalls via
13275084Sjohnlev * syscall instruction. The 32-bit syscalls are handled by
13285084Sjohnlev * interrupt gate above.
13295084Sjohnlev */
13305084Sjohnlev xen_set_callback(brand_sys_syscall, CALLBACKTYPE_syscall,
13315084Sjohnlev CALLBACKF_mask_events);
13325084Sjohnlev
13335084Sjohnlev #else
13345084Sjohnlev
1335*12826Skuriakose.kuruvilla@oracle.com if (is_x86_feature(x86_featureset, X86FSET_ASYSC)) {
13365084Sjohnlev wrmsr(MSR_AMD_LSTAR, (uintptr_t)brand_sys_syscall);
13375084Sjohnlev wrmsr(MSR_AMD_CSTAR, (uintptr_t)brand_sys_syscall32);
13385084Sjohnlev }
13395084Sjohnlev
13402712Snn35248 #endif
13415084Sjohnlev #endif /* __amd64 */
13422712Snn35248
1343*12826Skuriakose.kuruvilla@oracle.com if (is_x86_feature(x86_featureset, X86FSET_SEP))
13442712Snn35248 wrmsr(MSR_INTC_SEP_EIP, (uintptr_t)brand_sys_sysenter);
13452712Snn35248 }
13462712Snn35248
13472712Snn35248 /*
13482712Snn35248 * Disable interpositioning on the system call path by rewriting the
13492712Snn35248 * sys{call|enter} MSRs and the syscall-related entries in the IDT to use
13502712Snn35248 * the standard entry points, which bypass the interpositioning hooks.
13512712Snn35248 */
13522712Snn35248 void
brand_interpositioning_disable(void)13532712Snn35248 brand_interpositioning_disable(void)
13542712Snn35248 {
13555084Sjohnlev gate_desc_t *idt = CPU->cpu_idt;
13562712Snn35248 int i;
13572712Snn35248
13585084Sjohnlev ASSERT(curthread->t_preempt != 0 || getpil() >= DISP_LEVEL);
13595084Sjohnlev
13605084Sjohnlev for (i = 0; brand_tbl[i].ih_inum; i++) {
13615084Sjohnlev idt[brand_tbl[i].ih_inum] = brand_tbl[i].ih_default_desc;
13625084Sjohnlev #if defined(__xpv)
13635084Sjohnlev xen_idt_write(&idt[brand_tbl[i].ih_inum],
13645084Sjohnlev brand_tbl[i].ih_inum);
13655084Sjohnlev #endif
13665084Sjohnlev }
13672712Snn35248
13682712Snn35248 #if defined(__amd64)
13695084Sjohnlev #if defined(__xpv)
13705084Sjohnlev
13715084Sjohnlev /*
13725084Sjohnlev * See comment above in brand_interpositioning_enable.
13735084Sjohnlev */
13745084Sjohnlev xen_set_callback(sys_syscall, CALLBACKTYPE_syscall,
13755084Sjohnlev CALLBACKF_mask_events);
13765084Sjohnlev
13775084Sjohnlev #else
13785084Sjohnlev
1379*12826Skuriakose.kuruvilla@oracle.com if (is_x86_feature(x86_featureset, X86FSET_ASYSC)) {
13805084Sjohnlev wrmsr(MSR_AMD_LSTAR, (uintptr_t)sys_syscall);
13815084Sjohnlev wrmsr(MSR_AMD_CSTAR, (uintptr_t)sys_syscall32);
13825084Sjohnlev }
13835084Sjohnlev
13842712Snn35248 #endif
13855084Sjohnlev #endif /* __amd64 */
13862712Snn35248
1387*12826Skuriakose.kuruvilla@oracle.com if (is_x86_feature(x86_featureset, X86FSET_SEP))
13882712Snn35248 wrmsr(MSR_INTC_SEP_EIP, (uintptr_t)sys_sysenter);
13892712Snn35248 }
1390