1 /* $NetBSD: psl.h,v 1.49 2016/05/18 08:16:04 nakayama Exp $ */ 2 3 /* 4 * Copyright (c) 1992, 1993 5 * The Regents of the University of California. All rights reserved. 6 * 7 * This software was developed by the Computer Systems Engineering group 8 * at Lawrence Berkeley Laboratory under DARPA contract BG 91-66 and 9 * contributed to Berkeley. 10 * 11 * All advertising materials mentioning features or use of this software 12 * must display the following acknowledgement: 13 * This product includes software developed by the University of 14 * California, Lawrence Berkeley Laboratory. 15 * 16 * Redistribution and use in source and binary forms, with or without 17 * modification, are permitted provided that the following conditions 18 * are met: 19 * 1. Redistributions of source code must retain the above copyright 20 * notice, this list of conditions and the following disclaimer. 21 * 2. Redistributions in binary form must reproduce the above copyright 22 * notice, this list of conditions and the following disclaimer in the 23 * documentation and/or other materials provided with the distribution. 24 * 3. Neither the name of the University nor the names of its contributors 25 * may be used to endorse or promote products derived from this software 26 * without specific prior written permission. 27 * 28 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND 29 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 30 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 31 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE 32 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 33 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 34 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 35 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 36 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 37 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 38 * SUCH DAMAGE. 39 * 40 * @(#)psl.h 8.1 (Berkeley) 6/11/93 41 */ 42 43 #ifndef PSR_IMPL 44 45 /* 46 * SPARC Process Status Register (in psl.h for hysterical raisins). This 47 * doesn't exist on the V9. 48 * 49 * The picture in the Sun manuals looks like this: 50 * 1 1 51 * 31 28 27 24 23 20 19 14 3 2 11 8 7 6 5 4 0 52 * +-------+-------+-------+-----------+-+-+-------+-+-+-+---------+ 53 * | impl | ver | icc | reserved |E|E| pil |S|P|E| CWP | 54 * | | |n z v c| |C|F| | |S|T| | 55 * +-------+-------+-------+-----------+-+-+-------+-+-+-+---------+ 56 */ 57 58 #define PSR_IMPL 0xf0000000 /* implementation */ 59 #define PSR_VER 0x0f000000 /* version */ 60 #define PSR_ICC 0x00f00000 /* integer condition codes */ 61 #define PSR_N 0x00800000 /* negative */ 62 #define PSR_Z 0x00400000 /* zero */ 63 #define PSR_O 0x00200000 /* overflow */ 64 #define PSR_C 0x00100000 /* carry */ 65 #define PSR_EC 0x00002000 /* coprocessor enable */ 66 #define PSR_EF 0x00001000 /* FP enable */ 67 #define PSR_PIL 0x00000f00 /* interrupt level */ 68 #define PSR_S 0x00000080 /* supervisor (kernel) mode */ 69 #define PSR_PS 0x00000040 /* previous supervisor mode (traps) */ 70 #define PSR_ET 0x00000020 /* trap enable */ 71 #define PSR_CWP 0x0000001f /* current window pointer */ 72 73 #define PSR_BITS "\20\16EC\15EF\10S\7PS\6ET" 74 75 /* 76 * SPARC V9 CCR register 77 */ 78 79 #define ICC_C 0x01L 80 #define ICC_V 0x02L 81 #define ICC_Z 0x04L 82 #define ICC_N 0x08L 83 #define XCC_SHIFT 4 84 #define XCC_C (ICC_C<<XCC_SHIFT) 85 #define XCC_V (ICC_V<<XCC_SHIFT) 86 #define XCC_Z (ICC_Z<<XCC_SHIFT) 87 #define XCC_N (ICC_N<<XCC_SHIFT) 88 89 90 /* 91 * SPARC V9 PSTATE register (what replaces the PSR in V9) 92 * 93 * Here's the layout: 94 * 95 * 11 10 9 8 7 6 5 4 3 2 1 0 96 * +------------------------------------------------------------+ 97 * | IG | MG | CLE | TLE | MM | RED | PEF | AM | PRIV | IE | AG | 98 * +------------------------------------------------------------+ 99 */ 100 101 #define PSTATE_IG 0x800 /* enable spitfire interrupt globals */ 102 #define PSTATE_MG 0x400 /* enable spitfire MMU globals */ 103 #define PSTATE_CLE 0x200 /* current little endian */ 104 #define PSTATE_TLE 0x100 /* traps little endian */ 105 #define PSTATE_MM 0x0c0 /* memory model */ 106 #define PSTATE_MM_TSO 0x000 /* total store order */ 107 #define PSTATE_MM_PSO 0x040 /* partial store order */ 108 #define PSTATE_MM_RMO 0x080 /* Relaxed memory order */ 109 #define PSTATE_RED 0x020 /* RED state */ 110 #define PSTATE_PEF 0x010 /* enable floating point */ 111 #define PSTATE_AM 0x008 /* 32-bit address masking */ 112 #define PSTATE_PRIV 0x004 /* privileged mode */ 113 #define PSTATE_IE 0x002 /* interrupt enable */ 114 #define PSTATE_AG 0x001 /* enable alternate globals */ 115 116 #define PSTATE_BITS "\20\14IG\13MG\12CLE\11TLE\10\7MM\6RED\5PEF\4AM\3PRIV\2IE\1AG" 117 118 119 /* 120 * 32-bit code requires TSO or at best PSO since that's what's supported on 121 * SPARC V8 and earlier machines. 122 * 123 * 64-bit code sets the memory model in the ELF header. 124 * 125 * We're running kernel code in TSO for the moment so we don't need to worry 126 * about possible memory barrier bugs. 127 */ 128 129 #ifdef __arch64__ 130 #define PSTATE_PROM (PSTATE_MM_TSO|PSTATE_PRIV) 131 #define PSTATE_NUCLEUS (PSTATE_MM_TSO|PSTATE_PRIV|PSTATE_AG) 132 #define PSTATE_KERN (PSTATE_MM_TSO|PSTATE_PRIV) 133 #define PSTATE_INTR (PSTATE_KERN|PSTATE_IE) 134 #define PSTATE_USER32 (PSTATE_MM_TSO|PSTATE_AM|PSTATE_IE) 135 #define PSTATE_USER (PSTATE_MM_RMO|PSTATE_IE) 136 #else 137 #define PSTATE_PROM (PSTATE_MM_TSO|PSTATE_PRIV) 138 #define PSTATE_NUCLEUS (PSTATE_MM_TSO|PSTATE_AM|PSTATE_PRIV|PSTATE_AG) 139 #define PSTATE_KERN (PSTATE_MM_TSO|PSTATE_AM|PSTATE_PRIV) 140 #define PSTATE_INTR (PSTATE_KERN|PSTATE_IE) 141 #define PSTATE_USER32 (PSTATE_MM_TSO|PSTATE_AM|PSTATE_IE) 142 #define PSTATE_USER (PSTATE_MM_TSO|PSTATE_AM|PSTATE_IE) 143 #endif 144 145 146 /* 147 * SPARC V9 TSTATE register 148 * 149 * 39 32 31 24 23 20 19 8 7 5 4 0 150 * +-----+-----+-----+--------+---+-----+ 151 * | CCR | ASI | - | PSTATE | - | CWP | 152 * +-----+-----+-----+--------+---+-----+ 153 */ 154 155 #define TSTATE_CWP 0x01f 156 #define TSTATE_PSTATE 0xfff00 157 #define TSTATE_PSTATE_SHIFT 8 158 #define TSTATE_ASI 0xff000000LL 159 #define TSTATE_ASI_SHIFT 24 160 #define TSTATE_CCR 0xff00000000LL 161 #define TSTATE_CCR_SHIFT 32 162 163 #define PSRCC_TO_TSTATE(x) (((int64_t)(x)&PSR_ICC)<<(TSTATE_CCR_SHIFT-20)) 164 #define TSTATECCR_TO_PSR(x) (((x)&TSTATE_CCR)>>(TSTATE_CCR_SHIFT-20)) 165 166 /* 167 * These are here to simplify life. 168 */ 169 #define TSTATE_IG (PSTATE_IG<<TSTATE_PSTATE_SHIFT) 170 #define TSTATE_MG (PSTATE_MG<<TSTATE_PSTATE_SHIFT) 171 #define TSTATE_CLE (PSTATE_CLE<<TSTATE_PSTATE_SHIFT) 172 #define TSTATE_TLE (PSTATE_TLE<<TSTATE_PSTATE_SHIFT) 173 #define TSTATE_MM (PSTATE_MM<<TSTATE_PSTATE_SHIFT) 174 #define TSTATE_MM_TSO (PSTATE_MM_TSO<<TSTATE_PSTATE_SHIFT) 175 #define TSTATE_MM_PSO (PSTATE_MM_PSO<<TSTATE_PSTATE_SHIFT) 176 #define TSTATE_MM_RMO (PSTATE_MM_RMO<<TSTATE_PSTATE_SHIFT) 177 #define TSTATE_RED (PSTATE_RED<<TSTATE_PSTATE_SHIFT) 178 #define TSTATE_PEF (PSTATE_PEF<<TSTATE_PSTATE_SHIFT) 179 #define TSTATE_AM (PSTATE_AM<<TSTATE_PSTATE_SHIFT) 180 #define TSTATE_PRIV (PSTATE_PRIV<<TSTATE_PSTATE_SHIFT) 181 #define TSTATE_IE (PSTATE_IE<<TSTATE_PSTATE_SHIFT) 182 #define TSTATE_AG (PSTATE_AG<<TSTATE_PSTATE_SHIFT) 183 184 #define TSTATE_BITS "\20\14IG\13MG\12CLE\11TLE\10\7MM\6RED\5PEF\4AM\3PRIV\2IE\1AG" 185 186 #define TSTATE_KERN ((PSTATE_KERN)<<TSTATE_PSTATE_SHIFT) 187 #define TSTATE_USER ((PSTATE_USER)<<TSTATE_PSTATE_SHIFT) 188 /* 189 * SPARC V9 VER version register. 190 * 191 * 63 48 47 32 31 24 23 16 15 8 7 5 4 0 192 * +-------+------+------+-----+-------+---+--------+ 193 * | manuf | impl | mask | - | maxtl | - | maxwin | 194 * +-------+------+------+-----+-------+---+--------+ 195 * 196 */ 197 198 #define VER_MANUF 0xffff000000000000LL 199 #define VER_MANUF_SHIFT 48 200 #define VER_IMPL 0x0000ffff00000000LL 201 #define VER_IMPL_SHIFT 32 202 #define VER_MASK 0x00000000ff000000LL 203 #define VER_MASK_SHIFT 24 204 #define VER_MAXTL 0x000000000000ff00LL 205 #define VER_MAXTL_SHIFT 8 206 #define VER_MAXWIN 0x000000000000001fLL 207 208 /* 209 * Here are a few things to help us transition between user and kernel mode: 210 */ 211 212 /* Memory models */ 213 #define KERN_MM PSTATE_MM_TSO 214 #define USER_MM PSTATE_MM_RMO 215 216 /* 217 * Register window handlers. These point to generic routines that check the 218 * stack pointer and then vector to the real handler. We could optimize this 219 * if we could guarantee only 32-bit or 64-bit stacks. 220 */ 221 #define WSTATE_KERN 026 222 #define WSTATE_USER 022 223 224 #define CWP 0x01f 225 226 /* 64-byte alignment -- this seems the best place to put this. */ 227 #define SPARC64_BLOCK_SIZE 64 228 #define SPARC64_BLOCK_ALIGN 0x3f 229 230 #if defined(_KERNEL) && !defined(_LOCORE) 231 232 /* 233 * GCC pseudo-functions for manipulating PSR (primarily PIL field). 234 */ 235 static __inline __attribute__((__always_inline__)) int 236 getpsr(void) 237 { 238 int psr; 239 240 __asm volatile("rd %%psr,%0" : "=r" (psr)); 241 return (psr); 242 } 243 244 static __inline __attribute__((__always_inline__)) int 245 getmid(void) 246 { 247 int mid; 248 249 __asm volatile("rd %%tbr,%0" : "=r" (mid)); 250 return ((mid >> 20) & 0x3); 251 } 252 253 static __inline __attribute__((__always_inline__)) void 254 setpsr(int newpsr) 255 { 256 __asm volatile("wr %0,0,%%psr" : : "r" (newpsr) : "memory"); 257 __asm volatile("nop; nop; nop"); 258 } 259 260 static __inline __attribute__((__always_inline__)) void 261 spl0(void) 262 { 263 int psr, oldipl; 264 265 /* 266 * wrpsr xors two values: we choose old psr and old ipl here, 267 * which gives us the same value as the old psr but with all 268 * the old PIL bits turned off. 269 */ 270 __asm volatile("rd %%psr,%0" : "=r" (psr) : : "memory"); 271 oldipl = psr & PSR_PIL; 272 __asm volatile("wr %0,%1,%%psr" : : "r" (psr), "r" (oldipl)); 273 274 /* 275 * Three instructions must execute before we can depend 276 * on the bits to be changed. 277 */ 278 __asm volatile("nop; nop; nop"); 279 } 280 281 /* 282 * PIL 1 through 14 can use this macro. 283 * (spl0 and splhigh are special since they put all 0s or all 1s 284 * into the ipl field.) 285 */ 286 #define _SPLSET(name, newipl) \ 287 static __inline __attribute__((__always_inline__)) void name(void) \ 288 { \ 289 int psr; \ 290 __asm volatile("rd %%psr,%0" : "=r" (psr)); \ 291 psr &= ~PSR_PIL; \ 292 __asm volatile("wr %0,%1,%%psr" : : \ 293 "r" (psr), "n" ((newipl) << 8)); \ 294 __asm volatile("nop; nop; nop" : : : "memory"); \ 295 } 296 297 _SPLSET(spllowerschedclock, IPL_SCHED) 298 299 typedef uint8_t ipl_t; 300 typedef struct { 301 ipl_t _ipl; 302 } ipl_cookie_t; 303 304 static inline ipl_cookie_t 305 makeiplcookie(ipl_t ipl) 306 { 307 308 return (ipl_cookie_t){._ipl = ipl}; 309 } 310 311 /* Raise IPL and return previous value */ 312 static __inline int 313 splraiseipl(ipl_cookie_t icookie) 314 { 315 int newipl = icookie._ipl; 316 int psr, oldipl; 317 318 __asm volatile("rd %%psr,%0" : "=r" (psr)); 319 320 oldipl = psr & PSR_PIL; 321 newipl <<= 8; 322 if (newipl <= oldipl) 323 return (oldipl); 324 325 psr = (psr & ~oldipl) | newipl; 326 327 __asm volatile("wr %0,0,%%psr" : : "r" (psr)); 328 __asm volatile("nop; nop; nop" : : : "memory"); 329 330 return (oldipl); 331 } 332 333 #include <sys/spl.h> 334 335 #define splausoft() splraiseipl(makeiplcookie(IPL_SOFTAUDIO)) 336 #define splfdsoft() splraiseipl(makeiplcookie(IPL_SOFTFDC)) 337 338 #define splfd() splraiseipl(makeiplcookie(IPL_FD)) 339 #define splts102() splraiseipl(makeiplcookie(IPL_TS102)) 340 341 #define splzs() splraiseipl(makeiplcookie(IPL_ZS)) 342 343 /* splx does not have a return value */ 344 static __inline __attribute__((__always_inline__)) void 345 splx(int newipl) 346 { 347 int psr; 348 349 __asm volatile("rd %%psr,%0" : "=r" (psr) : : "memory"); 350 __asm volatile("wr %0,%1,%%psr" : : \ 351 "r" (psr & ~PSR_PIL), "rn" (newipl)); 352 __asm volatile("nop; nop; nop"); 353 } 354 #endif /* KERNEL && !_LOCORE */ 355 356 #endif /* PSR_IMPL */ 357