1 /* $NetBSD: psl.h,v 1.53 2024/04/07 17:08:00 rillig Exp $ */
2
3 /*
4 * Copyright (c) 1992, 1993
5 * The Regents of the University of California. All rights reserved.
6 *
7 * This software was developed by the Computer Systems Engineering group
8 * at Lawrence Berkeley Laboratory under DARPA contract BG 91-66 and
9 * contributed to Berkeley.
10 *
11 * All advertising materials mentioning features or use of this software
12 * must display the following acknowledgement:
13 * This product includes software developed by the University of
14 * California, Lawrence Berkeley Laboratory.
15 *
16 * Redistribution and use in source and binary forms, with or without
17 * modification, are permitted provided that the following conditions
18 * are met:
19 * 1. Redistributions of source code must retain the above copyright
20 * notice, this list of conditions and the following disclaimer.
21 * 2. Redistributions in binary form must reproduce the above copyright
22 * notice, this list of conditions and the following disclaimer in the
23 * documentation and/or other materials provided with the distribution.
24 * 3. Neither the name of the University nor the names of its contributors
25 * may be used to endorse or promote products derived from this software
26 * without specific prior written permission.
27 *
28 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
29 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
30 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
31 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
32 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
33 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
34 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
35 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
36 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
37 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
38 * SUCH DAMAGE.
39 *
40 * @(#)psl.h 8.1 (Berkeley) 6/11/93
41 */
42
43 #ifndef PSR_IMPL
44
45 /*
46 * SPARC Process Status Register (in psl.h for hysterical raisins). This
47 * doesn't exist on the V9.
48 *
49 * The picture in the Sun manuals looks like this:
50 * 1 1
51 * 31 28 27 24 23 20 19 14 3 2 11 8 7 6 5 4 0
52 * +-------+-------+-------+-----------+-+-+-------+-+-+-+---------+
53 * | impl | ver | icc | reserved |E|E| pil |S|P|E| CWP |
54 * | | |n z v c| |C|F| | |S|T| |
55 * +-------+-------+-------+-----------+-+-+-------+-+-+-+---------+
56 */
57
58 #define PSR_IMPL 0xf0000000 /* implementation */
59 #define PSR_VER 0x0f000000 /* version */
60 #define PSR_ICC 0x00f00000 /* integer condition codes */
61 #define PSR_N 0x00800000 /* negative */
62 #define PSR_Z 0x00400000 /* zero */
63 #define PSR_O 0x00200000 /* overflow */
64 #define PSR_C 0x00100000 /* carry */
65 #define PSR_EC 0x00002000 /* coprocessor enable */
66 #define PSR_EF 0x00001000 /* FP enable */
67 #define PSR_PIL 0x00000f00 /* interrupt level */
68 #define PSR_S 0x00000080 /* supervisor (kernel) mode */
69 #define PSR_PS 0x00000040 /* previous supervisor mode (traps) */
70 #define PSR_ET 0x00000020 /* trap enable */
71 #define PSR_CWP 0x0000001f /* current window pointer */
72
73 #define PSR_BITS "\20\16EC\15EF\10S\7PS\6ET"
74
75 /*
76 * SPARC V9 CCR register
77 */
78
79 #define ICC_C 0x01L
80 #define ICC_V 0x02L
81 #define ICC_Z 0x04L
82 #define ICC_N 0x08L
83 #define XCC_SHIFT 4
84 #define XCC_C (ICC_C<<XCC_SHIFT)
85 #define XCC_V (ICC_V<<XCC_SHIFT)
86 #define XCC_Z (ICC_Z<<XCC_SHIFT)
87 #define XCC_N (ICC_N<<XCC_SHIFT)
88
89
90 /*
91 * SPARC V9 PSTATE register (what replaces the PSR in V9)
92 *
93 * Here's the layout:
94 *
95 * 11 10 9 8 7 6 5 4 3 2 1 0
96 * +------------------------------------------------------------+
97 * | IG | MG | CLE | TLE | MM | RED | PEF | AM | PRIV | IE | AG |
98 * +------------------------------------------------------------+
99 */
100
101 #define PSTATE_IG 0x800 /* enable spitfire interrupt globals */
102 #define PSTATE_MG 0x400 /* enable spitfire MMU globals */
103 #define PSTATE_CLE 0x200 /* current little endian */
104 #define PSTATE_TLE 0x100 /* traps little endian */
105 #define PSTATE_MM 0x0c0 /* memory model */
106 #define PSTATE_MM_TSO 0x000 /* total store order */
107 #define PSTATE_MM_PSO 0x040 /* partial store order */
108 #define PSTATE_MM_RMO 0x080 /* Relaxed memory order */
109 #define PSTATE_RED 0x020 /* RED state */
110 #define PSTATE_PEF 0x010 /* enable floating point */
111 #define PSTATE_AM 0x008 /* 32-bit address masking */
112 #define PSTATE_PRIV 0x004 /* privileged mode */
113 #define PSTATE_IE 0x002 /* interrupt enable */
114 #define PSTATE_AG 0x001 /* enable alternate globals */
115
116 #define PSTATE_BITS "\177\020" \
117 "b\013IG\0" "b\012MG\0" "b\011CLE\0" "b\010TLE\0" \
118 "F\006\002\0" ":\000MM_TSO\0" ":\001MM_PSO\0" \
119 ":\002MM_RMO\0" "*?\0" "b\005RED\0" "b\004PEF\0" \
120 "b\003AM\0" "b\002PRIV\0" "b\001IE\0" "b\000AG\0"
121
122 /*
123 * 32-bit code requires TSO or at best PSO since that's what's supported on
124 * SPARC V8 and earlier machines.
125 *
126 * 64-bit code sets the memory model in the ELF header.
127 *
128 * We're running kernel code in TSO for the moment so we don't need to worry
129 * about possible memory barrier bugs.
130 */
131
132 #ifdef __arch64__
133 #define PSTATE_PROM (PSTATE_MM_TSO|PSTATE_PRIV)
134 #define PSTATE_NUCLEUS (PSTATE_MM_TSO|PSTATE_PRIV|PSTATE_AG)
135 #define PSTATE_KERN (PSTATE_MM_TSO|PSTATE_PRIV)
136 #define PSTATE_INTR (PSTATE_KERN|PSTATE_IE)
137 #define PSTATE_USER32 (PSTATE_MM_TSO|PSTATE_AM|PSTATE_IE)
138 #define PSTATE_USER (PSTATE_MM_RMO|PSTATE_IE)
139 #else
140 #define PSTATE_PROM (PSTATE_MM_TSO|PSTATE_PRIV)
141 #define PSTATE_NUCLEUS (PSTATE_MM_TSO|PSTATE_AM|PSTATE_PRIV|PSTATE_AG)
142 #define PSTATE_KERN (PSTATE_MM_TSO|PSTATE_AM|PSTATE_PRIV)
143 #define PSTATE_INTR (PSTATE_KERN|PSTATE_IE)
144 #define PSTATE_USER32 (PSTATE_MM_TSO|PSTATE_AM|PSTATE_IE)
145 #define PSTATE_USER (PSTATE_MM_TSO|PSTATE_AM|PSTATE_IE)
146 #endif
147
148
149 /*
150 * SPARC V9 TSTATE register
151 *
152 * 39 32 31 24 23 20 19 8 7 5 4 0
153 * +-----+-----+-----+--------+---+-----+
154 * | CCR | ASI | - | PSTATE | - | CWP |
155 * +-----+-----+-----+--------+---+-----+
156 */
157
158 #define TSTATE_CWP 0x01f
159 #define TSTATE_PSTATE 0xfff00
160 #define TSTATE_PSTATE_SHIFT 8
161 #define TSTATE_ASI 0xff000000LL
162 #define TSTATE_ASI_SHIFT 24
163 #define TSTATE_CCR 0xff00000000LL
164 #define TSTATE_CCR_SHIFT 32
165
166 #define PSRCC_TO_TSTATE(x) (((int64_t)(x)&PSR_ICC)<<(TSTATE_CCR_SHIFT-20))
167 #define TSTATECCR_TO_PSR(x) (((x)&TSTATE_CCR)>>(TSTATE_CCR_SHIFT-20))
168
169 /*
170 * These are here to simplify life.
171 */
172 #define TSTATE_IG (PSTATE_IG<<TSTATE_PSTATE_SHIFT)
173 #define TSTATE_MG (PSTATE_MG<<TSTATE_PSTATE_SHIFT)
174 #define TSTATE_CLE (PSTATE_CLE<<TSTATE_PSTATE_SHIFT)
175 #define TSTATE_TLE (PSTATE_TLE<<TSTATE_PSTATE_SHIFT)
176 #define TSTATE_MM (PSTATE_MM<<TSTATE_PSTATE_SHIFT)
177 #define TSTATE_MM_TSO (PSTATE_MM_TSO<<TSTATE_PSTATE_SHIFT)
178 #define TSTATE_MM_PSO (PSTATE_MM_PSO<<TSTATE_PSTATE_SHIFT)
179 #define TSTATE_MM_RMO (PSTATE_MM_RMO<<TSTATE_PSTATE_SHIFT)
180 #define TSTATE_RED (PSTATE_RED<<TSTATE_PSTATE_SHIFT)
181 #define TSTATE_PEF (PSTATE_PEF<<TSTATE_PSTATE_SHIFT)
182 #define TSTATE_AM (PSTATE_AM<<TSTATE_PSTATE_SHIFT)
183 #define TSTATE_PRIV (PSTATE_PRIV<<TSTATE_PSTATE_SHIFT)
184 #define TSTATE_IE (PSTATE_IE<<TSTATE_PSTATE_SHIFT)
185 #define TSTATE_AG (PSTATE_AG<<TSTATE_PSTATE_SHIFT)
186
187 #define TSTATE_BITS "\20\14IG\13MG\12CLE\11TLE\10\7MM\6RED\5PEF\4AM\3PRIV\2IE\1AG"
188
189 #define TSTATE_KERN ((PSTATE_KERN)<<TSTATE_PSTATE_SHIFT)
190 #define TSTATE_USER ((PSTATE_USER)<<TSTATE_PSTATE_SHIFT)
191 /*
192 * SPARC V9 VER version register.
193 *
194 * 63 48 47 32 31 24 23 16 15 8 7 5 4 0
195 * +-------+------+------+-----+-------+---+--------+
196 * | manuf | impl | mask | - | maxtl | - | maxwin |
197 * +-------+------+------+-----+-------+---+--------+
198 *
199 */
200
201 #define VER_MANUF 0xffff000000000000LL
202 #define VER_MANUF_SHIFT 48
203 #define VER_IMPL 0x0000ffff00000000LL
204 #define VER_IMPL_SHIFT 32
205 #define VER_MASK 0x00000000ff000000LL
206 #define VER_MASK_SHIFT 24
207 #define VER_MAXTL 0x000000000000ff00LL
208 #define VER_MAXTL_SHIFT 8
209 #define VER_MAXWIN 0x000000000000001fLL
210
211 /*
212 * Here are a few things to help us transition between user and kernel mode:
213 */
214
215 /* Memory models */
216 #define KERN_MM PSTATE_MM_TSO
217 #define USER_MM PSTATE_MM_RMO
218
219 /*
220 * Register window handlers. These point to generic routines that check the
221 * stack pointer and then vector to the real handler. We could optimize this
222 * if we could guarantee only 32-bit or 64-bit stacks.
223 */
224 #define WSTATE_KERN 026
225 #define WSTATE_USER 022
226
227 #define CWP 0x01f
228
229 /* 64-byte alignment -- this seems the best place to put this. */
230 #define SPARC64_BLOCK_SIZE 64
231 #define SPARC64_BLOCK_ALIGN 0x3f
232
233 #if (defined(_KERNEL) || defined(_KMEMUSER)) && !defined(_LOCORE)
234 typedef uint8_t ipl_t;
235 typedef struct {
236 ipl_t _ipl;
237 } ipl_cookie_t;
238 #endif /* _KERNEL|_KMEMUSER & !_LOCORE */
239
240 #if defined(_KERNEL) && !defined(_LOCORE)
241
242 /*
243 * GCC pseudo-functions for manipulating PSR (primarily PIL field).
244 */
245 static __inline __attribute__((__always_inline__)) int
getpsr(void)246 getpsr(void)
247 {
248 int psr;
249
250 __asm volatile("rd %%psr,%0" : "=r" (psr));
251 return (psr);
252 }
253
254 static __inline __attribute__((__always_inline__)) int
getmid(void)255 getmid(void)
256 {
257 int mid;
258
259 __asm volatile("rd %%tbr,%0" : "=r" (mid));
260 return ((mid >> 20) & 0x3);
261 }
262
263 static __inline __attribute__((__always_inline__)) void
setpsr(int newpsr)264 setpsr(int newpsr)
265 {
266 __asm volatile("wr %0,0,%%psr" : : "r" (newpsr) : "memory");
267 __asm volatile("nop; nop; nop");
268 }
269
270 static __inline __attribute__((__always_inline__)) void
spl0(void)271 spl0(void)
272 {
273 int psr, oldipl;
274
275 /*
276 * wrpsr xors two values: we choose old psr and old ipl here,
277 * which gives us the same value as the old psr but with all
278 * the old PIL bits turned off.
279 */
280 __asm volatile("rd %%psr,%0" : "=r" (psr) : : "memory");
281 oldipl = psr & PSR_PIL;
282 __asm volatile("wr %0,%1,%%psr" : : "r" (psr), "r" (oldipl));
283
284 /*
285 * Three instructions must execute before we can depend
286 * on the bits to be changed.
287 */
288 __asm volatile("nop; nop; nop");
289 }
290
291 /*
292 * PIL 1 through 14 can use this macro.
293 * (spl0 and splhigh are special since they put all 0s or all 1s
294 * into the ipl field.)
295 */
296 #define _SPLSET(name, newipl) \
297 static __inline __attribute__((__always_inline__)) void name(void) \
298 { \
299 int psr; \
300 __asm volatile("rd %%psr,%0" : "=r" (psr)); \
301 psr &= ~PSR_PIL; \
302 __asm volatile("wr %0,%1,%%psr" : : \
303 "r" (psr), "n" ((newipl) << 8)); \
304 __asm volatile("nop; nop; nop" : : : "memory"); \
305 }
306
_SPLSET(spllowerschedclock,IPL_SCHED)307 _SPLSET(spllowerschedclock, IPL_SCHED)
308
309 static inline __always_inline ipl_cookie_t
310 makeiplcookie(ipl_t ipl)
311 {
312
313 return (ipl_cookie_t){._ipl = ipl};
314 }
315
316 /* Raise IPL and return previous value */
317 static __inline __always_inline int
splraiseipl(ipl_cookie_t icookie)318 splraiseipl(ipl_cookie_t icookie)
319 {
320 int newipl = icookie._ipl;
321 int psr, oldipl;
322
323 __asm volatile("rd %%psr,%0" : "=r" (psr));
324
325 oldipl = psr & PSR_PIL;
326 newipl <<= 8;
327 if (newipl <= oldipl)
328 return (oldipl);
329
330 psr = (psr & ~oldipl) | newipl;
331
332 __asm volatile("wr %0,0,%%psr" : : "r" (psr));
333 __asm volatile("nop; nop; nop" : : : "memory");
334
335 return (oldipl);
336 }
337
338 #include <sys/spl.h>
339
340 #define splausoft() splraiseipl(makeiplcookie(IPL_SOFTAUDIO))
341 #define splfdsoft() splraiseipl(makeiplcookie(IPL_SOFTFDC))
342
343 #define splfd() splraiseipl(makeiplcookie(IPL_FD))
344 #define splts102() splraiseipl(makeiplcookie(IPL_TS102))
345
346 #define splzs() splraiseipl(makeiplcookie(IPL_ZS))
347
348 /* splx does not have a return value */
349 static __inline __attribute__((__always_inline__)) void
splx(int newipl)350 splx(int newipl)
351 {
352 int psr;
353
354 __asm volatile("rd %%psr,%0" : "=r" (psr) : : "memory");
355 __asm volatile("wr %0,%1,%%psr" : : \
356 "r" (psr & ~PSR_PIL), "rn" (newipl));
357 __asm volatile("nop; nop; nop");
358 }
359 #endif /* KERNEL && !_LOCORE */
360
361 #endif /* PSR_IMPL */
362