xref: /netbsd-src/sys/arch/sparc/include/psl.h (revision adf4d5a0e68273481db10d76435b2583b1f6813b)
1 /*	$NetBSD: psl.h,v 1.53 2024/04/07 17:08:00 rillig Exp $ */
2 
3 /*
4  * Copyright (c) 1992, 1993
5  *	The Regents of the University of California.  All rights reserved.
6  *
7  * This software was developed by the Computer Systems Engineering group
8  * at Lawrence Berkeley Laboratory under DARPA contract BG 91-66 and
9  * contributed to Berkeley.
10  *
11  * All advertising materials mentioning features or use of this software
12  * must display the following acknowledgement:
13  *	This product includes software developed by the University of
14  *	California, Lawrence Berkeley Laboratory.
15  *
16  * Redistribution and use in source and binary forms, with or without
17  * modification, are permitted provided that the following conditions
18  * are met:
19  * 1. Redistributions of source code must retain the above copyright
20  *    notice, this list of conditions and the following disclaimer.
21  * 2. Redistributions in binary form must reproduce the above copyright
22  *    notice, this list of conditions and the following disclaimer in the
23  *    documentation and/or other materials provided with the distribution.
24  * 3. Neither the name of the University nor the names of its contributors
25  *    may be used to endorse or promote products derived from this software
26  *    without specific prior written permission.
27  *
28  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
29  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
30  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
31  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
32  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
33  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
34  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
35  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
36  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
37  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
38  * SUCH DAMAGE.
39  *
40  *	@(#)psl.h	8.1 (Berkeley) 6/11/93
41  */
42 
43 #ifndef PSR_IMPL
44 
45 /*
46  * SPARC Process Status Register (in psl.h for hysterical raisins).  This
47  * doesn't exist on the V9.
48  *
49  * The picture in the Sun manuals looks like this:
50  *	                                     1 1
51  *	 31   28 27   24 23   20 19       14 3 2 11    8 7 6 5 4       0
52  *	+-------+-------+-------+-----------+-+-+-------+-+-+-+---------+
53  *	|  impl |  ver  |  icc  |  reserved |E|E|  pil  |S|P|E|   CWP   |
54  *	|       |       |n z v c|           |C|F|       | |S|T|         |
55  *	+-------+-------+-------+-----------+-+-+-------+-+-+-+---------+
56  */
57 
58 #define PSR_IMPL	0xf0000000	/* implementation */
59 #define PSR_VER		0x0f000000	/* version */
60 #define PSR_ICC		0x00f00000	/* integer condition codes */
61 #define PSR_N		0x00800000	/* negative */
62 #define PSR_Z		0x00400000	/* zero */
63 #define PSR_O		0x00200000	/* overflow */
64 #define PSR_C		0x00100000	/* carry */
65 #define PSR_EC		0x00002000	/* coprocessor enable */
66 #define PSR_EF		0x00001000	/* FP enable */
67 #define PSR_PIL		0x00000f00	/* interrupt level */
68 #define PSR_S		0x00000080	/* supervisor (kernel) mode */
69 #define PSR_PS		0x00000040	/* previous supervisor mode (traps) */
70 #define PSR_ET		0x00000020	/* trap enable */
71 #define PSR_CWP		0x0000001f	/* current window pointer */
72 
73 #define PSR_BITS "\20\16EC\15EF\10S\7PS\6ET"
74 
75 /*
76  * SPARC V9 CCR register
77  */
78 
79 #define ICC_C	0x01L
80 #define ICC_V	0x02L
81 #define ICC_Z	0x04L
82 #define ICC_N	0x08L
83 #define XCC_SHIFT	4
84 #define XCC_C	(ICC_C<<XCC_SHIFT)
85 #define XCC_V	(ICC_V<<XCC_SHIFT)
86 #define XCC_Z	(ICC_Z<<XCC_SHIFT)
87 #define XCC_N	(ICC_N<<XCC_SHIFT)
88 
89 
90 /*
91  * SPARC V9 PSTATE register (what replaces the PSR in V9)
92  *
93  * Here's the layout:
94  *
95  *    11   10    9     8   7  6   5     4     3     2     1   0
96  *  +------------------------------------------------------------+
97  *  | IG | MG | CLE | TLE | MM | RED | PEF | AM | PRIV | IE | AG |
98  *  +------------------------------------------------------------+
99  */
100 
101 #define PSTATE_IG	0x800	/* enable spitfire interrupt globals */
102 #define PSTATE_MG	0x400	/* enable spitfire MMU globals */
103 #define PSTATE_CLE	0x200	/* current little endian */
104 #define PSTATE_TLE	0x100	/* traps little endian */
105 #define PSTATE_MM	0x0c0	/* memory model */
106 #define PSTATE_MM_TSO	0x000	/* total store order */
107 #define PSTATE_MM_PSO	0x040	/* partial store order */
108 #define PSTATE_MM_RMO	0x080	/* Relaxed memory order */
109 #define PSTATE_RED	0x020	/* RED state */
110 #define PSTATE_PEF	0x010	/* enable floating point */
111 #define PSTATE_AM	0x008	/* 32-bit address masking */
112 #define PSTATE_PRIV	0x004	/* privileged mode */
113 #define PSTATE_IE	0x002	/* interrupt enable */
114 #define PSTATE_AG	0x001	/* enable alternate globals */
115 
116 #define PSTATE_BITS "\177\020"						\
117 	"b\013IG\0"	"b\012MG\0"	"b\011CLE\0"	"b\010TLE\0"	\
118 			"F\006\002\0"	":\000MM_TSO\0"	":\001MM_PSO\0"	\
119 	":\002MM_RMO\0"	"*?\0"		"b\005RED\0"	"b\004PEF\0"	\
120 	"b\003AM\0"	"b\002PRIV\0"	"b\001IE\0"	"b\000AG\0"
121 
122 /*
123  * 32-bit code requires TSO or at best PSO since that's what's supported on
124  * SPARC V8 and earlier machines.
125  *
126  * 64-bit code sets the memory model in the ELF header.
127  *
128  * We're running kernel code in TSO for the moment so we don't need to worry
129  * about possible memory barrier bugs.
130  */
131 
132 #ifdef __arch64__
133 #define PSTATE_PROM	(PSTATE_MM_TSO|PSTATE_PRIV)
134 #define PSTATE_NUCLEUS	(PSTATE_MM_TSO|PSTATE_PRIV|PSTATE_AG)
135 #define PSTATE_KERN	(PSTATE_MM_TSO|PSTATE_PRIV)
136 #define PSTATE_INTR	(PSTATE_KERN|PSTATE_IE)
137 #define PSTATE_USER32	(PSTATE_MM_TSO|PSTATE_AM|PSTATE_IE)
138 #define PSTATE_USER	(PSTATE_MM_RMO|PSTATE_IE)
139 #else
140 #define PSTATE_PROM	(PSTATE_MM_TSO|PSTATE_PRIV)
141 #define PSTATE_NUCLEUS	(PSTATE_MM_TSO|PSTATE_AM|PSTATE_PRIV|PSTATE_AG)
142 #define PSTATE_KERN	(PSTATE_MM_TSO|PSTATE_AM|PSTATE_PRIV)
143 #define PSTATE_INTR	(PSTATE_KERN|PSTATE_IE)
144 #define PSTATE_USER32	(PSTATE_MM_TSO|PSTATE_AM|PSTATE_IE)
145 #define PSTATE_USER	(PSTATE_MM_TSO|PSTATE_AM|PSTATE_IE)
146 #endif
147 
148 
149 /*
150  * SPARC V9 TSTATE register
151  *
152  *   39 32 31 24 23 20  19   8	7 5 4   0
153  *  +-----+-----+-----+--------+---+-----+
154  *  | CCR | ASI |  -  | PSTATE | - | CWP |
155  *  +-----+-----+-----+--------+---+-----+
156  */
157 
158 #define TSTATE_CWP		0x01f
159 #define TSTATE_PSTATE		0xfff00
160 #define TSTATE_PSTATE_SHIFT	8
161 #define TSTATE_ASI		0xff000000LL
162 #define TSTATE_ASI_SHIFT	24
163 #define TSTATE_CCR		0xff00000000LL
164 #define TSTATE_CCR_SHIFT	32
165 
166 #define PSRCC_TO_TSTATE(x)	(((int64_t)(x)&PSR_ICC)<<(TSTATE_CCR_SHIFT-20))
167 #define TSTATECCR_TO_PSR(x)	(((x)&TSTATE_CCR)>>(TSTATE_CCR_SHIFT-20))
168 
169 /*
170  * These are here to simplify life.
171  */
172 #define TSTATE_IG	(PSTATE_IG<<TSTATE_PSTATE_SHIFT)
173 #define TSTATE_MG	(PSTATE_MG<<TSTATE_PSTATE_SHIFT)
174 #define TSTATE_CLE	(PSTATE_CLE<<TSTATE_PSTATE_SHIFT)
175 #define TSTATE_TLE	(PSTATE_TLE<<TSTATE_PSTATE_SHIFT)
176 #define TSTATE_MM	(PSTATE_MM<<TSTATE_PSTATE_SHIFT)
177 #define TSTATE_MM_TSO	(PSTATE_MM_TSO<<TSTATE_PSTATE_SHIFT)
178 #define TSTATE_MM_PSO	(PSTATE_MM_PSO<<TSTATE_PSTATE_SHIFT)
179 #define TSTATE_MM_RMO	(PSTATE_MM_RMO<<TSTATE_PSTATE_SHIFT)
180 #define TSTATE_RED	(PSTATE_RED<<TSTATE_PSTATE_SHIFT)
181 #define TSTATE_PEF	(PSTATE_PEF<<TSTATE_PSTATE_SHIFT)
182 #define TSTATE_AM	(PSTATE_AM<<TSTATE_PSTATE_SHIFT)
183 #define TSTATE_PRIV	(PSTATE_PRIV<<TSTATE_PSTATE_SHIFT)
184 #define TSTATE_IE	(PSTATE_IE<<TSTATE_PSTATE_SHIFT)
185 #define TSTATE_AG	(PSTATE_AG<<TSTATE_PSTATE_SHIFT)
186 
187 #define TSTATE_BITS "\20\14IG\13MG\12CLE\11TLE\10\7MM\6RED\5PEF\4AM\3PRIV\2IE\1AG"
188 
189 #define TSTATE_KERN	((PSTATE_KERN)<<TSTATE_PSTATE_SHIFT)
190 #define TSTATE_USER	((PSTATE_USER)<<TSTATE_PSTATE_SHIFT)
191 /*
192  * SPARC V9 VER version register.
193  *
194  *  63   48 47  32 31  24 23 16 15    8 7 5 4      0
195  * +-------+------+------+-----+-------+---+--------+
196  * | manuf | impl | mask |  -  | maxtl | - | maxwin |
197  * +-------+------+------+-----+-------+---+--------+
198  *
199  */
200 
201 #define VER_MANUF	0xffff000000000000LL
202 #define VER_MANUF_SHIFT	48
203 #define VER_IMPL	0x0000ffff00000000LL
204 #define VER_IMPL_SHIFT	32
205 #define VER_MASK	0x00000000ff000000LL
206 #define VER_MASK_SHIFT	24
207 #define VER_MAXTL	0x000000000000ff00LL
208 #define VER_MAXTL_SHIFT	8
209 #define VER_MAXWIN	0x000000000000001fLL
210 
211 /*
212  * Here are a few things to help us transition between user and kernel mode:
213  */
214 
215 /* Memory models */
216 #define KERN_MM		PSTATE_MM_TSO
217 #define USER_MM		PSTATE_MM_RMO
218 
219 /*
220  * Register window handlers.  These point to generic routines that check the
221  * stack pointer and then vector to the real handler.  We could optimize this
222  * if we could guarantee only 32-bit or 64-bit stacks.
223  */
224 #define WSTATE_KERN	026
225 #define WSTATE_USER	022
226 
227 #define CWP		0x01f
228 
229 /* 64-byte alignment -- this seems the best place to put this. */
230 #define SPARC64_BLOCK_SIZE	64
231 #define SPARC64_BLOCK_ALIGN	0x3f
232 
233 #if (defined(_KERNEL) || defined(_KMEMUSER)) && !defined(_LOCORE)
234 typedef uint8_t ipl_t;
235 typedef struct {
236 	ipl_t _ipl;
237 } ipl_cookie_t;
238 #endif	/* _KERNEL|_KMEMUSER & !_LOCORE */
239 
240 #if defined(_KERNEL) && !defined(_LOCORE)
241 
242 /*
243  * GCC pseudo-functions for manipulating PSR (primarily PIL field).
244  */
245 static __inline __attribute__((__always_inline__)) int
getpsr(void)246 getpsr(void)
247 {
248 	int psr;
249 
250 	__asm volatile("rd %%psr,%0" : "=r" (psr));
251 	return (psr);
252 }
253 
254 static __inline __attribute__((__always_inline__)) int
getmid(void)255 getmid(void)
256 {
257 	int mid;
258 
259 	__asm volatile("rd %%tbr,%0" : "=r" (mid));
260 	return ((mid >> 20) & 0x3);
261 }
262 
263 static __inline __attribute__((__always_inline__)) void
setpsr(int newpsr)264 setpsr(int newpsr)
265 {
266 	__asm volatile("wr %0,0,%%psr" : : "r" (newpsr) : "memory");
267 	__asm volatile("nop; nop; nop");
268 }
269 
270 static __inline __attribute__((__always_inline__)) void
spl0(void)271 spl0(void)
272 {
273 	int psr, oldipl;
274 
275 	/*
276 	 * wrpsr xors two values: we choose old psr and old ipl here,
277 	 * which gives us the same value as the old psr but with all
278 	 * the old PIL bits turned off.
279 	 */
280 	__asm volatile("rd %%psr,%0" : "=r" (psr) : : "memory");
281 	oldipl = psr & PSR_PIL;
282 	__asm volatile("wr %0,%1,%%psr" : : "r" (psr), "r" (oldipl));
283 
284 	/*
285 	 * Three instructions must execute before we can depend
286 	 * on the bits to be changed.
287 	 */
288 	__asm volatile("nop; nop; nop");
289 }
290 
291 /*
292  * PIL 1 through 14 can use this macro.
293  * (spl0 and splhigh are special since they put all 0s or all 1s
294  * into the ipl field.)
295  */
296 #define	_SPLSET(name, newipl) \
297 static __inline __attribute__((__always_inline__)) void name(void) \
298 { \
299 	int psr; \
300 	__asm volatile("rd %%psr,%0" : "=r" (psr)); \
301 	psr &= ~PSR_PIL; \
302 	__asm volatile("wr %0,%1,%%psr" : : \
303 	    "r" (psr), "n" ((newipl) << 8)); \
304 	__asm volatile("nop; nop; nop" : : : "memory"); \
305 }
306 
_SPLSET(spllowerschedclock,IPL_SCHED)307 _SPLSET(spllowerschedclock, IPL_SCHED)
308 
309 static inline __always_inline ipl_cookie_t
310 makeiplcookie(ipl_t ipl)
311 {
312 
313 	return (ipl_cookie_t){._ipl = ipl};
314 }
315 
316 /* Raise IPL and return previous value */
317 static __inline __always_inline int
splraiseipl(ipl_cookie_t icookie)318 splraiseipl(ipl_cookie_t icookie)
319 {
320 	int newipl = icookie._ipl;
321 	int psr, oldipl;
322 
323 	__asm volatile("rd %%psr,%0" : "=r" (psr));
324 
325 	oldipl = psr & PSR_PIL;
326 	newipl <<= 8;
327 	if (newipl <= oldipl)
328 		return (oldipl);
329 
330 	psr = (psr & ~oldipl) | newipl;
331 
332 	__asm volatile("wr %0,0,%%psr" : : "r" (psr));
333 	__asm volatile("nop; nop; nop" : : : "memory");
334 
335 	return (oldipl);
336 }
337 
338 #include <sys/spl.h>
339 
340 #define	splausoft()	splraiseipl(makeiplcookie(IPL_SOFTAUDIO))
341 #define	splfdsoft()	splraiseipl(makeiplcookie(IPL_SOFTFDC))
342 
343 #define	splfd()		splraiseipl(makeiplcookie(IPL_FD))
344 #define	splts102()	splraiseipl(makeiplcookie(IPL_TS102))
345 
346 #define	splzs()		splraiseipl(makeiplcookie(IPL_ZS))
347 
348 /* splx does not have a return value */
349 static __inline __attribute__((__always_inline__)) void
splx(int newipl)350 splx(int newipl)
351 {
352 	int psr;
353 
354 	__asm volatile("rd %%psr,%0" : "=r" (psr) : : "memory");
355 	__asm volatile("wr %0,%1,%%psr" : : \
356 	    "r" (psr & ~PSR_PIL), "rn" (newipl));
357 	__asm volatile("nop; nop; nop");
358 }
359 #endif /* KERNEL && !_LOCORE */
360 
361 #endif /* PSR_IMPL */
362