xref: /netbsd-src/sys/arch/vax/include/lock.h (revision 7fa608457b817eca6e0977b37f758ae064f3c99c)
1 /*	$NetBSD: lock.h,v 1.27 2007/10/17 19:57:48 garbled Exp $	*/
2 
3 /*
4  * Copyright (c) 2000 Ludd, University of Lule}, Sweden.
5  * All rights reserved.
6  *
7  * Redistribution and use in source and binary forms, with or without
8  * modification, are permitted provided that the following conditions
9  * are met:
10  * 1. Redistributions of source code must retain the above copyright
11  *    notice, this list of conditions and the following disclaimer.
12  * 2. Redistributions in binary form must reproduce the above copyright
13  *    notice, this list of conditions and the following disclaimer in the
14  *    documentation and/or other materials provided with the distribution.
15  * 3. All advertising materials mentioning features or use of this software
16  *    must display the following acknowledgement:
17  *     This product includes software developed at Ludd, University of Lule}.
18  * 4. The name of the author may not be used to endorse or promote products
19  *    derived from this software without specific prior written permission
20  *
21  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
22  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
23  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
24  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
25  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
26  * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
27  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
28  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
29  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
30  * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
31  */
32 
33 #ifndef _VAX_LOCK_H_
34 #define _VAX_LOCK_H_
35 
36 #ifdef _KERNEL
37 #ifdef _KERNEL_OPT
38 #include "opt_multiprocessor.h"
39 #include <machine/intr.h>
40 #endif
41 #include <machine/cpu.h>
42 #endif
43 
44 static __inline int
45 __SIMPLELOCK_LOCKED_P(__cpu_simple_lock_t *__ptr)
46 {
47 	return *__ptr == __SIMPLELOCK_LOCKED;
48 }
49 
50 static __inline int
51 __SIMPLELOCK_UNLOCKED_P(__cpu_simple_lock_t *__ptr)
52 {
53 	return *__ptr == __SIMPLELOCK_UNLOCKED;
54 }
55 
56 static __inline void
57 __cpu_simple_lock_clear(__cpu_simple_lock_t *__ptr)
58 {
59 	*__ptr = __SIMPLELOCK_UNLOCKED;
60 }
61 
62 static __inline void
63 __cpu_simple_lock_set(__cpu_simple_lock_t *__ptr)
64 {
65 	*__ptr = __SIMPLELOCK_LOCKED;
66 }
67 
68 static __inline void __cpu_simple_lock_init(__cpu_simple_lock_t *);
69 static __inline void
70 __cpu_simple_lock_init(__cpu_simple_lock_t *__alp)
71 {
72 #ifdef _KERNEL
73 	__asm volatile ("movl %0,%%r1;jsb Sunlock"
74 		: /* No output */
75 		: "g"(__alp)
76 		: "r1","cc","memory");
77 #else
78 	__asm volatile ("bbcci $0,%0,1f;1:"
79 		: /* No output */
80 		: "m"(*__alp)
81 		: "cc");
82 #endif
83 }
84 
85 static __inline int __cpu_simple_lock_try(__cpu_simple_lock_t *);
86 static __inline int
87 __cpu_simple_lock_try(__cpu_simple_lock_t *__alp)
88 {
89 	int ret;
90 
91 #ifdef _KERNEL
92 	__asm volatile ("movl %1,%%r1;jsb Slocktry;movl %%r0,%0"
93 		: "=&r"(ret)
94 		: "g"(__alp)
95 		: "r0","r1","cc","memory");
96 #else
97 	__asm volatile ("clrl %0;bbssi $0,%1,1f;incl %0;1:"
98 		: "=&r"(ret)
99 		: "m"(*__alp)
100 		: "cc");
101 #endif
102 
103 	return ret;
104 }
105 
106 #ifdef _KERNEL
107 #if defined(MULTIPROCESSOR)
108 #define	VAX_LOCK_CHECKS ((1 << IPI_SEND_CNCHAR) | (1 << IPI_DDB))
109 #define	__cpu_simple_lock(__alp)					\
110 do {									\
111 	struct cpu_info *__ci = curcpu();				\
112 									\
113 	while (__cpu_simple_lock_try(__alp) == 0) {			\
114 		int __s;						\
115 									\
116 		if (__ci->ci_ipimsgs & VAX_LOCK_CHECKS) {		\
117 			__s = splipi();				\
118 			cpu_handle_ipi();				\
119 			splx(__s);					\
120 		}							\
121 	}								\
122 } while (/*CONSTCOND*/0)
123 #else /* MULTIPROCESSOR */
124 #define __cpu_simple_lock(__alp)					\
125 do {									\
126 	while (__cpu_simple_lock_try(__alp) == 0) {			\
127 		;							\
128 	}								\
129 } while (/*CONSTCOND*/0)
130 #endif
131 #else
132 static __inline void __cpu_simple_lock(__cpu_simple_lock_t *);
133 static __inline void
134 __cpu_simple_lock(__cpu_simple_lock_t *__alp)
135 {
136 	__asm volatile ("1:bbssi $0,%0,1b"
137 		: /* No outputs */
138 		: "m"(*__alp)
139 		: "cc");
140 }
141 #endif /* _KERNEL */
142 
143 #if 0
144 static __inline void __cpu_simple_lock(__cpu_simple_lock_t *);
145 static __inline void
146 __cpu_simple_lock(__cpu_simple_lock_t *__alp)
147 {
148 	struct cpu_info *ci = curcpu();
149 
150 	while (__cpu_simple_lock_try(__alp) == 0) {
151 		int s;
152 
153 		if (ci->ci_ipimsgs & IPI_SEND_CNCHAR) {
154 			s = splipi();
155 			cpu_handle_ipi();
156 			splx(s);
157 		}
158 	}
159 
160 #if 0
161 	__asm volatile ("movl %0,%%r1;jsb Slock"
162 		: /* No output */
163 		: "g"(__alp)
164 		: "r0","r1","cc","memory");
165 #endif
166 #if 0
167 	__asm volatile ("1:;bbssi $0, %0, 1b"
168 		: /* No output */
169 		: "m"(*__alp));
170 #endif
171 }
172 #endif
173 
174 static __inline void __cpu_simple_unlock(__cpu_simple_lock_t *);
175 static __inline void
176 __cpu_simple_unlock(__cpu_simple_lock_t *__alp)
177 {
178 #ifdef _KERNEL
179 	__asm volatile ("movl %0,%%r1;jsb Sunlock"
180 		: /* No output */
181 		: "g"(__alp)
182 		: "r1","cc","memory");
183 #else
184 	__asm volatile ("bbcci $0,%0,1f;1:"
185 		: /* No output */
186 		: "m"(*__alp)
187 		: "cc");
188 #endif
189 }
190 
191 #if defined(MULTIPROCESSOR)
192 /*
193  * On the Vax, interprocessor interrupts can come in at device priority
194  * level or lower. This can cause some problems while waiting for r/w
195  * spinlocks from a high'ish priority level: IPIs that come in will not
196  * be processed. This can lead to deadlock.
197  *
198  * This hook allows IPIs to be processed while a spinlock's interlock
199  * is released.
200  */
201 #define SPINLOCK_SPIN_HOOK						\
202 do {									\
203 	struct cpu_info *__ci = curcpu();				\
204 	int __s;							\
205 									\
206 	if (__ci->ci_ipimsgs != 0) {					\
207 		/* printf("CPU %lu has IPIs pending\n",			\
208 		    __ci->ci_cpuid); */					\
209 		__s = splipi();						\
210 		cpu_handle_ipi();					\
211 		splx(__s);						\
212 	}								\
213 } while (/*CONSTCOND*/0)
214 #endif /* MULTIPROCESSOR */
215 
216 static __inline void mb_read(void);
217 static __inline void
218 mb_read(void)
219 {
220 }
221 
222 static __inline void mb_write(void);
223 static __inline void
224 mb_write(void)
225 {
226 }
227 #endif /* _VAX_LOCK_H_ */
228