1 /* $NetBSD: lock.h,v 1.27 2007/10/17 19:57:48 garbled Exp $ */ 2 3 /* 4 * Copyright (c) 2000 Ludd, University of Lule}, Sweden. 5 * All rights reserved. 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions 9 * are met: 10 * 1. Redistributions of source code must retain the above copyright 11 * notice, this list of conditions and the following disclaimer. 12 * 2. Redistributions in binary form must reproduce the above copyright 13 * notice, this list of conditions and the following disclaimer in the 14 * documentation and/or other materials provided with the distribution. 15 * 3. All advertising materials mentioning features or use of this software 16 * must display the following acknowledgement: 17 * This product includes software developed at Ludd, University of Lule}. 18 * 4. The name of the author may not be used to endorse or promote products 19 * derived from this software without specific prior written permission 20 * 21 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 22 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 23 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 24 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, 25 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT 26 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 27 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 28 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 29 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 30 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 31 */ 32 33 #ifndef _VAX_LOCK_H_ 34 #define _VAX_LOCK_H_ 35 36 #ifdef _KERNEL 37 #ifdef _KERNEL_OPT 38 #include "opt_multiprocessor.h" 39 #include <machine/intr.h> 40 #endif 41 #include <machine/cpu.h> 42 #endif 43 44 static __inline int 45 __SIMPLELOCK_LOCKED_P(__cpu_simple_lock_t *__ptr) 46 { 47 return *__ptr == __SIMPLELOCK_LOCKED; 48 } 49 50 static __inline int 51 __SIMPLELOCK_UNLOCKED_P(__cpu_simple_lock_t *__ptr) 52 { 53 return *__ptr == __SIMPLELOCK_UNLOCKED; 54 } 55 56 static __inline void 57 __cpu_simple_lock_clear(__cpu_simple_lock_t *__ptr) 58 { 59 *__ptr = __SIMPLELOCK_UNLOCKED; 60 } 61 62 static __inline void 63 __cpu_simple_lock_set(__cpu_simple_lock_t *__ptr) 64 { 65 *__ptr = __SIMPLELOCK_LOCKED; 66 } 67 68 static __inline void __cpu_simple_lock_init(__cpu_simple_lock_t *); 69 static __inline void 70 __cpu_simple_lock_init(__cpu_simple_lock_t *__alp) 71 { 72 #ifdef _KERNEL 73 __asm volatile ("movl %0,%%r1;jsb Sunlock" 74 : /* No output */ 75 : "g"(__alp) 76 : "r1","cc","memory"); 77 #else 78 __asm volatile ("bbcci $0,%0,1f;1:" 79 : /* No output */ 80 : "m"(*__alp) 81 : "cc"); 82 #endif 83 } 84 85 static __inline int __cpu_simple_lock_try(__cpu_simple_lock_t *); 86 static __inline int 87 __cpu_simple_lock_try(__cpu_simple_lock_t *__alp) 88 { 89 int ret; 90 91 #ifdef _KERNEL 92 __asm volatile ("movl %1,%%r1;jsb Slocktry;movl %%r0,%0" 93 : "=&r"(ret) 94 : "g"(__alp) 95 : "r0","r1","cc","memory"); 96 #else 97 __asm volatile ("clrl %0;bbssi $0,%1,1f;incl %0;1:" 98 : "=&r"(ret) 99 : "m"(*__alp) 100 : "cc"); 101 #endif 102 103 return ret; 104 } 105 106 #ifdef _KERNEL 107 #if defined(MULTIPROCESSOR) 108 #define VAX_LOCK_CHECKS ((1 << IPI_SEND_CNCHAR) | (1 << IPI_DDB)) 109 #define __cpu_simple_lock(__alp) \ 110 do { \ 111 struct cpu_info *__ci = curcpu(); \ 112 \ 113 while (__cpu_simple_lock_try(__alp) == 0) { \ 114 int __s; \ 115 \ 116 if (__ci->ci_ipimsgs & VAX_LOCK_CHECKS) { \ 117 __s = splipi(); \ 118 cpu_handle_ipi(); \ 119 splx(__s); \ 120 } \ 121 } \ 122 } while (/*CONSTCOND*/0) 123 #else /* MULTIPROCESSOR */ 124 #define __cpu_simple_lock(__alp) \ 125 do { \ 126 while (__cpu_simple_lock_try(__alp) == 0) { \ 127 ; \ 128 } \ 129 } while (/*CONSTCOND*/0) 130 #endif 131 #else 132 static __inline void __cpu_simple_lock(__cpu_simple_lock_t *); 133 static __inline void 134 __cpu_simple_lock(__cpu_simple_lock_t *__alp) 135 { 136 __asm volatile ("1:bbssi $0,%0,1b" 137 : /* No outputs */ 138 : "m"(*__alp) 139 : "cc"); 140 } 141 #endif /* _KERNEL */ 142 143 #if 0 144 static __inline void __cpu_simple_lock(__cpu_simple_lock_t *); 145 static __inline void 146 __cpu_simple_lock(__cpu_simple_lock_t *__alp) 147 { 148 struct cpu_info *ci = curcpu(); 149 150 while (__cpu_simple_lock_try(__alp) == 0) { 151 int s; 152 153 if (ci->ci_ipimsgs & IPI_SEND_CNCHAR) { 154 s = splipi(); 155 cpu_handle_ipi(); 156 splx(s); 157 } 158 } 159 160 #if 0 161 __asm volatile ("movl %0,%%r1;jsb Slock" 162 : /* No output */ 163 : "g"(__alp) 164 : "r0","r1","cc","memory"); 165 #endif 166 #if 0 167 __asm volatile ("1:;bbssi $0, %0, 1b" 168 : /* No output */ 169 : "m"(*__alp)); 170 #endif 171 } 172 #endif 173 174 static __inline void __cpu_simple_unlock(__cpu_simple_lock_t *); 175 static __inline void 176 __cpu_simple_unlock(__cpu_simple_lock_t *__alp) 177 { 178 #ifdef _KERNEL 179 __asm volatile ("movl %0,%%r1;jsb Sunlock" 180 : /* No output */ 181 : "g"(__alp) 182 : "r1","cc","memory"); 183 #else 184 __asm volatile ("bbcci $0,%0,1f;1:" 185 : /* No output */ 186 : "m"(*__alp) 187 : "cc"); 188 #endif 189 } 190 191 #if defined(MULTIPROCESSOR) 192 /* 193 * On the Vax, interprocessor interrupts can come in at device priority 194 * level or lower. This can cause some problems while waiting for r/w 195 * spinlocks from a high'ish priority level: IPIs that come in will not 196 * be processed. This can lead to deadlock. 197 * 198 * This hook allows IPIs to be processed while a spinlock's interlock 199 * is released. 200 */ 201 #define SPINLOCK_SPIN_HOOK \ 202 do { \ 203 struct cpu_info *__ci = curcpu(); \ 204 int __s; \ 205 \ 206 if (__ci->ci_ipimsgs != 0) { \ 207 /* printf("CPU %lu has IPIs pending\n", \ 208 __ci->ci_cpuid); */ \ 209 __s = splipi(); \ 210 cpu_handle_ipi(); \ 211 splx(__s); \ 212 } \ 213 } while (/*CONSTCOND*/0) 214 #endif /* MULTIPROCESSOR */ 215 216 static __inline void mb_read(void); 217 static __inline void 218 mb_read(void) 219 { 220 } 221 222 static __inline void mb_write(void); 223 static __inline void 224 mb_write(void) 225 { 226 } 227 #endif /* _VAX_LOCK_H_ */ 228