1 /* $NetBSD: profile.h,v 1.11 2007/11/24 18:55:41 bouyer Exp $ */ 2 3 /* 4 * Copyright (c) 1992, 1993 5 * The Regents of the University of California. All rights reserved. 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions 9 * are met: 10 * 1. Redistributions of source code must retain the above copyright 11 * notice, this list of conditions and the following disclaimer. 12 * 2. Redistributions in binary form must reproduce the above copyright 13 * notice, this list of conditions and the following disclaimer in the 14 * documentation and/or other materials provided with the distribution. 15 * 3. Neither the name of the University nor the names of its contributors 16 * may be used to endorse or promote products derived from this software 17 * without specific prior written permission. 18 * 19 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND 20 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 21 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 22 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE 23 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 24 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 25 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 26 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 27 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 28 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 29 * SUCH DAMAGE. 30 * 31 * @(#)profile.h 8.1 (Berkeley) 6/11/93 32 */ 33 34 #ifdef _KERNEL_OPT 35 #include "opt_multiprocessor.h" 36 #include "opt_xen.h" 37 #endif 38 39 #include <machine/atomic.h> 40 41 #define _MCOUNT_DECL void _mcount 42 43 #define EPROL_EXPORT __asm(".globl _eprol") 44 45 #ifdef PIC 46 #define __MCPLT "@PLT" 47 #else 48 #define __MCPLT 49 #endif 50 51 #define MCOUNT \ 52 __weak_alias(mcount, __mcount) \ 53 __asm(" .globl __mcount \n" \ 54 " .type __mcount,@function\n" \ 55 "__mcount: \n" \ 56 " pushq %rbp \n" \ 57 " movq %rsp,%rbp \n" \ 58 " subq $56,%rsp \n" \ 59 " movq %rdi,0(%rsp) \n" \ 60 " movq %rsi,8(%rsp) \n" \ 61 " movq %rdx,16(%rsp) \n" \ 62 " movq %rcx,24(%rsp) \n" \ 63 " movq %r8,32(%rsp) \n" \ 64 " movq %r9,40(%rsp) \n" \ 65 " movq %rax,48(%rsp) \n" \ 66 " movq 0(%rbp),%r11 \n" \ 67 " movq 8(%r11),%rdi \n" \ 68 " movq 8(%rbp),%rsi \n" \ 69 " call _mcount"__MCPLT " \n" \ 70 " movq 0(%rsp),%rdi \n" \ 71 " movq 8(%rsp),%rsi \n" \ 72 " movq 16(%rsp),%rdx \n" \ 73 " movq 24(%rsp),%rcx \n" \ 74 " movq 32(%rsp),%r8 \n" \ 75 " movq 40(%rsp),%r9 \n" \ 76 " movq 48(%rsp),%rax \n" \ 77 " leave \n" \ 78 " ret \n" \ 79 " .size __mcount,.-__mcount"); 80 81 82 #ifdef _KERNEL 83 #ifdef MULTIPROCESSOR 84 __cpu_simple_lock_t __mcount_lock; 85 86 static inline void 87 MCOUNT_ENTER_MP(void) 88 { 89 while (x86_atomic_testset_b(&__mcount_lock, __SIMPLELOCK_LOCKED) 90 != __SIMPLELOCK_UNLOCKED) { 91 while (__mcount_lock == __SIMPLELOCK_LOCKED) 92 ; 93 } 94 __insn_barrier(); 95 } 96 97 static inline void 98 MCOUNT_EXIT_MP(void) 99 { 100 __insn_barrier(); 101 __mcount_lock = __SIMPLELOCK_UNLOCKED; 102 } 103 #else 104 #define MCOUNT_ENTER_MP() 105 #define MCOUNT_EXIT_MP() 106 #endif 107 108 #ifdef XEN 109 static inline void 110 mcount_disable_intr(void) 111 { 112 /* works because __cli is a macro */ 113 __cli(); 114 } 115 116 static inline u_long 117 mcount_read_psl(void) 118 { 119 return (HYPERVISOR_shared_info->vcpu_info[0].evtchn_upcall_mask); 120 } 121 122 static inline void 123 mcount_write_psl(u_long psl) 124 { 125 HYPERVISOR_shared_info->vcpu_info[0].evtchn_upcall_mask = psl; 126 x86_lfence(); 127 /* XXX can't call hypervisor_force_callback() because we're in mcount*/ 128 } 129 130 #else /* XEN */ 131 static inline void 132 mcount_disable_intr(void) 133 { 134 __asm volatile("cli"); 135 } 136 137 static inline u_long 138 mcount_read_psl(void) 139 { 140 u_long ef; 141 142 __asm volatile("pushfl; popl %0" : "=r" (ef)); 143 return (ef); 144 } 145 146 static inline void 147 mcount_write_psl(u_long ef) 148 { 149 __asm volatile("pushl %0; popfl" : : "r" (ef)); 150 } 151 152 #endif /* XEN */ 153 #define MCOUNT_ENTER \ 154 s = (int)mcount_read_psl(); \ 155 mcount_disable_intr(); \ 156 MCOUNT_ENTER_MP(); 157 158 #define MCOUNT_EXIT \ 159 MCOUNT_EXIT_MP(); \ 160 mcount_write_psl(s); 161 162 #endif /* _KERNEL */ 163