/* $NetBSD: profile.h,v 1.18 2016/04/11 14:15:30 bouyer Exp $ */ /* * Copyright (c) 1992, 1993 * The Regents of the University of California. All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. Neither the name of the University nor the names of its contributors * may be used to endorse or promote products derived from this software * without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * * @(#)profile.h 8.1 (Berkeley) 6/11/93 */ #ifdef __x86_64__ #ifdef _KERNEL_OPT #include "opt_xen.h" #endif #define _MCOUNT_DECL void _mcount #define EPROL_EXPORT __asm(".globl _eprol") #ifdef __PIC__ #define __MCPLT "@PLT" #else #define __MCPLT #endif #define MCOUNT \ __weak_alias(mcount, __mcount) \ __asm(" .globl __mcount \n" \ " .type __mcount,@function\n" \ "__mcount: \n" \ " pushq %rbp \n" \ " movq %rsp,%rbp \n" \ " subq $56,%rsp \n" \ " movq %rdi,0(%rsp) \n" \ " movq %rsi,8(%rsp) \n" \ " movq %rdx,16(%rsp) \n" \ " movq %rcx,24(%rsp) \n" \ " movq %r8,32(%rsp) \n" \ " movq %r9,40(%rsp) \n" \ " movq %rax,48(%rsp) \n" \ " movq 0(%rbp),%r11 \n" \ " movq 8(%r11),%rdi \n" \ " movq 8(%rbp),%rsi \n" \ " call _mcount"__MCPLT " \n" \ " movq 0(%rsp),%rdi \n" \ " movq 8(%rsp),%rsi \n" \ " movq 16(%rsp),%rdx \n" \ " movq 24(%rsp),%rcx \n" \ " movq 32(%rsp),%r8 \n" \ " movq 40(%rsp),%r9 \n" \ " movq 48(%rsp),%rax \n" \ " leave \n" \ " ret \n" \ " .size __mcount,.-__mcount"); #ifdef _KERNEL #ifdef XEN static inline void mcount_disable_intr(void) { /* should be __cli() but this calls x86_lfence() which calls mcount */ curcpu()->ci_vcpu->evtchn_upcall_mask = 1; __asm volatile("lfence" ::: "memory"); /* x86_lfence() */ } static inline u_long mcount_read_psl(void) { return (curcpu()->ci_vcpu->evtchn_upcall_mask); } static inline void mcount_write_psl(u_long psl) { curcpu()->ci_vcpu->evtchn_upcall_mask = psl; /* can't call x86_lfence because it calls mcount() */ __asm volatile("lfence" ::: "memory"); /* x86_lfence() */ /* XXX can't call hypervisor_force_callback() because we're in mcount*/ } #else /* XEN */ static inline void mcount_disable_intr(void) { __asm volatile("cli"); } static inline u_long mcount_read_psl(void) { u_long ef; __asm volatile("pushfq; popq %0" : "=r" (ef)); return (ef); } static inline void mcount_write_psl(u_long ef) { __asm volatile("pushq %0; popfq" : : "r" (ef)); } #endif /* XEN */ #define MCOUNT_ENTER \ do { s = (int)mcount_read_psl(); mcount_disable_intr(); } while (0) #define MCOUNT_EXIT do { mcount_write_psl(s); } while (0) #endif /* _KERNEL */ #else /* __x86_64__ */ #include #endif /* __x86_64__ */