Home | History | Annotate | Line # | Download | only in include
      1  1.4       ryo /* $NetBSD: profile.h,v 1.4 2021/02/10 12:31:34 ryo Exp $ */
      2  1.1      matt 
      3  1.1      matt /*-
      4  1.1      matt  * Copyright (c) 2014 The NetBSD Foundation, Inc.
      5  1.1      matt  * All rights reserved.
      6  1.1      matt  *
      7  1.1      matt  * This code is derived from software contributed to The NetBSD Foundation
      8  1.1      matt  * by Matt Thomas of 3am Software Foundry.
      9  1.1      matt  *
     10  1.1      matt  * Redistribution and use in source and binary forms, with or without
     11  1.1      matt  * modification, are permitted provided that the following conditions
     12  1.1      matt  * are met:
     13  1.1      matt  * 1. Redistributions of source code must retain the above copyright
     14  1.1      matt  *    notice, this list of conditions and the following disclaimer.
     15  1.1      matt  * 2. Redistributions in binary form must reproduce the above copyright
     16  1.1      matt  *    notice, this list of conditions and the following disclaimer in the
     17  1.1      matt  *    documentation and/or other materials provided with the distribution.
     18  1.1      matt  *
     19  1.1      matt  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
     20  1.1      matt  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
     21  1.1      matt  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
     22  1.1      matt  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
     23  1.1      matt  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
     24  1.1      matt  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
     25  1.1      matt  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     26  1.1      matt  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
     27  1.1      matt  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
     28  1.1      matt  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     29  1.1      matt  * POSSIBILITY OF SUCH DAMAGE.
     30  1.1      matt  */
     31  1.1      matt 
     32  1.1      matt #ifdef __aarch64__
     33  1.1      matt 
     34  1.4       ryo #define	_MCOUNT_DECL void mcount
     35  1.1      matt 
     36  1.1      matt /*
     37  1.1      matt  * Cannot implement mcount in C as GCC will trash the ip register when it
     38  1.1      matt  * pushes a trapframe. Pity we cannot insert assembly before the function
     39  1.1      matt  * prologue.
     40  1.1      matt  */
     41  1.1      matt 
     42  1.4       ryo #define MCOUNT_ASM_NAME "_mcount"		/* gcc */
     43  1.4       ryo #define MCOUNT_ASM_NAME_ALIAS "__mcount"	/* llvm */
     44  1.1      matt #define	PLTSYM
     45  1.1      matt 
     46  1.1      matt #define	MCOUNT								\
     47  1.1      matt 	__asm(".text");							\
     48  1.3       ryo 	__asm(".align	6");						\
     49  1.1      matt 	__asm(".type	" MCOUNT_ASM_NAME ",@function");		\
     50  1.1      matt 	__asm(".global	" MCOUNT_ASM_NAME);				\
     51  1.4       ryo 	__asm(".global	" MCOUNT_ASM_NAME_ALIAS);			\
     52  1.1      matt 	__asm(MCOUNT_ASM_NAME ":");					\
     53  1.4       ryo 	__asm(MCOUNT_ASM_NAME_ALIAS ":");				\
     54  1.1      matt 	/*								\
     55  1.1      matt 	 * Preserve registers that are trashed during mcount		\
     56  1.1      matt 	 */								\
     57  1.2  jakllsch 	__asm("stp	x29, x30, [sp, #-80]!");			\
     58  1.2  jakllsch 	__asm("stp	x0, x1, [sp, #16]");				\
     59  1.2  jakllsch 	__asm("stp	x2, x3, [sp, #32]");				\
     60  1.2  jakllsch 	__asm("stp	x4, x5, [sp, #48]");				\
     61  1.2  jakllsch 	__asm("stp	x6, x7, [sp, #64]");				\
     62  1.1      matt 	/*								\
     63  1.1      matt 	 * find the return address for mcount,				\
     64  1.1      matt 	 * and the return address for mcount's caller.			\
     65  1.1      matt 	 *								\
     66  1.1      matt 	 * frompcindex = pc pushed by call into self.			\
     67  1.1      matt 	 */								\
     68  1.2  jakllsch 	__asm("ldr	x0, [x29, #8]");				\
     69  1.1      matt 	/*								\
     70  1.1      matt 	 * selfpc = pc pushed by mcount call				\
     71  1.1      matt 	 */								\
     72  1.1      matt 	__asm("mov	x1, x30");					\
     73  1.1      matt 	/*								\
     74  1.1      matt 	 * Call the real mcount code					\
     75  1.1      matt 	 */								\
     76  1.4       ryo 	__asm("bl	" ___STRING(_C_LABEL(mcount)));			\
     77  1.1      matt 	/*								\
     78  1.1      matt 	 * Restore registers that were trashed during mcount		\
     79  1.1      matt 	 */								\
     80  1.2  jakllsch 	__asm("ldp	x0, x1, [sp, #16]");				\
     81  1.2  jakllsch 	__asm("ldp	x2, x3, [sp, #32]");				\
     82  1.2  jakllsch 	__asm("ldp	x4, x5, [sp, #48]");				\
     83  1.2  jakllsch 	__asm("ldp	x6, x7, [sp, #64]");				\
     84  1.2  jakllsch 	__asm("ldp	x29, x30, [sp], #80");				\
     85  1.1      matt 	__asm("ret");							\
     86  1.1      matt 	__asm(".size	" MCOUNT_ASM_NAME ", .-" MCOUNT_ASM_NAME);
     87  1.1      matt 
     88  1.3       ryo #ifdef _KERNEL
     89  1.3       ryo #define MCOUNT_ENTER	\
     90  1.3       ryo 	__asm __volatile ("mrs %x0, daif; msr daifset, #3": "=r"(s):: "memory")
     91  1.3       ryo #define MCOUNT_EXIT	\
     92  1.3       ryo 	__asm __volatile ("msr daif, %x0":: "r"(s): "memory")
     93  1.3       ryo #endif /* _KERNEL */
     94  1.3       ryo 
     95  1.1      matt #elif defined(__arm__)
     96  1.1      matt 
     97  1.1      matt #include <arm/profile.h>
     98  1.1      matt 
     99  1.1      matt #endif
    100  1.1      matt 
    101