Home | History | Annotate | Line # | Download | only in include
frameasm.h revision 1.50
      1 /*	$NetBSD: frameasm.h,v 1.50 2020/06/01 22:58:06 ad Exp $	*/
      2 
      3 #ifndef _AMD64_MACHINE_FRAMEASM_H
      4 #define _AMD64_MACHINE_FRAMEASM_H
      5 
      6 #ifdef _KERNEL_OPT
      7 #include "opt_xen.h"
      8 #include "opt_svs.h"
      9 #include "opt_kcov.h"
     10 #include "opt_kmsan.h"
     11 #endif
     12 
     13 /*
     14  * Macros to define pushing/popping frames for interrupts, traps
     15  * and system calls. Currently all the same; will diverge later.
     16  */
     17 
     18 #ifdef XENPV
     19 #define HYPERVISOR_iret hypercall_page + (__HYPERVISOR_iret * 32)
     20 /* Xen do not need swapgs, done by hypervisor */
     21 #define swapgs
     22 #define iretq	pushq $0 ; jmp HYPERVISOR_iret
     23 #define	XEN_ONLY2(x,y)	x,y
     24 #define	NOT_XEN(x)
     25 
     26 #define CLI(temp_reg) \
     27  	movq CPUVAR(VCPU),%r ## temp_reg ;			\
     28 	movb $1,EVTCHN_UPCALL_MASK(%r ## temp_reg);
     29 
     30 #define STI(temp_reg) \
     31  	movq CPUVAR(VCPU),%r ## temp_reg ;			\
     32 	movb $0,EVTCHN_UPCALL_MASK(%r ## temp_reg);
     33 
     34 #define PUSHF(temp_reg) \
     35  	movq CPUVAR(VCPU),%r ## temp_reg ;			\
     36 	movzbl EVTCHN_UPCALL_MASK(%r ## temp_reg), %e ## temp_reg; \
     37 	pushq %r ## temp_reg
     38 
     39 #define POPF \
     40 	popq %rdi; \
     41 	call _C_LABEL(xen_write_psl)
     42 
     43 
     44 #else /* XENPV */
     45 #define	XEN_ONLY2(x,y)
     46 #define	NOT_XEN(x)	x
     47 #define CLI(temp_reg) cli
     48 #define STI(temp_reg) sti
     49 #define PUSHF(temp_reg) pushf
     50 #define POPL popl
     51 #endif	/* XEN */
     52 
     53 #define HP_NAME_CLAC		1
     54 #define HP_NAME_STAC		2
     55 #define HP_NAME_NOLOCK		3
     56 #define HP_NAME_RETFENCE	4
     57 #define HP_NAME_SVS_ENTER	5
     58 #define HP_NAME_SVS_LEAVE	6
     59 #define HP_NAME_SVS_ENTER_ALT	7
     60 #define HP_NAME_SVS_LEAVE_ALT	8
     61 #define HP_NAME_IBRS_ENTER	9
     62 #define HP_NAME_IBRS_LEAVE	10
     63 #define HP_NAME_SVS_ENTER_NMI	11
     64 #define HP_NAME_SVS_LEAVE_NMI	12
     65 #define HP_NAME_MDS_LEAVE	13
     66 #define HP_NAME_SSE2_LFENCE	14
     67 #define HP_NAME_SSE2_MFENCE	15
     68 
     69 #define HOTPATCH(name, size) \
     70 123:						; \
     71 	.pushsection	.rodata.hotpatch, "a"	; \
     72 	.byte		name			; \
     73 	.byte		size			; \
     74 	.quad		123b			; \
     75 	.popsection
     76 
     77 #define SMAP_ENABLE \
     78 	HOTPATCH(HP_NAME_CLAC, 3)		; \
     79 	.byte 0x0F, 0x1F, 0x00			; \
     80 
     81 #define SMAP_DISABLE \
     82 	HOTPATCH(HP_NAME_STAC, 3)		; \
     83 	.byte 0x0F, 0x1F, 0x00			; \
     84 
     85 /*
     86  * IBRS
     87  */
     88 
     89 #define IBRS_ENTER_BYTES	12
     90 #define IBRS_ENTER \
     91 	HOTPATCH(HP_NAME_IBRS_ENTER, IBRS_ENTER_BYTES)		; \
     92 	NOIBRS_ENTER
     93 #define NOIBRS_ENTER \
     94 	.byte 0xEB, (IBRS_ENTER_BYTES-2)	/* jmp */	; \
     95 	.fill	(IBRS_ENTER_BYTES-2),1,0xCC
     96 
     97 #define IBRS_LEAVE_BYTES	12
     98 #define IBRS_LEAVE \
     99 	HOTPATCH(HP_NAME_IBRS_LEAVE, IBRS_LEAVE_BYTES)		; \
    100 	NOIBRS_LEAVE
    101 #define NOIBRS_LEAVE \
    102 	.byte 0xEB, (IBRS_LEAVE_BYTES-2)	/* jmp */	; \
    103 	.fill	(IBRS_LEAVE_BYTES-2),1,0xCC
    104 
    105 /*
    106  * MDS
    107  */
    108 
    109 #define MDS_LEAVE_BYTES	10
    110 #define MDS_LEAVE \
    111 	HOTPATCH(HP_NAME_MDS_LEAVE, MDS_LEAVE_BYTES)		; \
    112 	NOMDS_LEAVE
    113 #define NOMDS_LEAVE \
    114 	.byte 0xEB, (MDS_LEAVE_BYTES-2)	/* jmp */		; \
    115 	.fill	(MDS_LEAVE_BYTES-2),1,0xCC
    116 
    117 #define	SWAPGS	NOT_XEN(swapgs)
    118 
    119 /*
    120  * These are used on interrupt or trap entry or exit.
    121  */
    122 #define INTR_SAVE_GPRS \
    123 	movq	%rdi,TF_RDI(%rsp)	; \
    124 	movq	%rsi,TF_RSI(%rsp)	; \
    125 	movq	%rdx,TF_RDX(%rsp)	; \
    126 	movq	%rcx,TF_RCX(%rsp)	; \
    127 	movq	%r8,TF_R8(%rsp)		; \
    128 	movq	%r9,TF_R9(%rsp)		; \
    129 	movq	%r10,TF_R10(%rsp)	; \
    130 	movq	%r11,TF_R11(%rsp)	; \
    131 	movq	%r12,TF_R12(%rsp)	; \
    132 	movq	%r13,TF_R13(%rsp)	; \
    133 	movq	%r14,TF_R14(%rsp)	; \
    134 	movq	%r15,TF_R15(%rsp)	; \
    135 	movq	%rbp,TF_RBP(%rsp)	; \
    136 	movq	%rbx,TF_RBX(%rsp)	; \
    137 	movq	%rax,TF_RAX(%rsp)
    138 
    139 #define	INTR_RESTORE_GPRS \
    140 	movq	TF_RDI(%rsp),%rdi	; \
    141 	movq	TF_RSI(%rsp),%rsi	; \
    142 	movq	TF_RDX(%rsp),%rdx	; \
    143 	movq	TF_RCX(%rsp),%rcx	; \
    144 	movq	TF_R8(%rsp),%r8		; \
    145 	movq	TF_R9(%rsp),%r9		; \
    146 	movq	TF_R10(%rsp),%r10	; \
    147 	movq	TF_R11(%rsp),%r11	; \
    148 	movq	TF_R12(%rsp),%r12	; \
    149 	movq	TF_R13(%rsp),%r13	; \
    150 	movq	TF_R14(%rsp),%r14	; \
    151 	movq	TF_R15(%rsp),%r15	; \
    152 	movq	TF_RBP(%rsp),%rbp	; \
    153 	movq	TF_RBX(%rsp),%rbx	; \
    154 	movq	TF_RAX(%rsp),%rax
    155 
    156 #define TEXT_USER_BEGIN	.pushsection	.text.user, "ax"
    157 #define TEXT_USER_END	.popsection
    158 
    159 #ifdef SVS
    160 
    161 /* XXX: put this somewhere else */
    162 #define SVS_UTLS		0xffffff0000000000 /* PMAP_PCPU_BASE */
    163 #define UTLS_KPDIRPA		0
    164 #define UTLS_SCRATCH		8
    165 #define UTLS_RSP0		16
    166 
    167 #define SVS_ENTER_BYTES	22
    168 #define NOSVS_ENTER \
    169 	.byte 0xEB, (SVS_ENTER_BYTES-2)	/* jmp */	; \
    170 	.fill	(SVS_ENTER_BYTES-2),1,0xCC
    171 #define SVS_ENTER \
    172 	HOTPATCH(HP_NAME_SVS_ENTER, SVS_ENTER_BYTES)	; \
    173 	NOSVS_ENTER
    174 
    175 #define SVS_LEAVE_BYTES	21
    176 #define NOSVS_LEAVE \
    177 	.byte 0xEB, (SVS_LEAVE_BYTES-2)	/* jmp */	; \
    178 	.fill	(SVS_LEAVE_BYTES-2),1,0xCC
    179 #define SVS_LEAVE \
    180 	HOTPATCH(HP_NAME_SVS_LEAVE, SVS_LEAVE_BYTES)	; \
    181 	NOSVS_LEAVE
    182 
    183 #define SVS_ENTER_ALT_BYTES	23
    184 #define NOSVS_ENTER_ALTSTACK \
    185 	.byte 0xEB, (SVS_ENTER_ALT_BYTES-2)	/* jmp */	; \
    186 	.fill	(SVS_ENTER_ALT_BYTES-2),1,0xCC
    187 #define SVS_ENTER_ALTSTACK \
    188 	HOTPATCH(HP_NAME_SVS_ENTER_ALT, SVS_ENTER_ALT_BYTES)	; \
    189 	NOSVS_ENTER_ALTSTACK
    190 
    191 #define SVS_LEAVE_ALT_BYTES	22
    192 #define NOSVS_LEAVE_ALTSTACK \
    193 	.byte 0xEB, (SVS_LEAVE_ALT_BYTES-2)	/* jmp */	; \
    194 	.fill	(SVS_LEAVE_ALT_BYTES-2),1,0xCC
    195 #define SVS_LEAVE_ALTSTACK \
    196 	HOTPATCH(HP_NAME_SVS_LEAVE_ALT, SVS_LEAVE_ALT_BYTES)	; \
    197 	NOSVS_LEAVE_ALTSTACK
    198 
    199 #define SVS_ENTER_NMI_BYTES	22
    200 #define NOSVS_ENTER_NMI \
    201 	.byte 0xEB, (SVS_ENTER_NMI_BYTES-2)	/* jmp */	; \
    202 	.fill	(SVS_ENTER_NMI_BYTES-2),1,0xCC
    203 #define SVS_ENTER_NMI \
    204 	HOTPATCH(HP_NAME_SVS_ENTER_NMI, SVS_ENTER_NMI_BYTES)	; \
    205 	NOSVS_ENTER_NMI
    206 
    207 #define SVS_LEAVE_NMI_BYTES	11
    208 #define NOSVS_LEAVE_NMI \
    209 	.byte 0xEB, (SVS_LEAVE_NMI_BYTES-2)	/* jmp */	; \
    210 	.fill	(SVS_LEAVE_NMI_BYTES-2),1,0xCC
    211 #define SVS_LEAVE_NMI \
    212 	HOTPATCH(HP_NAME_SVS_LEAVE_NMI, SVS_LEAVE_NMI_BYTES)	; \
    213 	NOSVS_LEAVE_NMI
    214 
    215 #else
    216 #define SVS_ENTER	/* nothing */
    217 #define SVS_ENTER_NMI	/* nothing */
    218 #define SVS_LEAVE	/* nothing */
    219 #define SVS_LEAVE_NMI	/* nothing */
    220 #define SVS_ENTER_ALTSTACK	/* nothing */
    221 #define SVS_LEAVE_ALTSTACK	/* nothing */
    222 #endif
    223 
    224 #ifdef KMSAN
    225 /* XXX this belongs somewhere else. */
    226 #define KMSAN_ENTER	\
    227 	movq	%rsp,%rdi		; \
    228 	movq	$TF_REGSIZE+16+40,%rsi	; \
    229 	xorq	%rdx,%rdx		; \
    230 	callq	kmsan_mark		; \
    231 	callq	kmsan_intr_enter
    232 #define KMSAN_LEAVE	\
    233 	pushq	%rbp			; \
    234 	movq	%rsp,%rbp		; \
    235 	callq	kmsan_intr_leave	; \
    236 	popq	%rbp
    237 #define KMSAN_INIT_ARG(sz)	\
    238 	pushq	%rax			; \
    239 	pushq	%rcx			; \
    240 	pushq	%rdx			; \
    241 	pushq	%rsi			; \
    242 	pushq	%rdi			; \
    243 	pushq	%r8			; \
    244 	pushq	%r9			; \
    245 	pushq	%r10			; \
    246 	pushq	%r11			; \
    247 	movq	$sz,%rdi		; \
    248 	callq	_C_LABEL(kmsan_init_arg); \
    249 	popq	%r11			; \
    250 	popq	%r10			; \
    251 	popq	%r9			; \
    252 	popq	%r8			; \
    253 	popq	%rdi			; \
    254 	popq	%rsi			; \
    255 	popq	%rdx			; \
    256 	popq	%rcx			; \
    257 	popq	%rax
    258 #define KMSAN_INIT_RET(sz)	\
    259 	pushq	%rax			; \
    260 	pushq	%rcx			; \
    261 	pushq	%rdx			; \
    262 	pushq	%rsi			; \
    263 	pushq	%rdi			; \
    264 	pushq	%r8			; \
    265 	pushq	%r9			; \
    266 	pushq	%r10			; \
    267 	pushq	%r11			; \
    268 	movq	$sz,%rdi		; \
    269 	callq	_C_LABEL(kmsan_init_ret); \
    270 	popq	%r11			; \
    271 	popq	%r10			; \
    272 	popq	%r9			; \
    273 	popq	%r8			; \
    274 	popq	%rdi			; \
    275 	popq	%rsi			; \
    276 	popq	%rdx			; \
    277 	popq	%rcx			; \
    278 	popq	%rax
    279 #define KMSAN_REP_STOS(scale)	\
    280 	pushq	%rax			; \
    281 	pushq	%rcx			; \
    282 	pushq	%rdx			; \
    283 	pushq	%rsi			; \
    284 	pushq	%rdi			; \
    285 	pushq	%r8			; \
    286 	pushq	%r9			; \
    287 	pushq	%r10			; \
    288 	pushq	%r11			; \
    289 	leaq	(,%rcx,scale),%rsi	; \
    290 	callq	_C_LABEL(__msan_instrument_asm_store); \
    291 	popq	%r11			; \
    292 	popq	%r10			; \
    293 	popq	%r9			; \
    294 	popq	%r8			; \
    295 	popq	%rdi			; \
    296 	popq	%rsi			; \
    297 	popq	%rdx			; \
    298 	popq	%rcx			; \
    299 	popq	%rax
    300 #else
    301 #define KMSAN_ENTER		/* nothing */
    302 #define KMSAN_LEAVE		/* nothing */
    303 #define KMSAN_INIT_ARG(sz)	/* nothing */
    304 #define KMSAN_INIT_RET(sz)	/* nothing */
    305 #define	KMSAN_REP_STOS(scale)	/* nothing */
    306 #endif
    307 
    308 #ifdef KCOV
    309 #define KCOV_DISABLE			\
    310 	incl	CPUVAR(IDEPTH)
    311 #define KCOV_ENABLE			\
    312 	decl	CPUVAR(IDEPTH)
    313 #else
    314 #define KCOV_DISABLE		/* nothing */
    315 #define KCOV_ENABLE		/* nothing */
    316 #endif
    317 
    318 #define	INTRENTRY \
    319 	subq	$TF_REGSIZE,%rsp	; \
    320 	INTR_SAVE_GPRS			; \
    321 	cld				; \
    322 	SMAP_ENABLE			; \
    323 	testb	$SEL_UPL,TF_CS(%rsp)	; \
    324 	je	98f			; \
    325 	SWAPGS				; \
    326 	IBRS_ENTER			; \
    327 	SVS_ENTER			; \
    328 	movw	%gs,TF_GS(%rsp)		; \
    329 	movw	%fs,TF_FS(%rsp)		; \
    330 	movw	%es,TF_ES(%rsp)		; \
    331 	movw	%ds,TF_DS(%rsp)		; \
    332 98:	KMSAN_ENTER
    333 
    334 #define INTRFASTEXIT \
    335 	jmp	intrfastexit
    336 
    337 #define INTR_RECURSE_HWFRAME \
    338 	movq	%rsp,%r10		; \
    339 	movl	%ss,%r11d		; \
    340 	pushq	%r11			; \
    341 	pushq	%r10			; \
    342 	pushfq				; \
    343 	pushq	$GSEL(GCODE_SEL,SEL_KPL); \
    344 /* XEN: We must fixup CS, as even kernel mode runs at CPL 3 */ \
    345  	XEN_ONLY2(andb	$0xfc,(%rsp);)	  \
    346 	pushq	%r13			;
    347 
    348 #define INTR_RECURSE_ENTRY \
    349 	subq	$TF_REGSIZE,%rsp	; \
    350 	INTR_SAVE_GPRS			; \
    351 	cld				; \
    352 	KMSAN_ENTER
    353 
    354 #define	CHECK_DEFERRED_SWITCH \
    355 	cmpl	$0, CPUVAR(WANT_PMAPLOAD)
    356 
    357 #define CHECK_ASTPENDING(reg)	cmpl	$0, L_MD_ASTPENDING(reg)
    358 #define CLEAR_ASTPENDING(reg)	movl	$0, L_MD_ASTPENDING(reg)
    359 
    360 /*
    361  * If the FPU state is not in the CPU, restore it. Executed with interrupts
    362  * disabled.
    363  *
    364  *     %r14 is curlwp, must not be modified
    365  *     %rbx must not be modified
    366  */
    367 #define HANDLE_DEFERRED_FPU	\
    368 	testl	$MDL_FPU_IN_CPU,L_MD_FLAGS(%r14)	; \
    369 	jnz	1f					; \
    370 	call	_C_LABEL(fpu_handle_deferred)		; \
    371 	orl	$MDL_FPU_IN_CPU,L_MD_FLAGS(%r14)	; \
    372 1:
    373 
    374 #endif /* _AMD64_MACHINE_FRAMEASM_H */
    375