Home | History | Annotate | Line # | Download | only in i386
morestack.S revision 1.9
      1 # x86/x86_64 support for -fsplit-stack.
      2 # Copyright (C) 2009-2020 Free Software Foundation, Inc.
      3 # Contributed by Ian Lance Taylor <iant@google.com>.
      4 
      5 # This file is part of GCC.
      6 
      7 # GCC is free software; you can redistribute it and/or modify it under
      8 # the terms of the GNU General Public License as published by the Free
      9 # Software Foundation; either version 3, or (at your option) any later
     10 # version.
     11 
     12 # GCC is distributed in the hope that it will be useful, but WITHOUT ANY
     13 # WARRANTY; without even the implied warranty of MERCHANTABILITY or
     14 # FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
     15 # for more details.
     16 
     17 # Under Section 7 of GPL version 3, you are granted additional
     18 # permissions described in the GCC Runtime Library Exception, version
     19 # 3.1, as published by the Free Software Foundation.
     20 
     21 # You should have received a copy of the GNU General Public License and
     22 # a copy of the GCC Runtime Library Exception along with this program;
     23 # see the files COPYING3 and COPYING.RUNTIME respectively.  If not, see
     24 # <http://www.gnu.org/licenses/>.
     25 
     26 
     27 # Support for allocating more stack space when using -fsplit-stack.
     28 # When a function discovers that it needs more stack space, it will
     29 # call __morestack with the size of the stack frame and the size of
     30 # the parameters to copy from the old stack frame to the new one.
     31 # The __morestack function preserves the parameter registers and
     32 # calls __generic_morestack to actually allocate the stack space.
     33 
     34 # When this is called stack space is very low, but we ensure that
     35 # there is enough space to push the parameter registers and to call
     36 # __generic_morestack.
     37 
     38 # When calling __generic_morestack, FRAME_SIZE points to the size of
     39 # the desired frame when the function is called, and the function
     40 # sets it to the size of the allocated stack.  OLD_STACK points to
     41 # the parameters on the old stack and PARAM_SIZE is the number of
     42 # bytes of parameters to copy to the new stack.  These are the
     43 # parameters of the function that called __morestack.  The
     44 # __generic_morestack function returns the new stack pointer,
     45 # pointing to the address of the first copied parameter.  The return
     46 # value minus the returned *FRAME_SIZE will be the first address on
     47 # the stack which we should not use.
     48 
     49 # void *__generic_morestack (size_t *frame_size, void *old_stack,
     50 #			     size_t param_size);
     51 
     52 # The __morestack routine has to arrange for the caller to return to a
     53 # stub on the new stack.  The stub is responsible for restoring the
     54 # old stack pointer and returning to the caller's caller.  This calls
     55 # __generic_releasestack to retrieve the old stack pointer and release
     56 # the newly allocated stack.
     57 
     58 # void *__generic_releasestack (size_t *available);
     59 
     60 # We do a little dance so that the processor's call/return return
     61 # address prediction works out.  The compiler arranges for the caller
     62 # to look like this:
     63 #   call __generic_morestack
     64 #   ret
     65 #  L:
     66 #   // carry on with function
     67 # After we allocate more stack, we call L, which is in our caller.
     68 # When that returns (to the predicted instruction), we release the
     69 # stack segment and reset the stack pointer.  We then return to the
     70 # predicted instruction, namely the ret instruction immediately after
     71 # the call to __generic_morestack.  That then returns to the caller of
     72 # the original caller.
     73 
     74 
     75 # The amount of extra space we ask for.  In general this has to be
     76 # enough for the dynamic loader to find a symbol and for a signal
     77 # handler to run.
     78 
     79 #ifndef __x86_64__
     80 #define BACKOFF (1024)
     81 #else
     82 #define BACKOFF (1536)
     83 #endif
     84 
     85 
     86 # The amount of space we ask for when calling non-split-stack code.
     87 #define NON_SPLIT_STACK 0x100000
     88 
     89 # This entry point is for split-stack code which calls non-split-stack
     90 # code.  When the linker sees this case, it converts the call to
     91 # __morestack to call __morestack_non_split instead.  We just bump the
     92 # requested stack space by 16K.
     93 
     94 #include <cet.h>
     95 
     96 	.global __morestack_non_split
     97 	.hidden	__morestack_non_split
     98 
     99 #ifdef __ELF__
    100        .type	__morestack_non_split,@function
    101 #endif
    102 
    103 __morestack_non_split:
    104 	.cfi_startproc
    105 
    106 #ifndef __x86_64__
    107 
    108 	# See below for an extended explanation of this.
    109 	.cfi_def_cfa %esp,16
    110 
    111 	pushl	%eax			# Save %eax in case it is a parameter.
    112 
    113 	.cfi_adjust_cfa_offset 4	# Account for pushed register.
    114 
    115 	movl	%esp,%eax		# Current stack,
    116 	subl	8(%esp),%eax		# less required stack frame size,
    117 	subl	$NON_SPLIT_STACK,%eax	# less space for non-split code.
    118 	cmpl	%gs:0x30,%eax		# See if we have enough space.
    119 	jb	2f			# Get more space if we need it.
    120 
    121 	# Here the stack is
    122 	#	%esp + 20:	stack pointer after two returns
    123 	#	%esp + 16:	return address of morestack caller's caller
    124 	#	%esp + 12:	size of parameters
    125 	#	%esp + 8:	new stack frame size
    126 	#	%esp + 4:	return address of this function
    127 	#	%esp:		saved %eax
    128 	#
    129 	# Since we aren't doing a full split stack, we don't need to
    130 	# do anything when our caller returns.  So we return to our
    131 	# caller rather than calling it, and let it return as usual.
    132 	# To make that work we adjust the return address.
    133 
    134 	# This breaks call/return address prediction for the call to
    135 	# this function.  I can't figure out a way to make it work
    136 	# short of copying the parameters down the stack, which will
    137 	# probably take more clock cycles than we will lose breaking
    138 	# call/return address prediction.  We will only break
    139 	# prediction for this call, not for our caller.
    140 
    141 	movl	4(%esp),%eax		# Increment the return address
    142 	cmpb	$0xc3,(%eax)		# to skip the ret instruction;
    143 	je	1f			# see above.
    144 	addl	$2,%eax
    145 1:	inc	%eax
    146 
    147 	# If the instruction that we return to is
    148 	#   leal  20(%ebp),{%eax,%ecx,%edx}
    149 	# then we have been called by a varargs function that expects
    150 	# %ebp to hold a real value.  That can only work if we do the
    151 	# full stack split routine.  FIXME: This is fragile.
    152 	cmpb	$0x8d,(%eax)
    153 	jne	3f
    154 	cmpb	$0x14,2(%eax)
    155 	jne	3f
    156 	cmpb	$0x45,1(%eax)
    157 	je	2f
    158 	cmpb	$0x4d,1(%eax)
    159 	je	2f
    160 	cmpb	$0x55,1(%eax)
    161 	je	2f
    162 
    163 3:
    164 	movl	%eax,4(%esp)		# Update return address.
    165 
    166 	popl	%eax			# Restore %eax and stack.
    167 
    168 	.cfi_adjust_cfa_offset -4	# Account for popped register.
    169 
    170 	ret	$8			# Return to caller, popping args.
    171 
    172 2:
    173 	.cfi_adjust_cfa_offset 4	# Back to where we were.
    174 
    175 	popl	%eax			# Restore %eax and stack.
    176 
    177 	.cfi_adjust_cfa_offset -4	# Account for popped register.
    178 
    179 	# Increment space we request.
    180 	addl	$NON_SPLIT_STACK+0x1000+BACKOFF,4(%esp)
    181 
    182 	# Fall through into morestack.
    183 
    184 #else
    185 
    186 	# See below for an extended explanation of this.
    187 	.cfi_def_cfa %rsp,16
    188 
    189 	pushq	%rax			# Save %rax in case caller is using
    190 					# it to preserve original %r10.
    191 	.cfi_adjust_cfa_offset 8	# Adjust for pushed register.
    192 
    193 	movq	%rsp,%rax		# Current stack,
    194 	subq	%r10,%rax		# less required stack frame size,
    195 	subq	$NON_SPLIT_STACK,%rax	# less space for non-split code.
    196 
    197 #ifdef __LP64__
    198 	cmpq	%fs:0x70,%rax		# See if we have enough space.
    199 #else
    200 	cmpl	%fs:0x40,%eax
    201 #endif
    202 
    203 	jb	2f			# Get more space if we need it.
    204 
    205 	# If the instruction that we return to is
    206 	#   leaq  24(%rbp), %r11n
    207 	# then we have been called by a varargs function that expects
    208 	# %ebp to hold a real value.  That can only work if we do the
    209 	# full stack split routine.  FIXME: This is fragile.
    210 	movq	8(%rsp),%rax
    211 	incq	%rax			# Skip ret instruction in caller.
    212 	cmpl	$0x185d8d4c,(%rax)
    213 	je	2f
    214 
    215 	# This breaks call/return prediction, as described above.
    216 	incq	8(%rsp)			# Increment the return address.
    217 
    218 	popq	%rax			# Restore register.
    219 
    220 	.cfi_adjust_cfa_offset -8	# Adjust for popped register.
    221 
    222 	ret				# Return to caller.
    223 
    224 2:
    225 	popq	%rax			# Restore register.
    226 
    227 	.cfi_adjust_cfa_offset -8	# Adjust for popped register.
    228 
    229 	# Increment space we request.
    230 	addq	$NON_SPLIT_STACK+0x1000+BACKOFF,%r10
    231 
    232 	# Fall through into morestack.
    233 
    234 #endif
    235 
    236 	.cfi_endproc
    237 #ifdef __ELF__
    238 	.size	__morestack_non_split, . - __morestack_non_split
    239 #endif
    240 
    241 # __morestack_non_split falls through into __morestack.
    242 
    243 
    244 # The __morestack function.
    245 
    246 	.global	__morestack
    247 	.hidden	__morestack
    248 
    249 #ifdef __ELF__
    250 	.type	__morestack,@function
    251 #endif
    252 
    253 __morestack:
    254 .LFB1:
    255 	.cfi_startproc
    256 
    257 
    258 #ifndef __x86_64__
    259 
    260 
    261 # The 32-bit __morestack function.
    262 
    263 	# We use a cleanup to restore the stack guard if an exception
    264 	# is thrown through this code.
    265 #ifndef __PIC__
    266 	.cfi_personality 0,__gcc_personality_v0
    267 	.cfi_lsda 0,.LLSDA1
    268 #else
    269 	.cfi_personality 0x9b,DW.ref.__gcc_personality_v0
    270 	.cfi_lsda 0x1b,.LLSDA1
    271 #endif
    272 
    273 	# We return below with a ret $8.  We will return to a single
    274 	# return instruction, which will return to the caller of our
    275 	# caller.  We let the unwinder skip that single return
    276 	# instruction, and just return to the real caller.
    277 
    278 	# Here CFA points just past the return address on the stack,
    279 	# e.g., on function entry it is %esp + 4.  The stack looks
    280 	# like this:
    281 	#	CFA + 12:	stack pointer after two returns
    282 	#	CFA + 8:	return address of morestack caller's caller
    283 	#	CFA + 4:	size of parameters
    284 	#	CFA:		new stack frame size
    285 	#	CFA - 4:	return address of this function
    286 	#	CFA - 8:	previous value of %ebp; %ebp points here
    287 	# Setting the new CFA to be the current CFA + 12 (i.e., %esp +
    288 	# 16) will make the unwinder pick up the right return address.
    289 
    290 	.cfi_def_cfa %esp,16
    291 
    292 	pushl	%ebp
    293 	.cfi_adjust_cfa_offset 4
    294 	.cfi_offset %ebp, -20
    295 	movl	%esp,%ebp
    296 	.cfi_def_cfa_register %ebp
    297 
    298 	# In 32-bit mode the parameters are pushed on the stack.  The
    299 	# argument size is pushed then the new stack frame size is
    300 	# pushed.
    301 
    302 	# In the body of a non-leaf function, the stack pointer will
    303 	# be aligned to a 16-byte boundary.  That is CFA + 12 in the
    304 	# stack picture above: (CFA + 12) % 16 == 0.  At this point we
    305 	# have %esp == CFA - 8, so %esp % 16 == 12.  We need some
    306 	# space for saving registers and passing parameters, and we
    307 	# need to wind up with %esp % 16 == 0.
    308 	subl	$44,%esp
    309 
    310 	# Because our cleanup code may need to clobber %ebx, we need
    311 	# to save it here so the unwinder can restore the value used
    312 	# by the caller.  Note that we don't have to restore the
    313 	# register, since we don't change it, we just have to save it
    314 	# for the unwinder.
    315 	movl	%ebx,-4(%ebp)
    316 	.cfi_offset %ebx, -24
    317 
    318 	# In 32-bit mode the registers %eax, %edx, and %ecx may be
    319 	# used for parameters, depending on the regparm and fastcall
    320 	# attributes.
    321 
    322 	movl	%eax,-8(%ebp)
    323 	movl	%edx,-12(%ebp)
    324 	movl	%ecx,-16(%ebp)
    325 
    326 	call	__morestack_block_signals
    327 
    328 	movl	12(%ebp),%eax		# The size of the parameters.
    329 	movl	%eax,8(%esp)
    330 	leal	20(%ebp),%eax		# Address of caller's parameters.
    331 	movl	%eax,4(%esp)
    332 	addl	$BACKOFF,8(%ebp)	# Ask for backoff bytes.
    333 	leal	8(%ebp),%eax		# The address of the new frame size.
    334 	movl	%eax,(%esp)
    335 
    336 	call	__generic_morestack
    337 
    338 	movl	%eax,%esp		# Switch to the new stack.
    339 	subl	8(%ebp),%eax		# The end of the stack space.
    340 	addl	$BACKOFF,%eax		# Back off 512 bytes.
    341 
    342 .LEHB0:
    343 	# FIXME: The offset must match
    344 	# TARGET_THREAD_SPLIT_STACK_OFFSET in
    345 	# gcc/config/i386/linux.h.
    346 	movl	%eax,%gs:0x30		# Save the new stack boundary.
    347 
    348 	call	__morestack_unblock_signals
    349 
    350 	movl	-12(%ebp),%edx		# Restore registers.
    351 	movl	-16(%ebp),%ecx
    352 
    353 	movl	4(%ebp),%eax		# Increment the return address
    354 	cmpb	$0xc3,(%eax)		# to skip the ret instruction;
    355 	je	1f			# see above.
    356 	addl	$2,%eax
    357 1:	inc	%eax
    358 
    359 	movl	%eax,-12(%ebp)		# Store return address in an
    360 					# unused slot.
    361 
    362 	movl	-8(%ebp),%eax		# Restore the last register.
    363 
    364 	call	*-12(%ebp)		# Call our caller!
    365 
    366 	# The caller will return here, as predicted.
    367 
    368 	# Save the registers which may hold a return value.  We
    369 	# assume that __generic_releasestack does not touch any
    370 	# floating point or vector registers.
    371 	pushl	%eax
    372 	pushl	%edx
    373 
    374 	# Push the arguments to __generic_releasestack now so that the
    375 	# stack is at a 16-byte boundary for
    376 	# __morestack_block_signals.
    377 	pushl	$0			# Where the available space is returned.
    378 	leal	0(%esp),%eax		# Push its address.
    379 	push	%eax
    380 
    381 	call	__morestack_block_signals
    382 
    383 	call	__generic_releasestack
    384 
    385 	subl	4(%esp),%eax		# Subtract available space.
    386 	addl	$BACKOFF,%eax		# Back off 512 bytes.
    387 .LEHE0:
    388 	movl	%eax,%gs:0x30		# Save the new stack boundary.
    389 
    390 	addl	$8,%esp			# Remove values from stack.
    391 
    392 	# We need to restore the old stack pointer, which is in %rbp,
    393 	# before we unblock signals.  We also need to restore %eax and
    394 	# %edx after we unblock signals but before we return.  Do this
    395 	# by moving %eax and %edx from the current stack to the old
    396 	# stack.
    397 
    398 	popl	%edx			# Pop return value from current stack.
    399 	popl	%eax
    400 
    401 	movl	%ebp,%esp		# Restore stack pointer.
    402 
    403 	# As before, we now have %esp % 16 == 12.
    404 
    405 	pushl	%eax			# Push return value on old stack.
    406 	pushl	%edx
    407 	subl	$4,%esp			# Align stack to 16-byte boundary.
    408 
    409 	call	__morestack_unblock_signals
    410 
    411 	addl	$4,%esp
    412 	popl	%edx			# Restore return value.
    413 	popl	%eax
    414 
    415 	.cfi_remember_state
    416 
    417 	# We never changed %ebx, so we don't have to actually restore it.
    418 	.cfi_restore %ebx
    419 
    420 	popl	%ebp
    421 	.cfi_restore %ebp
    422 	.cfi_def_cfa %esp, 16
    423 	ret	$8			# Return to caller, which will
    424 					# immediately return.  Pop
    425 					# arguments as we go.
    426 
    427 # This is the cleanup code called by the stack unwinder when unwinding
    428 # through the code between .LEHB0 and .LEHE0 above.
    429 
    430 .L1:
    431 	.cfi_restore_state
    432 	subl	$16,%esp		# Maintain 16 byte alignment.
    433 	movl	%eax,4(%esp)		# Save exception header.
    434 	movl	%ebp,(%esp)		# Stack pointer after resume.
    435 	call	__generic_findstack
    436 	movl	%ebp,%ecx		# Get the stack pointer.
    437 	subl	%eax,%ecx		# Subtract available space.
    438 	addl	$BACKOFF,%ecx		# Back off 512 bytes.
    439 	movl	%ecx,%gs:0x30		# Save new stack boundary.
    440 	movl	4(%esp),%eax		# Function argument.
    441 	movl	%eax,(%esp)
    442 #ifdef __PIC__
    443 	call	__x86.get_pc_thunk.bx	# %ebx may not be set up for us.
    444 	addl	$_GLOBAL_OFFSET_TABLE_, %ebx
    445 	call	_Unwind_Resume@PLT	# Resume unwinding.
    446 #else
    447 	call	_Unwind_Resume
    448 #endif
    449 
    450 #else /* defined(__x86_64__) */
    451 
    452 
    453 # The 64-bit __morestack function.
    454 
    455 	# We use a cleanup to restore the stack guard if an exception
    456 	# is thrown through this code.
    457 #ifndef __PIC__
    458 	.cfi_personality 0x3,__gcc_personality_v0
    459 	.cfi_lsda 0x3,.LLSDA1
    460 #else
    461 	.cfi_personality 0x9b,DW.ref.__gcc_personality_v0
    462 	.cfi_lsda 0x1b,.LLSDA1
    463 #endif
    464 
    465 	# We will return a single return instruction, which will
    466 	# return to the caller of our caller.  Let the unwinder skip
    467 	# that single return instruction, and just return to the real
    468 	# caller.
    469 	.cfi_def_cfa %rsp,16
    470 
    471 	# Set up a normal backtrace.
    472 	pushq	%rbp
    473 	.cfi_adjust_cfa_offset 8
    474 	.cfi_offset %rbp, -24
    475 	movq	%rsp, %rbp
    476 	.cfi_def_cfa_register %rbp
    477 
    478 	# In 64-bit mode the new stack frame size is passed in r10
    479         # and the argument size is passed in r11.
    480 
    481 	addq	$BACKOFF,%r10		# Ask for backoff bytes.
    482 	pushq	%r10			# Save new frame size.
    483 
    484 	# In 64-bit mode the registers %rdi, %rsi, %rdx, %rcx, %r8,
    485 	# and %r9 may be used for parameters.  We also preserve %rax
    486 	# which the caller may use to hold %r10.
    487 
    488 	pushq	%rax
    489 	pushq	%rdi
    490 	pushq	%rsi
    491 	pushq	%rdx
    492 	pushq	%rcx
    493 	pushq	%r8
    494 	pushq	%r9
    495 
    496 	pushq	%r11
    497 
    498 	# We entered morestack with the stack pointer aligned to a
    499 	# 16-byte boundary (the call to morestack's caller used 8
    500 	# bytes, and the call to morestack used 8 bytes).  We have now
    501 	# pushed 10 registers, so we are still aligned to a 16-byte
    502 	# boundary.
    503 
    504 	call	__morestack_block_signals
    505 
    506 	leaq	-8(%rbp),%rdi		# Address of new frame size.
    507 	leaq	24(%rbp),%rsi		# The caller's parameters.
    508 	popq	%rdx			# The size of the parameters.
    509 
    510 	subq	$8,%rsp			# Align stack.
    511 
    512 	call	__generic_morestack
    513 
    514 	movq	-8(%rbp),%r10		# Reload modified frame size
    515 	movq	%rax,%rsp		# Switch to the new stack.
    516 	subq	%r10,%rax		# The end of the stack space.
    517 	addq	$BACKOFF,%rax		# Back off 1024 bytes.
    518 
    519 .LEHB0:
    520 	# FIXME: The offset must match
    521 	# TARGET_THREAD_SPLIT_STACK_OFFSET in
    522 	# gcc/config/i386/linux64.h.
    523 	# Macro to save the new stack boundary.
    524 #ifdef __LP64__
    525 #define X86_64_SAVE_NEW_STACK_BOUNDARY(reg)	movq	%r##reg,%fs:0x70
    526 #else
    527 #define X86_64_SAVE_NEW_STACK_BOUNDARY(reg)	movl	%e##reg,%fs:0x40
    528 #endif
    529 	X86_64_SAVE_NEW_STACK_BOUNDARY (ax)
    530 
    531 	call	__morestack_unblock_signals
    532 
    533 	movq	-24(%rbp),%rdi		# Restore registers.
    534 	movq	-32(%rbp),%rsi
    535 	movq	-40(%rbp),%rdx
    536 	movq	-48(%rbp),%rcx
    537 	movq	-56(%rbp),%r8
    538 	movq	-64(%rbp),%r9
    539 
    540 	movq	8(%rbp),%r10		# Increment the return address
    541 	incq	%r10			# to skip the ret instruction;
    542 					# see above.
    543 
    544 	movq	-16(%rbp),%rax		# Restore caller's %rax.
    545 
    546 	call	*%r10			# Call our caller!
    547 
    548 	# The caller will return here, as predicted.
    549 
    550 	# Save the registers which may hold a return value.  We
    551 	# assume that __generic_releasestack does not touch any
    552 	# floating point or vector registers.
    553 	pushq	%rax
    554 	pushq	%rdx
    555 
    556 	call	__morestack_block_signals
    557 
    558 	pushq	$0			# For alignment.
    559 	pushq	$0			# Where the available space is returned.
    560 	leaq	0(%rsp),%rdi		# Pass its address.
    561 
    562 	call	__generic_releasestack
    563 
    564 	subq	0(%rsp),%rax		# Subtract available space.
    565 	addq	$BACKOFF,%rax		# Back off 1024 bytes.
    566 .LEHE0:
    567 	X86_64_SAVE_NEW_STACK_BOUNDARY (ax)
    568 
    569 	addq	$16,%rsp		# Remove values from stack.
    570 
    571 	# We need to restore the old stack pointer, which is in %rbp,
    572 	# before we unblock signals.  We also need to restore %rax and
    573 	# %rdx after we unblock signals but before we return.  Do this
    574 	# by moving %rax and %rdx from the current stack to the old
    575 	# stack.
    576 
    577 	popq	%rdx			# Pop return value from current stack.
    578 	popq	%rax
    579 
    580 	movq	%rbp,%rsp		# Restore stack pointer.
    581 
    582 	# Now (%rsp & 16) == 8.
    583 
    584 	subq	$8,%rsp			# For alignment.
    585 	pushq	%rax			# Push return value on old stack.
    586 	pushq	%rdx
    587 
    588 	call	__morestack_unblock_signals
    589 
    590 	popq	%rdx			# Restore return value.
    591 	popq	%rax
    592 	addq	$8,%rsp
    593 
    594 	.cfi_remember_state
    595 	popq	%rbp
    596 	.cfi_restore %rbp
    597 	.cfi_def_cfa %rsp, 16
    598 	ret				# Return to caller, which will
    599 					# immediately return.
    600 
    601 # This is the cleanup code called by the stack unwinder when unwinding
    602 # through the code between .LEHB0 and .LEHE0 above.
    603 
    604 .L1:
    605 	.cfi_restore_state
    606 	subq	$16,%rsp		# Maintain 16 byte alignment.
    607 	movq	%rax,(%rsp)		# Save exception header.
    608 	movq	%rbp,%rdi		# Stack pointer after resume.
    609 	call	__generic_findstack
    610 	movq	%rbp,%rcx		# Get the stack pointer.
    611 	subq	%rax,%rcx		# Subtract available space.
    612 	addq	$BACKOFF,%rcx		# Back off 1024 bytes.
    613 	X86_64_SAVE_NEW_STACK_BOUNDARY (cx)
    614 	movq	(%rsp),%rdi		# Restore exception data for call.
    615 #ifdef __PIC__
    616 	call	_Unwind_Resume@PLT	# Resume unwinding.
    617 #else
    618 	call	_Unwind_Resume		# Resume unwinding.
    619 #endif
    620 
    621 #endif /* defined(__x86_64__) */
    622 
    623 	.cfi_endproc
    624 #ifdef __ELF__
    625 	.size	__morestack, . - __morestack
    626 #endif
    627 
    628 #if !defined(__x86_64__) && defined(__PIC__)
    629 # Output the thunk to get PC into bx, since we use it above.
    630 	.section	.text.__x86.get_pc_thunk.bx,"axG",@progbits,__x86.get_pc_thunk.bx,comdat
    631 	.globl	__x86.get_pc_thunk.bx
    632 	.hidden	__x86.get_pc_thunk.bx
    633 #ifdef __ELF__
    634 	.type	__x86.get_pc_thunk.bx, @function
    635 #endif
    636 __x86.get_pc_thunk.bx:
    637 	.cfi_startproc
    638 	movl	(%esp), %ebx
    639 	ret
    640 	.cfi_endproc
    641 #ifdef __ELF__
    642 	.size	__x86.get_pc_thunk.bx, . - __x86.get_pc_thunk.bx
    643 #endif
    644 #endif
    645 
    646 # The exception table.  This tells the personality routine to execute
    647 # the exception handler.
    648 
    649 	.section	.gcc_except_table,"a",@progbits
    650 	.align	4
    651 .LLSDA1:
    652 	.byte	0xff	# @LPStart format (omit)
    653 	.byte	0xff	# @TType format (omit)
    654 	.byte	0x1	# call-site format (uleb128)
    655 	.uleb128 .LLSDACSE1-.LLSDACSB1	# Call-site table length
    656 .LLSDACSB1:
    657 	.uleb128 .LEHB0-.LFB1	# region 0 start
    658 	.uleb128 .LEHE0-.LEHB0	# length
    659 	.uleb128 .L1-.LFB1	# landing pad
    660 	.uleb128 0		# action
    661 .LLSDACSE1:
    662 
    663 
    664 	.global __gcc_personality_v0
    665 #ifdef __PIC__
    666 	# Build a position independent reference to the basic
    667         # personality function.
    668 	.hidden DW.ref.__gcc_personality_v0
    669 	.weak   DW.ref.__gcc_personality_v0
    670 	.section .data.DW.ref.__gcc_personality_v0,"awG",@progbits,DW.ref.__gcc_personality_v0,comdat
    671 	.type	DW.ref.__gcc_personality_v0, @object
    672 DW.ref.__gcc_personality_v0:
    673 #ifndef __LP64__
    674 	.align 4
    675 	.size	DW.ref.__gcc_personality_v0, 4
    676 	.long	__gcc_personality_v0
    677 #else
    678 	.align 8
    679 	.size	DW.ref.__gcc_personality_v0, 8
    680 	.quad	__gcc_personality_v0
    681 #endif
    682 #endif
    683 
    684 #if defined __x86_64__ && defined __LP64__
    685 
    686 # This entry point is used for the large model.  With this entry point
    687 # the upper 32 bits of %r10 hold the argument size and the lower 32
    688 # bits hold the new stack frame size.  There doesn't seem to be a way
    689 # to know in the assembler code that we are assembling for the large
    690 # model, and there doesn't seem to be a large model multilib anyhow.
    691 # If one is developed, then the non-PIC code is probably OK since we
    692 # will probably be close to the morestack code, but the PIC code
    693 # almost certainly needs to be changed.  FIXME.
    694 
    695 	.text
    696 	.global	__morestack_large_model
    697 	.hidden	__morestack_large_model
    698 
    699 #ifdef __ELF__
    700 	.type	__morestack_large_model,@function
    701 #endif
    702 
    703 __morestack_large_model:
    704 
    705 	.cfi_startproc
    706 	_CET_ENDBR
    707 
    708 	movq	%r10, %r11
    709 	andl	$0xffffffff, %r10d
    710 	sarq	$32, %r11
    711 	jmp	__morestack
    712 
    713 	.cfi_endproc
    714 #ifdef __ELF__
    715        .size	__morestack_large_model, . - __morestack_large_model
    716 #endif
    717 
    718 #endif /* __x86_64__ && __LP64__ */
    719 
    720 # Initialize the stack test value when the program starts or when a
    721 # new thread starts.  We don't know how large the main stack is, so we
    722 # guess conservatively.  We might be able to use getrlimit here.
    723 
    724 	.text
    725 	.global	__stack_split_initialize
    726 	.hidden	__stack_split_initialize
    727 
    728 #ifdef __ELF__
    729 	.type	__stack_split_initialize, @function
    730 #endif
    731 
    732 __stack_split_initialize:
    733 	_CET_ENDBR
    734 
    735 #ifndef __x86_64__
    736 
    737 	leal	-16000(%esp),%eax	# We should have at least 16K.
    738 	movl	%eax,%gs:0x30
    739 	subl	$4,%esp			# Align stack.
    740 	pushl	$16000
    741 	pushl	%esp
    742 #ifdef __PIC__
    743 	call	__generic_morestack_set_initial_sp@PLT
    744 #else
    745 	call	__generic_morestack_set_initial_sp
    746 #endif
    747 	addl	$12,%esp
    748 	ret
    749 
    750 #else /* defined(__x86_64__) */
    751 
    752 	leaq	-16000(%rsp),%rax	# We should have at least 16K.
    753 	X86_64_SAVE_NEW_STACK_BOUNDARY (ax)
    754 	subq	$8,%rsp			# Align stack.
    755 	movq	%rsp,%rdi
    756 	movq	$16000,%rsi
    757 #ifdef __PIC__
    758 	call	__generic_morestack_set_initial_sp@PLT
    759 #else
    760 	call	__generic_morestack_set_initial_sp
    761 #endif
    762 	addq	$8,%rsp
    763 	ret
    764 
    765 #endif /* defined(__x86_64__) */
    766 
    767 #ifdef __ELF__
    768 	.size	__stack_split_initialize, . - __stack_split_initialize
    769 #endif
    770 
    771 # Routines to get and set the guard, for __splitstack_getcontext,
    772 # __splitstack_setcontext, and __splitstack_makecontext.
    773 
    774 # void *__morestack_get_guard (void) returns the current stack guard.
    775 	.text
    776 	.global	__morestack_get_guard
    777 	.hidden	__morestack_get_guard
    778 
    779 #ifdef __ELF__
    780 	.type	__morestack_get_guard,@function
    781 #endif
    782 
    783 __morestack_get_guard:
    784 
    785 #ifndef __x86_64__
    786 	movl	%gs:0x30,%eax
    787 #else
    788 #ifdef __LP64__
    789 	movq	%fs:0x70,%rax
    790 #else
    791 	movl	%fs:0x40,%eax
    792 #endif
    793 #endif
    794 	ret
    795 
    796 #ifdef __ELF__
    797 	.size	__morestack_get_guard, . - __morestack_get_guard
    798 #endif
    799 
    800 # void __morestack_set_guard (void *) sets the stack guard.
    801 	.global	__morestack_set_guard
    802 	.hidden	__morestack_set_guard
    803 
    804 #ifdef __ELF__
    805 	.type	__morestack_set_guard,@function
    806 #endif
    807 
    808 __morestack_set_guard:
    809 
    810 #ifndef __x86_64__
    811 	movl	4(%esp),%eax
    812 	movl	%eax,%gs:0x30
    813 #else
    814 	X86_64_SAVE_NEW_STACK_BOUNDARY (di)
    815 #endif
    816 	ret
    817 
    818 #ifdef __ELF__
    819 	.size	__morestack_set_guard, . - __morestack_set_guard
    820 #endif
    821 
    822 # void *__morestack_make_guard (void *, size_t) returns the stack
    823 # guard value for a stack.
    824 	.global	__morestack_make_guard
    825 	.hidden	__morestack_make_guard
    826 
    827 #ifdef __ELF__
    828 	.type	__morestack_make_guard,@function
    829 #endif
    830 
    831 __morestack_make_guard:
    832 
    833 #ifndef __x86_64__
    834 	movl	4(%esp),%eax
    835 	subl	8(%esp),%eax
    836 	addl	$BACKOFF,%eax
    837 #else
    838 	subq	%rsi,%rdi
    839 	addq	$BACKOFF,%rdi
    840 	movq	%rdi,%rax
    841 #endif
    842 	ret
    843 
    844 #ifdef __ELF__
    845 	.size	__morestack_make_guard, . - __morestack_make_guard
    846 #endif
    847 
    848 # Make __stack_split_initialize a high priority constructor.  FIXME:
    849 # This is ELF specific.
    850 
    851 	.section	.ctors.65535,"aw",@progbits
    852 
    853 #ifndef __LP64__
    854 	.align	4
    855 	.long	__stack_split_initialize
    856 	.long	__morestack_load_mmap
    857 #else
    858 	.align	8
    859 	.quad	__stack_split_initialize
    860 	.quad	__morestack_load_mmap
    861 #endif
    862 
    863 #if defined(__ELF__) && defined(__linux__)
    864 	.section	.note.GNU-stack,"",@progbits
    865 	.section	.note.GNU-split-stack,"",@progbits
    866 	.section	.note.GNU-no-split-stack,"",@progbits
    867 #endif
    868