Subversion Repositories Kolibri OS

Rev

Details | Last modification | View Log | RSS feed

Rev Author Line No. Line
6515 serge 1
# x86/x86_64 support for -fsplit-stack.
2
# Copyright (C) 2009-2015 Free Software Foundation, Inc.
3
# Contributed by Ian Lance Taylor .
4
 
5
# This file is part of GCC.
6
 
7
# GCC is free software; you can redistribute it and/or modify it under
8
# the terms of the GNU General Public License as published by the Free
9
# Software Foundation; either version 3, or (at your option) any later
10
# version.
11
 
12
# GCC is distributed in the hope that it will be useful, but WITHOUT ANY
13
# WARRANTY; without even the implied warranty of MERCHANTABILITY or
14
# FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
15
# for more details.
16
 
17
# Under Section 7 of GPL version 3, you are granted additional
18
# permissions described in the GCC Runtime Library Exception, version
19
# 3.1, as published by the Free Software Foundation.
20
 
21
# You should have received a copy of the GNU General Public License and
22
# a copy of the GCC Runtime Library Exception along with this program;
23
# see the files COPYING3 and COPYING.RUNTIME respectively.  If not, see
24
# .
25
 
26
 
27
# Support for allocating more stack space when using -fsplit-stack.
28
# When a function discovers that it needs more stack space, it will
29
# call __morestack with the size of the stack frame and the size of
30
# the parameters to copy from the old stack frame to the new one.
31
# The __morestack function preserves the parameter registers and
32
# calls __generic_morestack to actually allocate the stack space.
33
 
34
# When this is called stack space is very low, but we ensure that
35
# there is enough space to push the parameter registers and to call
36
# __generic_morestack.
37
 
38
# When calling __generic_morestack, FRAME_SIZE points to the size of
39
# the desired frame when the function is called, and the function
40
# sets it to the size of the allocated stack.  OLD_STACK points to
41
# the parameters on the old stack and PARAM_SIZE is the number of
42
# bytes of parameters to copy to the new stack.  These are the
43
# parameters of the function that called __morestack.  The
44
# __generic_morestack function returns the new stack pointer,
45
# pointing to the address of the first copied parameter.  The return
46
# value minus the returned *FRAME_SIZE will be the first address on
47
# the stack which we should not use.
48
 
49
# void *__generic_morestack (size_t *frame_size, void *old_stack,
50
#			     size_t param_size);
51
 
52
# The __morestack routine has to arrange for the caller to return to a
53
# stub on the new stack.  The stub is responsible for restoring the
54
# old stack pointer and returning to the caller's caller.  This calls
55
# __generic_releasestack to retrieve the old stack pointer and release
56
# the newly allocated stack.
57
 
58
# void *__generic_releasestack (size_t *available);
59
 
60
# We do a little dance so that the processor's call/return return
61
# address prediction works out.  The compiler arranges for the caller
62
# to look like this:
63
#   call __generic_morestack
64
#   ret
65
#  L:
66
#   // carry on with function
67
# After we allocate more stack, we call L, which is in our caller.
68
# When that returns (to the predicted instruction), we release the
69
# stack segment and reset the stack pointer.  We then return to the
70
# predicted instruction, namely the ret instruction immediately after
71
# the call to __generic_morestack.  That then returns to the caller of
72
# the original caller.
73
 
74
 
75
# The amount of extra space we ask for.  In general this has to be
76
# enough for the dynamic loader to find a symbol and for a signal
77
# handler to run.
78
 
79
#ifndef __x86_64__
80
#define BACKOFF (1024)
81
#else
82
#define BACKOFF (1536)
83
#endif
84
 
85
 
86
# The amount of space we ask for when calling non-split-stack code.
87
#define NON_SPLIT_STACK 0x100000
88
 
89
# This entry point is for split-stack code which calls non-split-stack
90
# code.  When the linker sees this case, it converts the call to
91
# __morestack to call __morestack_non_split instead.  We just bump the
92
# requested stack space by 16K.
93
 
94
	.global __morestack_non_split
95
	.hidden	__morestack_non_split
96
 
97
#ifdef __ELF__
98
       .type	__morestack_non_split,@function
99
#endif
100
 
101
__morestack_non_split:
102
	.cfi_startproc
103
 
104
#ifndef __x86_64__
105
 
106
	# See below for an extended explanation of this.
107
	.cfi_def_cfa %esp,16
108
 
109
	pushl	%eax			# Save %eax in case it is a parameter.
110
 
111
	.cfi_adjust_cfa_offset 4	# Account for pushed register.
112
 
113
	movl	%esp,%eax		# Current stack,
114
	subl	8(%esp),%eax		# less required stack frame size,
115
	subl	$NON_SPLIT_STACK,%eax	# less space for non-split code.
116
	cmpl	%gs:0x30,%eax		# See if we have enough space.
117
	jb	2f			# Get more space if we need it.
118
 
119
	# Here the stack is
120
	#	%esp + 20:	stack pointer after two returns
121
	#	%esp + 16:	return address of morestack caller's caller
122
	#	%esp + 12:	size of parameters
123
	#	%esp + 8:	new stack frame size
124
	#	%esp + 4:	return address of this function
125
	#	%esp:		saved %eax
126
	#
127
	# Since we aren't doing a full split stack, we don't need to
128
	# do anything when our caller returns.  So we return to our
129
	# caller rather than calling it, and let it return as usual.
130
	# To make that work we adjust the return address.
131
 
132
	# This breaks call/return address prediction for the call to
133
	# this function.  I can't figure out a way to make it work
134
	# short of copying the parameters down the stack, which will
135
	# probably take more clock cycles than we will lose breaking
136
	# call/return address prediction.  We will only break
137
	# prediction for this call, not for our caller.
138
 
139
	movl	4(%esp),%eax		# Increment the return address
140
	cmpb	$0xc3,(%eax)		# to skip the ret instruction;
141
	je	1f			# see above.
142
	addl	$2,%eax
143
1:	inc	%eax
144
 
145
	# If the instruction that we return to is
146
	#   leal  20(%ebp),{%eax,%ecx,%edx}
147
	# then we have been called by a varargs function that expects
148
	# %ebp to hold a real value.  That can only work if we do the
149
	# full stack split routine.  FIXME: This is fragile.
150
	cmpb	$0x8d,(%eax)
151
	jne	3f
152
	cmpb	$0x14,2(%eax)
153
	jne	3f
154
	cmpb	$0x45,1(%eax)
155
	je	2f
156
	cmpb	$0x4d,1(%eax)
157
	je	2f
158
	cmpb	$0x55,1(%eax)
159
	je	2f
160
 
161
3:
162
	movl	%eax,4(%esp)		# Update return address.
163
 
164
	popl	%eax			# Restore %eax and stack.
165
 
166
	.cfi_adjust_cfa_offset -4	# Account for popped register.
167
 
168
	ret	$8			# Return to caller, popping args.
169
 
170
2:
171
	.cfi_adjust_cfa_offset 4	# Back to where we were.
172
 
173
	popl	%eax			# Restore %eax and stack.
174
 
175
	.cfi_adjust_cfa_offset -4	# Account for popped register.
176
 
177
	# Increment space we request.
178
	addl	$NON_SPLIT_STACK+0x1000+BACKOFF,4(%esp)
179
 
180
	# Fall through into morestack.
181
 
182
#else
183
 
184
	# See below for an extended explanation of this.
185
	.cfi_def_cfa %rsp,16
186
 
187
	pushq	%rax			# Save %rax in case caller is using
188
					# it to preserve original %r10.
189
	.cfi_adjust_cfa_offset 8	# Adjust for pushed register.
190
 
191
	movq	%rsp,%rax		# Current stack,
192
	subq	%r10,%rax		# less required stack frame size,
193
	subq	$NON_SPLIT_STACK,%rax	# less space for non-split code.
194
 
195
#ifdef __LP64__
196
	cmpq	%fs:0x70,%rax		# See if we have enough space.
197
#else
198
	cmpl	%fs:0x40,%eax
199
#endif
200
 
201
	jb	2f			# Get more space if we need it.
202
 
203
	# If the instruction that we return to is
204
	#   leaq  24(%rbp), %r11n
205
	# then we have been called by a varargs function that expects
206
	# %ebp to hold a real value.  That can only work if we do the
207
	# full stack split routine.  FIXME: This is fragile.
208
	movq	8(%rsp),%rax
209
	incq	%rax			# Skip ret instruction in caller.
210
	cmpl	$0x185d8d4c,(%rax)
211
	je	2f
212
 
213
	# This breaks call/return prediction, as described above.
214
	incq	8(%rsp)			# Increment the return address.
215
 
216
	popq	%rax			# Restore register.
217
 
218
	.cfi_adjust_cfa_offset -8	# Adjust for popped register.
219
 
220
	ret				# Return to caller.
221
 
222
2:
223
	popq	%rax			# Restore register.
224
 
225
	.cfi_adjust_cfa_offset -8	# Adjust for popped register.
226
 
227
	# Increment space we request.
228
	addq	$NON_SPLIT_STACK+0x1000+BACKOFF,%r10
229
 
230
	# Fall through into morestack.
231
 
232
#endif
233
 
234
	.cfi_endproc
235
#ifdef __ELF__
236
	.size	__morestack_non_split, . - __morestack_non_split
237
#endif
238
 
239
# __morestack_non_split falls through into __morestack.
240
 
241
 
242
# The __morestack function.
243
 
244
	.global	__morestack
245
	.hidden	__morestack
246
 
247
#ifdef __ELF__
248
	.type	__morestack,@function
249
#endif
250
 
251
__morestack:
252
.LFB1:
253
	.cfi_startproc
254
 
255
 
256
#ifndef __x86_64__
257
 
258
 
259
# The 32-bit __morestack function.
260
 
261
	# We use a cleanup to restore the stack guard if an exception
262
	# is thrown through this code.
263
#ifndef __PIC__
264
	.cfi_personality 0,__gcc_personality_v0
265
	.cfi_lsda 0,.LLSDA1
266
#else
267
	.cfi_personality 0x9b,DW.ref.__gcc_personality_v0
268
	.cfi_lsda 0x1b,.LLSDA1
269
#endif
270
 
271
	# We return below with a ret $8.  We will return to a single
272
	# return instruction, which will return to the caller of our
273
	# caller.  We let the unwinder skip that single return
274
	# instruction, and just return to the real caller.
275
 
276
	# Here CFA points just past the return address on the stack,
277
	# e.g., on function entry it is %esp + 4.  The stack looks
278
	# like this:
279
	#	CFA + 12:	stack pointer after two returns
280
	#	CFA + 8:	return address of morestack caller's caller
281
	#	CFA + 4:	size of parameters
282
	#	CFA:		new stack frame size
283
	#	CFA - 4:	return address of this function
284
	#	CFA - 8:	previous value of %ebp; %ebp points here
285
	# Setting the new CFA to be the current CFA + 12 (i.e., %esp +
286
	# 16) will make the unwinder pick up the right return address.
287
 
288
	.cfi_def_cfa %esp,16
289
 
290
	pushl	%ebp
291
	.cfi_adjust_cfa_offset 4
292
	.cfi_offset %ebp, -20
293
	movl	%esp,%ebp
294
	.cfi_def_cfa_register %ebp
295
 
296
	# In 32-bit mode the parameters are pushed on the stack.  The
297
	# argument size is pushed then the new stack frame size is
298
	# pushed.
299
 
300
	# In the body of a non-leaf function, the stack pointer will
301
	# be aligned to a 16-byte boundary.  That is CFA + 12 in the
302
	# stack picture above: (CFA + 12) % 16 == 0.  At this point we
303
	# have %esp == CFA - 8, so %esp % 16 == 12.  We need some
304
	# space for saving registers and passing parameters, and we
305
	# need to wind up with %esp % 16 == 0.
306
	subl	$44,%esp
307
 
308
	# Because our cleanup code may need to clobber %ebx, we need
309
	# to save it here so the unwinder can restore the value used
310
	# by the caller.  Note that we don't have to restore the
311
	# register, since we don't change it, we just have to save it
312
	# for the unwinder.
313
	movl	%ebx,-4(%ebp)
314
	.cfi_offset %ebx, -24
315
 
316
	# In 32-bit mode the registers %eax, %edx, and %ecx may be
317
	# used for parameters, depending on the regparm and fastcall
318
	# attributes.
319
 
320
	movl	%eax,-8(%ebp)
321
	movl	%edx,-12(%ebp)
322
	movl	%ecx,-16(%ebp)
323
 
324
	call	__morestack_block_signals
325
 
326
	movl	12(%ebp),%eax		# The size of the parameters.
327
	movl	%eax,8(%esp)
328
	leal	20(%ebp),%eax		# Address of caller's parameters.
329
	movl	%eax,4(%esp)
330
	addl	$BACKOFF,8(%ebp)	# Ask for backoff bytes.
331
	leal	8(%ebp),%eax		# The address of the new frame size.
332
	movl	%eax,(%esp)
333
 
334
	call	__generic_morestack
335
 
336
	movl	%eax,%esp		# Switch to the new stack.
337
	subl	8(%ebp),%eax		# The end of the stack space.
338
	addl	$BACKOFF,%eax		# Back off 512 bytes.
339
 
340
.LEHB0:
341
	# FIXME: The offset must match
342
	# TARGET_THREAD_SPLIT_STACK_OFFSET in
343
	# gcc/config/i386/linux.h.
344
	movl	%eax,%gs:0x30		# Save the new stack boundary.
345
 
346
	call	__morestack_unblock_signals
347
 
348
	movl	-12(%ebp),%edx		# Restore registers.
349
	movl	-16(%ebp),%ecx
350
 
351
	movl	4(%ebp),%eax		# Increment the return address
352
	cmpb	$0xc3,(%eax)		# to skip the ret instruction;
353
	je	1f			# see above.
354
	addl	$2,%eax
355
1:	inc	%eax
356
 
357
	movl	%eax,-12(%ebp)		# Store return address in an
358
					# unused slot.
359
 
360
	movl	-8(%ebp),%eax		# Restore the last register.
361
 
362
	call	*-12(%ebp)		# Call our caller!
363
 
364
	# The caller will return here, as predicted.
365
 
366
	# Save the registers which may hold a return value.  We
367
	# assume that __generic_releasestack does not touch any
368
	# floating point or vector registers.
369
	pushl	%eax
370
	pushl	%edx
371
 
372
	# Push the arguments to __generic_releasestack now so that the
373
	# stack is at a 16-byte boundary for
374
	# __morestack_block_signals.
375
	pushl	$0			# Where the available space is returned.
376
	leal	0(%esp),%eax		# Push its address.
377
	push	%eax
378
 
379
	call	__morestack_block_signals
380
 
381
	call	__generic_releasestack
382
 
383
	subl	4(%esp),%eax		# Subtract available space.
384
	addl	$BACKOFF,%eax		# Back off 512 bytes.
385
.LEHE0:
386
	movl	%eax,%gs:0x30		# Save the new stack boundary.
387
 
388
	addl	$8,%esp			# Remove values from stack.
389
 
390
	# We need to restore the old stack pointer, which is in %rbp,
391
	# before we unblock signals.  We also need to restore %eax and
392
	# %edx after we unblock signals but before we return.  Do this
393
	# by moving %eax and %edx from the current stack to the old
394
	# stack.
395
 
396
	popl	%edx			# Pop return value from current stack.
397
	popl	%eax
398
 
399
	movl	%ebp,%esp		# Restore stack pointer.
400
 
401
	# As before, we now have %esp % 16 == 12.
402
 
403
	pushl	%eax			# Push return value on old stack.
404
	pushl	%edx
405
	subl	$4,%esp			# Align stack to 16-byte boundary.
406
 
407
	call	__morestack_unblock_signals
408
 
409
	addl	$4,%esp
410
	popl	%edx			# Restore return value.
411
	popl	%eax
412
 
413
	.cfi_remember_state
414
 
415
	# We never changed %ebx, so we don't have to actually restore it.
416
	.cfi_restore %ebx
417
 
418
	popl	%ebp
419
	.cfi_restore %ebp
420
	.cfi_def_cfa %esp, 16
421
	ret	$8			# Return to caller, which will
422
					# immediately return.  Pop
423
					# arguments as we go.
424
 
425
# This is the cleanup code called by the stack unwinder when unwinding
426
# through the code between .LEHB0 and .LEHE0 above.
427
 
428
.L1:
429
	.cfi_restore_state
430
	subl	$16,%esp		# Maintain 16 byte alignment.
431
	movl	%eax,4(%esp)		# Save exception header.
432
	movl	%ebp,(%esp)		# Stack pointer after resume.
433
	call	__generic_findstack
434
	movl	%ebp,%ecx		# Get the stack pointer.
435
	subl	%eax,%ecx		# Subtract available space.
436
	addl	$BACKOFF,%ecx		# Back off 512 bytes.
437
	movl	%ecx,%gs:0x30		# Save new stack boundary.
438
	movl	4(%esp),%eax		# Function argument.
439
	movl	%eax,(%esp)
440
#ifdef __PIC__
441
	call	__x86.get_pc_thunk.bx	# %ebx may not be set up for us.
442
	addl	$_GLOBAL_OFFSET_TABLE_, %ebx
443
	call	_Unwind_Resume@PLT	# Resume unwinding.
444
#else
445
	call	_Unwind_Resume
446
#endif
447
 
448
#else /* defined(__x86_64__) */
449
 
450
 
451
# The 64-bit __morestack function.
452
 
453
	# We use a cleanup to restore the stack guard if an exception
454
	# is thrown through this code.
455
#ifndef __PIC__
456
	.cfi_personality 0x3,__gcc_personality_v0
457
	.cfi_lsda 0x3,.LLSDA1
458
#else
459
	.cfi_personality 0x9b,DW.ref.__gcc_personality_v0
460
	.cfi_lsda 0x1b,.LLSDA1
461
#endif
462
 
463
	# We will return a single return instruction, which will
464
	# return to the caller of our caller.  Let the unwinder skip
465
	# that single return instruction, and just return to the real
466
	# caller.
467
	.cfi_def_cfa %rsp,16
468
 
469
	# Set up a normal backtrace.
470
	pushq	%rbp
471
	.cfi_adjust_cfa_offset 8
472
	.cfi_offset %rbp, -24
473
	movq	%rsp, %rbp
474
	.cfi_def_cfa_register %rbp
475
 
476
	# In 64-bit mode the new stack frame size is passed in r10
477
        # and the argument size is passed in r11.
478
 
479
	addq	$BACKOFF,%r10		# Ask for backoff bytes.
480
	pushq	%r10			# Save new frame size.
481
 
482
	# In 64-bit mode the registers %rdi, %rsi, %rdx, %rcx, %r8,
483
	# and %r9 may be used for parameters.  We also preserve %rax
484
	# which the caller may use to hold %r10.
485
 
486
	pushq	%rax
487
	pushq	%rdi
488
	pushq	%rsi
489
	pushq	%rdx
490
	pushq	%rcx
491
	pushq	%r8
492
	pushq	%r9
493
 
494
	pushq	%r11
495
 
496
	# We entered morestack with the stack pointer aligned to a
497
	# 16-byte boundary (the call to morestack's caller used 8
498
	# bytes, and the call to morestack used 8 bytes).  We have now
499
	# pushed 10 registers, so we are still aligned to a 16-byte
500
	# boundary.
501
 
502
	call	__morestack_block_signals
503
 
504
	leaq	-8(%rbp),%rdi		# Address of new frame size.
505
	leaq	24(%rbp),%rsi		# The caller's parameters.
506
	popq	%rdx			# The size of the parameters.
507
 
508
	subq	$8,%rsp			# Align stack.
509
 
510
	call	__generic_morestack
511
 
512
	movq	-8(%rbp),%r10		# Reload modified frame size
513
	movq	%rax,%rsp		# Switch to the new stack.
514
	subq	%r10,%rax		# The end of the stack space.
515
	addq	$BACKOFF,%rax		# Back off 1024 bytes.
516
 
517
.LEHB0:
518
	# FIXME: The offset must match
519
	# TARGET_THREAD_SPLIT_STACK_OFFSET in
520
	# gcc/config/i386/linux64.h.
521
	# Macro to save the new stack boundary.
522
#ifdef __LP64__
523
#define X86_64_SAVE_NEW_STACK_BOUNDARY(reg)	movq	%r##reg,%fs:0x70
524
#else
525
#define X86_64_SAVE_NEW_STACK_BOUNDARY(reg)	movl	%e##reg,%fs:0x40
526
#endif
527
	X86_64_SAVE_NEW_STACK_BOUNDARY (ax)
528
 
529
	call	__morestack_unblock_signals
530
 
531
	movq	-24(%rbp),%rdi		# Restore registers.
532
	movq	-32(%rbp),%rsi
533
	movq	-40(%rbp),%rdx
534
	movq	-48(%rbp),%rcx
535
	movq	-56(%rbp),%r8
536
	movq	-64(%rbp),%r9
537
 
538
	movq	8(%rbp),%r10		# Increment the return address
539
	incq	%r10			# to skip the ret instruction;
540
					# see above.
541
 
542
	movq	-16(%rbp),%rax		# Restore caller's %rax.
543
 
544
	call	*%r10			# Call our caller!
545
 
546
	# The caller will return here, as predicted.
547
 
548
	# Save the registers which may hold a return value.  We
549
	# assume that __generic_releasestack does not touch any
550
	# floating point or vector registers.
551
	pushq	%rax
552
	pushq	%rdx
553
 
554
	call	__morestack_block_signals
555
 
556
	pushq	$0			# For alignment.
557
	pushq	$0			# Where the available space is returned.
558
	leaq	0(%rsp),%rdi		# Pass its address.
559
 
560
	call	__generic_releasestack
561
 
562
	subq	0(%rsp),%rax		# Subtract available space.
563
	addq	$BACKOFF,%rax		# Back off 1024 bytes.
564
.LEHE0:
565
	X86_64_SAVE_NEW_STACK_BOUNDARY (ax)
566
 
567
	addq	$16,%rsp		# Remove values from stack.
568
 
569
	# We need to restore the old stack pointer, which is in %rbp,
570
	# before we unblock signals.  We also need to restore %rax and
571
	# %rdx after we unblock signals but before we return.  Do this
572
	# by moving %rax and %rdx from the current stack to the old
573
	# stack.
574
 
575
	popq	%rdx			# Pop return value from current stack.
576
	popq	%rax
577
 
578
	movq	%rbp,%rsp		# Restore stack pointer.
579
 
580
	# Now (%rsp & 16) == 8.
581
 
582
	subq	$8,%rsp			# For alignment.
583
	pushq	%rax			# Push return value on old stack.
584
	pushq	%rdx
585
 
586
	call	__morestack_unblock_signals
587
 
588
	popq	%rdx			# Restore return value.
589
	popq	%rax
590
	addq	$8,%rsp
591
 
592
	.cfi_remember_state
593
	popq	%rbp
594
	.cfi_restore %rbp
595
	.cfi_def_cfa %rsp, 16
596
	ret				# Return to caller, which will
597
					# immediately return.
598
 
599
# This is the cleanup code called by the stack unwinder when unwinding
600
# through the code between .LEHB0 and .LEHE0 above.
601
 
602
.L1:
603
	.cfi_restore_state
604
	subq	$16,%rsp		# Maintain 16 byte alignment.
605
	movq	%rax,(%rsp)		# Save exception header.
606
	movq	%rbp,%rdi		# Stack pointer after resume.
607
	call	__generic_findstack
608
	movq	%rbp,%rcx		# Get the stack pointer.
609
	subq	%rax,%rcx		# Subtract available space.
610
	addq	$BACKOFF,%rcx		# Back off 1024 bytes.
611
	X86_64_SAVE_NEW_STACK_BOUNDARY (cx)
612
	movq	(%rsp),%rdi		# Restore exception data for call.
613
#ifdef __PIC__
614
	call	_Unwind_Resume@PLT	# Resume unwinding.
615
#else
616
	call	_Unwind_Resume		# Resume unwinding.
617
#endif
618
 
619
#endif /* defined(__x86_64__) */
620
 
621
	.cfi_endproc
622
#ifdef __ELF__
623
	.size	__morestack, . - __morestack
624
#endif
625
 
626
#if !defined(__x86_64__) && defined(__PIC__)
627
# Output the thunk to get PC into bx, since we use it above.
628
	.section	.text.__x86.get_pc_thunk.bx,"axG",@progbits,__x86.get_pc_thunk.bx,comdat
629
	.globl	__x86.get_pc_thunk.bx
630
	.hidden	__x86.get_pc_thunk.bx
631
#ifdef __ELF__
632
	.type	__x86.get_pc_thunk.bx, @function
633
#endif
634
__x86.get_pc_thunk.bx:
635
	.cfi_startproc
636
	movl	(%esp), %ebx
637
	ret
638
	.cfi_endproc
639
#ifdef __ELF__
640
	.size	__x86.get_pc_thunk.bx, . - __x86.get_pc_thunk.bx
641
#endif
642
#endif
643
 
644
# The exception table.  This tells the personality routine to execute
645
# the exception handler.
646
 
647
	.section	.gcc_except_table,"a",@progbits
648
	.align	4
649
.LLSDA1:
650
	.byte	0xff	# @LPStart format (omit)
651
	.byte	0xff	# @TType format (omit)
652
	.byte	0x1	# call-site format (uleb128)
653
	.uleb128 .LLSDACSE1-.LLSDACSB1	# Call-site table length
654
.LLSDACSB1:
655
	.uleb128 .LEHB0-.LFB1	# region 0 start
656
	.uleb128 .LEHE0-.LEHB0	# length
657
	.uleb128 .L1-.LFB1	# landing pad
658
	.uleb128 0		# action
659
.LLSDACSE1:
660
 
661
 
662
	.global __gcc_personality_v0
663
#ifdef __PIC__
664
	# Build a position independent reference to the basic
665
        # personality function.
666
	.hidden DW.ref.__gcc_personality_v0
667
	.weak   DW.ref.__gcc_personality_v0
668
	.section .data.DW.ref.__gcc_personality_v0,"awG",@progbits,DW.ref.__gcc_personality_v0,comdat
669
	.type	DW.ref.__gcc_personality_v0, @object
670
DW.ref.__gcc_personality_v0:
671
#ifndef __LP64__
672
	.align 4
673
	.size	DW.ref.__gcc_personality_v0, 4
674
	.long	__gcc_personality_v0
675
#else
676
	.align 8
677
	.size	DW.ref.__gcc_personality_v0, 8
678
	.quad	__gcc_personality_v0
679
#endif
680
#endif
681
 
682
#if defined __x86_64__ && defined __LP64__
683
 
684
# This entry point is used for the large model.  With this entry point
685
# the upper 32 bits of %r10 hold the argument size and the lower 32
686
# bits hold the new stack frame size.  There doesn't seem to be a way
687
# to know in the assembler code that we are assembling for the large
688
# model, and there doesn't seem to be a large model multilib anyhow.
689
# If one is developed, then the non-PIC code is probably OK since we
690
# will probably be close to the morestack code, but the PIC code
691
# almost certainly needs to be changed.  FIXME.
692
 
693
	.text
694
	.global	__morestack_large_model
695
	.hidden	__morestack_large_model
696
 
697
#ifdef __ELF__
698
	.type	__morestack_large_model,@function
699
#endif
700
 
701
__morestack_large_model:
702
 
703
	.cfi_startproc
704
 
705
	movq	%r10, %r11
706
	andl	$0xffffffff, %r10d
707
	sarq	$32, %r11
708
	jmp	__morestack
709
 
710
	.cfi_endproc
711
#ifdef __ELF__
712
       .size	__morestack_large_model, . - __morestack_large_model
713
#endif
714
 
715
#endif /* __x86_64__ && __LP64__ */
716
 
717
# Initialize the stack test value when the program starts or when a
718
# new thread starts.  We don't know how large the main stack is, so we
719
# guess conservatively.  We might be able to use getrlimit here.
720
 
721
	.text
722
	.global	__stack_split_initialize
723
	.hidden	__stack_split_initialize
724
 
725
#ifdef __ELF__
726
	.type	__stack_split_initialize, @function
727
#endif
728
 
729
__stack_split_initialize:
730
 
731
#ifndef __x86_64__
732
 
733
	leal	-16000(%esp),%eax	# We should have at least 16K.
734
	movl	%eax,%gs:0x30
735
	subl	$4,%esp			# Align stack.
736
	pushl	$16000
737
	pushl	%esp
738
#ifdef __PIC__
739
	call	__generic_morestack_set_initial_sp@PLT
740
#else
741
	call	__generic_morestack_set_initial_sp
742
#endif
743
	addl	$12,%esp
744
	ret
745
 
746
#else /* defined(__x86_64__) */
747
 
748
	leaq	-16000(%rsp),%rax	# We should have at least 16K.
749
	X86_64_SAVE_NEW_STACK_BOUNDARY (ax)
750
	subq	$8,%rsp			# Align stack.
751
	movq	%rsp,%rdi
752
	movq	$16000,%rsi
753
#ifdef __PIC__
754
	call	__generic_morestack_set_initial_sp@PLT
755
#else
756
	call	__generic_morestack_set_initial_sp
757
#endif
758
	addq	$8,%rsp
759
	ret
760
 
761
#endif /* defined(__x86_64__) */
762
 
763
#ifdef __ELF__
764
	.size	__stack_split_initialize, . - __stack_split_initialize
765
#endif
766
 
767
# Routines to get and set the guard, for __splitstack_getcontext,
768
# __splitstack_setcontext, and __splitstack_makecontext.
769
 
770
# void *__morestack_get_guard (void) returns the current stack guard.
771
	.text
772
	.global	__morestack_get_guard
773
	.hidden	__morestack_get_guard
774
 
775
#ifdef __ELF__
776
	.type	__morestack_get_guard,@function
777
#endif
778
 
779
__morestack_get_guard:
780
 
781
#ifndef __x86_64__
782
	movl	%gs:0x30,%eax
783
#else
784
#ifdef __LP64__
785
	movq	%fs:0x70,%rax
786
#else
787
	movl	%fs:0x40,%eax
788
#endif
789
#endif
790
	ret
791
 
792
#ifdef __ELF__
793
	.size	__morestack_get_guard, . - __morestack_get_guard
794
#endif
795
 
796
# void __morestack_set_guard (void *) sets the stack guard.
797
	.global	__morestack_set_guard
798
	.hidden	__morestack_set_guard
799
 
800
#ifdef __ELF__
801
	.type	__morestack_set_guard,@function
802
#endif
803
 
804
__morestack_set_guard:
805
 
806
#ifndef __x86_64__
807
	movl	4(%esp),%eax
808
	movl	%eax,%gs:0x30
809
#else
810
	X86_64_SAVE_NEW_STACK_BOUNDARY (di)
811
#endif
812
	ret
813
 
814
#ifdef __ELF__
815
	.size	__morestack_set_guard, . - __morestack_set_guard
816
#endif
817
 
818
# void *__morestack_make_guard (void *, size_t) returns the stack
819
# guard value for a stack.
820
	.global	__morestack_make_guard
821
	.hidden	__morestack_make_guard
822
 
823
#ifdef __ELF__
824
	.type	__morestack_make_guard,@function
825
#endif
826
 
827
__morestack_make_guard:
828
 
829
#ifndef __x86_64__
830
	movl	4(%esp),%eax
831
	subl	8(%esp),%eax
832
	addl	$BACKOFF,%eax
833
#else
834
	subq	%rsi,%rdi
835
	addq	$BACKOFF,%rdi
836
	movq	%rdi,%rax
837
#endif
838
	ret
839
 
840
#ifdef __ELF__
841
	.size	__morestack_make_guard, . - __morestack_make_guard
842
#endif
843
 
844
# Make __stack_split_initialize a high priority constructor.  FIXME:
845
# This is ELF specific.
846
 
847
	.section	.ctors.65535,"aw",@progbits
848
 
849
#ifndef __LP64__
850
	.align	4
851
	.long	__stack_split_initialize
852
	.long	__morestack_load_mmap
853
#else
854
	.align	8
855
	.quad	__stack_split_initialize
856
	.quad	__morestack_load_mmap
857
#endif
858
 
859
#ifdef __ELF__
860
	.section	.note.GNU-stack,"",@progbits
861
	.section	.note.GNU-split-stack,"",@progbits
862
	.section	.note.GNU-no-split-stack,"",@progbits
863
#endif