Subversion Repositories Kolibri OS

Rev

Blame | Last modification | View Log | RSS feed

  1. ;*****************************************************************************
  2. ;* x86inc.asm: x264asm abstraction layer
  3. ;*****************************************************************************
  4. ;* Copyright (C) 2005-2013 x264 project
  5. ;*
  6. ;* Authors: Loren Merritt <lorenm@u.washington.edu>
  7. ;*          Anton Mitrofanov <BugMaster@narod.ru>
  8. ;*          Jason Garrett-Glaser <darkshikari@gmail.com>
  9. ;*          Henrik Gramner <henrik@gramner.com>
  10. ;*
  11. ;* Permission to use, copy, modify, and/or distribute this software for any
  12. ;* purpose with or without fee is hereby granted, provided that the above
  13. ;* copyright notice and this permission notice appear in all copies.
  14. ;*
  15. ;* THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
  16. ;* WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
  17. ;* MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
  18. ;* ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
  19. ;* WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
  20. ;* ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
  21. ;* OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
  22. ;*****************************************************************************
  23.  
  24. ; This is a header file for the x264ASM assembly language, which uses
  25. ; NASM/YASM syntax combined with a large number of macros to provide easy
  26. ; abstraction between different calling conventions (x86_32, win64, linux64).
  27. ; It also has various other useful features to simplify writing the kind of
  28. ; DSP functions that are most often used in x264.
  29.  
  30. ; Unlike the rest of x264, this file is available under an ISC license, as it
  31. ; has significant usefulness outside of x264 and we want it to be available
  32. ; to the largest audience possible.  Of course, if you modify it for your own
  33. ; purposes to add a new feature, we strongly encourage contributing a patch
  34. ; as this feature might be useful for others as well.  Send patches or ideas
  35. ; to x264-devel@videolan.org .
  36.  
  37. %ifndef private_prefix
  38.     %define private_prefix x264
  39. %endif
  40.  
  41. %ifndef public_prefix
  42.     %define public_prefix private_prefix
  43. %endif
  44.  
  45. %define WIN64  0
  46. %define UNIX64 0
  47. %if ARCH_X86_64
  48.     %ifidn __OUTPUT_FORMAT__,win32
  49.         %define WIN64  1
  50.     %elifidn __OUTPUT_FORMAT__,win64
  51.         %define WIN64  1
  52.     %elifidn __OUTPUT_FORMAT__,x64
  53.         %define WIN64  1
  54.     %else
  55.         %define UNIX64 1
  56.     %endif
  57. %endif
  58.  
  59. %ifdef PREFIX
  60.     %define mangle(x) _ %+ x
  61. %else
  62.     %define mangle(x) x
  63. %endif
  64.  
  65. ; aout does not support align=
  66. ; NOTE: This section is out of sync with x264, in order to
  67. ; keep supporting OS/2.
  68. %macro SECTION_RODATA 0-1 16
  69.     %ifidn __OUTPUT_FORMAT__,aout
  70.         section .text
  71.     %else
  72.         SECTION .rodata align=%1
  73.     %endif
  74. %endmacro
  75.  
  76. %macro SECTION_TEXT 0-1 16
  77.     %ifidn __OUTPUT_FORMAT__,aout
  78.         SECTION .text
  79.     %else
  80.         SECTION .text align=%1
  81.     %endif
  82. %endmacro
  83.  
  84. %if WIN64
  85.     %define PIC
  86. %elif ARCH_X86_64 == 0
  87. ; x86_32 doesn't require PIC.
  88. ; Some distros prefer shared objects to be PIC, but nothing breaks if
  89. ; the code contains a few textrels, so we'll skip that complexity.
  90.     %undef PIC
  91. %endif
  92. %ifdef PIC
  93.     default rel
  94. %endif
  95.  
  96. %macro CPUNOP 1
  97.     %if HAVE_CPUNOP
  98.         CPU %1
  99.     %endif
  100. %endmacro
  101.  
  102. ; Always use long nops (reduces 0x90 spam in disassembly on x86_32)
  103. CPUNOP amdnop
  104.  
  105. ; Macros to eliminate most code duplication between x86_32 and x86_64:
  106. ; Currently this works only for leaf functions which load all their arguments
  107. ; into registers at the start, and make no other use of the stack. Luckily that
  108. ; covers most of x264's asm.
  109.  
  110. ; PROLOGUE:
  111. ; %1 = number of arguments. loads them from stack if needed.
  112. ; %2 = number of registers used. pushes callee-saved regs if needed.
  113. ; %3 = number of xmm registers used. pushes callee-saved xmm regs if needed.
  114. ; %4 = (optional) stack size to be allocated. If not aligned (x86-32 ICC 10.x,
  115. ;      MSVC or YMM), the stack will be manually aligned (to 16 or 32 bytes),
  116. ;      and an extra register will be allocated to hold the original stack
  117. ;      pointer (to not invalidate r0m etc.). To prevent the use of an extra
  118. ;      register as stack pointer, request a negative stack size.
  119. ; %4+/%5+ = list of names to define to registers
  120. ; PROLOGUE can also be invoked by adding the same options to cglobal
  121.  
  122. ; e.g.
  123. ; cglobal foo, 2,3,0, dst, src, tmp
  124. ; declares a function (foo), taking two args (dst and src) and one local variable (tmp)
  125.  
  126. ; TODO Some functions can use some args directly from the stack. If they're the
  127. ; last args then you can just not declare them, but if they're in the middle
  128. ; we need more flexible macro.
  129.  
  130. ; RET:
  131. ; Pops anything that was pushed by PROLOGUE, and returns.
  132.  
  133. ; REP_RET:
  134. ; Use this instead of RET if it's a branch target.
  135.  
  136. ; registers:
  137. ; rN and rNq are the native-size register holding function argument N
  138. ; rNd, rNw, rNb are dword, word, and byte size
  139. ; rNh is the high 8 bits of the word size
  140. ; rNm is the original location of arg N (a register or on the stack), dword
  141. ; rNmp is native size
  142.  
  143. %macro DECLARE_REG 2-3
  144.     %define r%1q %2
  145.     %define r%1d %2d
  146.     %define r%1w %2w
  147.     %define r%1b %2b
  148.     %define r%1h %2h
  149.     %define %2q %2
  150.     %if %0 == 2
  151.         %define r%1m  %2d
  152.         %define r%1mp %2
  153.     %elif ARCH_X86_64 ; memory
  154.         %define r%1m [rstk + stack_offset + %3]
  155.         %define r%1mp qword r %+ %1 %+ m
  156.     %else
  157.         %define r%1m [rstk + stack_offset + %3]
  158.         %define r%1mp dword r %+ %1 %+ m
  159.     %endif
  160.     %define r%1  %2
  161. %endmacro
  162.  
  163. %macro DECLARE_REG_SIZE 3
  164.     %define r%1q r%1
  165.     %define e%1q r%1
  166.     %define r%1d e%1
  167.     %define e%1d e%1
  168.     %define r%1w %1
  169.     %define e%1w %1
  170.     %define r%1h %3
  171.     %define e%1h %3
  172.     %define r%1b %2
  173.     %define e%1b %2
  174. %if ARCH_X86_64 == 0
  175.     %define r%1  e%1
  176. %endif
  177. %endmacro
  178.  
  179. DECLARE_REG_SIZE ax, al, ah
  180. DECLARE_REG_SIZE bx, bl, bh
  181. DECLARE_REG_SIZE cx, cl, ch
  182. DECLARE_REG_SIZE dx, dl, dh
  183. DECLARE_REG_SIZE si, sil, null
  184. DECLARE_REG_SIZE di, dil, null
  185. DECLARE_REG_SIZE bp, bpl, null
  186.  
  187. ; t# defines for when per-arch register allocation is more complex than just function arguments
  188.  
  189. %macro DECLARE_REG_TMP 1-*
  190.     %assign %%i 0
  191.     %rep %0
  192.         CAT_XDEFINE t, %%i, r%1
  193.         %assign %%i %%i+1
  194.         %rotate 1
  195.     %endrep
  196. %endmacro
  197.  
  198. %macro DECLARE_REG_TMP_SIZE 0-*
  199.     %rep %0
  200.         %define t%1q t%1 %+ q
  201.         %define t%1d t%1 %+ d
  202.         %define t%1w t%1 %+ w
  203.         %define t%1h t%1 %+ h
  204.         %define t%1b t%1 %+ b
  205.         %rotate 1
  206.     %endrep
  207. %endmacro
  208.  
  209. DECLARE_REG_TMP_SIZE 0,1,2,3,4,5,6,7,8,9,10,11,12,13,14
  210.  
  211. %if ARCH_X86_64
  212.     %define gprsize 8
  213. %else
  214.     %define gprsize 4
  215. %endif
  216.  
  217. %macro PUSH 1
  218.     push %1
  219.     %ifidn rstk, rsp
  220.         %assign stack_offset stack_offset+gprsize
  221.     %endif
  222. %endmacro
  223.  
  224. %macro POP 1
  225.     pop %1
  226.     %ifidn rstk, rsp
  227.         %assign stack_offset stack_offset-gprsize
  228.     %endif
  229. %endmacro
  230.  
  231. %macro PUSH_IF_USED 1-*
  232.     %rep %0
  233.         %if %1 < regs_used
  234.             PUSH r%1
  235.         %endif
  236.         %rotate 1
  237.     %endrep
  238. %endmacro
  239.  
  240. %macro POP_IF_USED 1-*
  241.     %rep %0
  242.         %if %1 < regs_used
  243.             pop r%1
  244.         %endif
  245.         %rotate 1
  246.     %endrep
  247. %endmacro
  248.  
  249. %macro LOAD_IF_USED 1-*
  250.     %rep %0
  251.         %if %1 < num_args
  252.             mov r%1, r %+ %1 %+ mp
  253.         %endif
  254.         %rotate 1
  255.     %endrep
  256. %endmacro
  257.  
  258. %macro SUB 2
  259.     sub %1, %2
  260.     %ifidn %1, rstk
  261.         %assign stack_offset stack_offset+(%2)
  262.     %endif
  263. %endmacro
  264.  
  265. %macro ADD 2
  266.     add %1, %2
  267.     %ifidn %1, rstk
  268.         %assign stack_offset stack_offset-(%2)
  269.     %endif
  270. %endmacro
  271.  
  272. %macro movifnidn 2
  273.     %ifnidn %1, %2
  274.         mov %1, %2
  275.     %endif
  276. %endmacro
  277.  
  278. %macro movsxdifnidn 2
  279.     %ifnidn %1, %2
  280.         movsxd %1, %2
  281.     %endif
  282. %endmacro
  283.  
  284. %macro ASSERT 1
  285.     %if (%1) == 0
  286.         %error assert failed
  287.     %endif
  288. %endmacro
  289.  
  290. %macro DEFINE_ARGS 0-*
  291.     %ifdef n_arg_names
  292.         %assign %%i 0
  293.         %rep n_arg_names
  294.             CAT_UNDEF arg_name %+ %%i, q
  295.             CAT_UNDEF arg_name %+ %%i, d
  296.             CAT_UNDEF arg_name %+ %%i, w
  297.             CAT_UNDEF arg_name %+ %%i, h
  298.             CAT_UNDEF arg_name %+ %%i, b
  299.             CAT_UNDEF arg_name %+ %%i, m
  300.             CAT_UNDEF arg_name %+ %%i, mp
  301.             CAT_UNDEF arg_name, %%i
  302.             %assign %%i %%i+1
  303.         %endrep
  304.     %endif
  305.  
  306.     %xdefine %%stack_offset stack_offset
  307.     %undef stack_offset ; so that the current value of stack_offset doesn't get baked in by xdefine
  308.     %assign %%i 0
  309.     %rep %0
  310.         %xdefine %1q r %+ %%i %+ q
  311.         %xdefine %1d r %+ %%i %+ d
  312.         %xdefine %1w r %+ %%i %+ w
  313.         %xdefine %1h r %+ %%i %+ h
  314.         %xdefine %1b r %+ %%i %+ b
  315.         %xdefine %1m r %+ %%i %+ m
  316.         %xdefine %1mp r %+ %%i %+ mp
  317.         CAT_XDEFINE arg_name, %%i, %1
  318.         %assign %%i %%i+1
  319.         %rotate 1
  320.     %endrep
  321.     %xdefine stack_offset %%stack_offset
  322.     %assign n_arg_names %0
  323. %endmacro
  324.  
  325. %macro ALLOC_STACK 1-2 0 ; stack_size, n_xmm_regs (for win64 only)
  326.     %ifnum %1
  327.         %if %1 != 0
  328.             %assign %%stack_alignment ((mmsize + 15) & ~15)
  329.             %assign stack_size %1
  330.             %if stack_size < 0
  331.                 %assign stack_size -stack_size
  332.             %endif
  333.             %assign stack_size_padded stack_size
  334.             %if WIN64
  335.                 %assign stack_size_padded stack_size_padded + 32 ; reserve 32 bytes for shadow space
  336.                 %if mmsize != 8
  337.                     %assign xmm_regs_used %2
  338.                     %if xmm_regs_used > 8
  339.                         %assign stack_size_padded stack_size_padded + (xmm_regs_used-8)*16
  340.                     %endif
  341.                 %endif
  342.             %endif
  343.             %if mmsize <= 16 && HAVE_ALIGNED_STACK
  344.                 %assign stack_size_padded stack_size_padded + %%stack_alignment - gprsize - (stack_offset & (%%stack_alignment - 1))
  345.                 SUB rsp, stack_size_padded
  346.             %else
  347.                 %assign %%reg_num (regs_used - 1)
  348.                 %xdefine rstk r %+ %%reg_num
  349.                 ; align stack, and save original stack location directly above
  350.                 ; it, i.e. in [rsp+stack_size_padded], so we can restore the
  351.                 ; stack in a single instruction (i.e. mov rsp, rstk or mov
  352.                 ; rsp, [rsp+stack_size_padded])
  353.                 mov  rstk, rsp
  354.                 %if %1 < 0 ; need to store rsp on stack
  355.                     sub  rsp, gprsize+stack_size_padded
  356.                     and  rsp, ~(%%stack_alignment-1)
  357.                     %xdefine rstkm [rsp+stack_size_padded]
  358.                     mov rstkm, rstk
  359.                 %else ; can keep rsp in rstk during whole function
  360.                     sub  rsp, stack_size_padded
  361.                     and  rsp, ~(%%stack_alignment-1)
  362.                     %xdefine rstkm rstk
  363.                 %endif
  364.             %endif
  365.             WIN64_PUSH_XMM
  366.         %endif
  367.     %endif
  368. %endmacro
  369.  
  370. %macro SETUP_STACK_POINTER 1
  371.     %ifnum %1
  372.         %if %1 != 0 && (HAVE_ALIGNED_STACK == 0 || mmsize == 32)
  373.             %if %1 > 0
  374.                 %assign regs_used (regs_used + 1)
  375.             %elif ARCH_X86_64 && regs_used == num_args && num_args <= 4 + UNIX64 * 2
  376.                 %warning "Stack pointer will overwrite register argument"
  377.             %endif
  378.         %endif
  379.     %endif
  380. %endmacro
  381.  
  382. %macro DEFINE_ARGS_INTERNAL 3+
  383.     %ifnum %2
  384.         DEFINE_ARGS %3
  385.     %elif %1 == 4
  386.         DEFINE_ARGS %2
  387.     %elif %1 > 4
  388.         DEFINE_ARGS %2, %3
  389.     %endif
  390. %endmacro
  391.  
  392. %if WIN64 ; Windows x64 ;=================================================
  393.  
  394. DECLARE_REG 0,  rcx
  395. DECLARE_REG 1,  rdx
  396. DECLARE_REG 2,  R8
  397. DECLARE_REG 3,  R9
  398. DECLARE_REG 4,  R10, 40
  399. DECLARE_REG 5,  R11, 48
  400. DECLARE_REG 6,  rax, 56
  401. DECLARE_REG 7,  rdi, 64
  402. DECLARE_REG 8,  rsi, 72
  403. DECLARE_REG 9,  rbx, 80
  404. DECLARE_REG 10, rbp, 88
  405. DECLARE_REG 11, R12, 96
  406. DECLARE_REG 12, R13, 104
  407. DECLARE_REG 13, R14, 112
  408. DECLARE_REG 14, R15, 120
  409.  
  410. %macro PROLOGUE 2-5+ 0 ; #args, #regs, #xmm_regs, [stack_size,] arg_names...
  411.     %assign num_args %1
  412.     %assign regs_used %2
  413.     ASSERT regs_used >= num_args
  414.     SETUP_STACK_POINTER %4
  415.     ASSERT regs_used <= 15
  416.     PUSH_IF_USED 7, 8, 9, 10, 11, 12, 13, 14
  417.     ALLOC_STACK %4, %3
  418.     %if mmsize != 8 && stack_size == 0
  419.         WIN64_SPILL_XMM %3
  420.     %endif
  421.     LOAD_IF_USED 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14
  422.     DEFINE_ARGS_INTERNAL %0, %4, %5
  423. %endmacro
  424.  
  425. %macro WIN64_PUSH_XMM 0
  426.     ; Use the shadow space to store XMM6 and XMM7, the rest needs stack space allocated.
  427.     %if xmm_regs_used > 6
  428.         movaps [rstk + stack_offset +  8], xmm6
  429.     %endif
  430.     %if xmm_regs_used > 7
  431.         movaps [rstk + stack_offset + 24], xmm7
  432.     %endif
  433.     %if xmm_regs_used > 8
  434.         %assign %%i 8
  435.         %rep xmm_regs_used-8
  436.             movaps [rsp + (%%i-8)*16 + stack_size + 32], xmm %+ %%i
  437.             %assign %%i %%i+1
  438.         %endrep
  439.     %endif
  440. %endmacro
  441.  
  442. %macro WIN64_SPILL_XMM 1
  443.     %assign xmm_regs_used %1
  444.     ASSERT xmm_regs_used <= 16
  445.     %if xmm_regs_used > 8
  446.         %assign stack_size_padded (xmm_regs_used-8)*16 + (~stack_offset&8) + 32
  447.         SUB rsp, stack_size_padded
  448.     %endif
  449.     WIN64_PUSH_XMM
  450. %endmacro
  451.  
  452. %macro WIN64_RESTORE_XMM_INTERNAL 1
  453.     %assign %%pad_size 0
  454.     %if xmm_regs_used > 8
  455.         %assign %%i xmm_regs_used
  456.         %rep xmm_regs_used-8
  457.             %assign %%i %%i-1
  458.             movaps xmm %+ %%i, [%1 + (%%i-8)*16 + stack_size + 32]
  459.         %endrep
  460.     %endif
  461.     %if stack_size_padded > 0
  462.         %if stack_size > 0 && (mmsize == 32 || HAVE_ALIGNED_STACK == 0)
  463.             mov rsp, rstkm
  464.         %else
  465.             add %1, stack_size_padded
  466.             %assign %%pad_size stack_size_padded
  467.         %endif
  468.     %endif
  469.     %if xmm_regs_used > 7
  470.         movaps xmm7, [%1 + stack_offset - %%pad_size + 24]
  471.     %endif
  472.     %if xmm_regs_used > 6
  473.         movaps xmm6, [%1 + stack_offset - %%pad_size +  8]
  474.     %endif
  475. %endmacro
  476.  
  477. %macro WIN64_RESTORE_XMM 1
  478.     WIN64_RESTORE_XMM_INTERNAL %1
  479.     %assign stack_offset (stack_offset-stack_size_padded)
  480.     %assign xmm_regs_used 0
  481. %endmacro
  482.  
  483. %define has_epilogue regs_used > 7 || xmm_regs_used > 6 || mmsize == 32 || stack_size > 0
  484.  
  485. %macro RET 0
  486.     WIN64_RESTORE_XMM_INTERNAL rsp
  487.     POP_IF_USED 14, 13, 12, 11, 10, 9, 8, 7
  488. %if mmsize == 32
  489.     vzeroupper
  490. %endif
  491.     AUTO_REP_RET
  492. %endmacro
  493.  
  494. %elif ARCH_X86_64 ; *nix x64 ;=============================================
  495.  
  496. DECLARE_REG 0,  rdi
  497. DECLARE_REG 1,  rsi
  498. DECLARE_REG 2,  rdx
  499. DECLARE_REG 3,  rcx
  500. DECLARE_REG 4,  R8
  501. DECLARE_REG 5,  R9
  502. DECLARE_REG 6,  rax, 8
  503. DECLARE_REG 7,  R10, 16
  504. DECLARE_REG 8,  R11, 24
  505. DECLARE_REG 9,  rbx, 32
  506. DECLARE_REG 10, rbp, 40
  507. DECLARE_REG 11, R12, 48
  508. DECLARE_REG 12, R13, 56
  509. DECLARE_REG 13, R14, 64
  510. DECLARE_REG 14, R15, 72
  511.  
  512. %macro PROLOGUE 2-5+ ; #args, #regs, #xmm_regs, [stack_size,] arg_names...
  513.     %assign num_args %1
  514.     %assign regs_used %2
  515.     ASSERT regs_used >= num_args
  516.     SETUP_STACK_POINTER %4
  517.     ASSERT regs_used <= 15
  518.     PUSH_IF_USED 9, 10, 11, 12, 13, 14
  519.     ALLOC_STACK %4
  520.     LOAD_IF_USED 6, 7, 8, 9, 10, 11, 12, 13, 14
  521.     DEFINE_ARGS_INTERNAL %0, %4, %5
  522. %endmacro
  523.  
  524. %define has_epilogue regs_used > 9 || mmsize == 32 || stack_size > 0
  525.  
  526. %macro RET 0
  527. %if stack_size_padded > 0
  528. %if mmsize == 32 || HAVE_ALIGNED_STACK == 0
  529.     mov rsp, rstkm
  530. %else
  531.     add rsp, stack_size_padded
  532. %endif
  533. %endif
  534.     POP_IF_USED 14, 13, 12, 11, 10, 9
  535. %if mmsize == 32
  536.     vzeroupper
  537. %endif
  538.     AUTO_REP_RET
  539. %endmacro
  540.  
  541. %else ; X86_32 ;==============================================================
  542.  
  543. DECLARE_REG 0, eax, 4
  544. DECLARE_REG 1, ecx, 8
  545. DECLARE_REG 2, edx, 12
  546. DECLARE_REG 3, ebx, 16
  547. DECLARE_REG 4, esi, 20
  548. DECLARE_REG 5, edi, 24
  549. DECLARE_REG 6, ebp, 28
  550. %define rsp esp
  551.  
  552. %macro DECLARE_ARG 1-*
  553.     %rep %0
  554.         %define r%1m [rstk + stack_offset + 4*%1 + 4]
  555.         %define r%1mp dword r%1m
  556.         %rotate 1
  557.     %endrep
  558. %endmacro
  559.  
  560. DECLARE_ARG 7, 8, 9, 10, 11, 12, 13, 14
  561.  
  562. %macro PROLOGUE 2-5+ ; #args, #regs, #xmm_regs, [stack_size,] arg_names...
  563.     %assign num_args %1
  564.     %assign regs_used %2
  565.     ASSERT regs_used >= num_args
  566.     %if num_args > 7
  567.         %assign num_args 7
  568.     %endif
  569.     %if regs_used > 7
  570.         %assign regs_used 7
  571.     %endif
  572.     SETUP_STACK_POINTER %4
  573.     ASSERT regs_used <= 7
  574.     PUSH_IF_USED 3, 4, 5, 6
  575.     ALLOC_STACK %4
  576.     LOAD_IF_USED 0, 1, 2, 3, 4, 5, 6
  577.     DEFINE_ARGS_INTERNAL %0, %4, %5
  578. %endmacro
  579.  
  580. %define has_epilogue regs_used > 3 || mmsize == 32 || stack_size > 0
  581.  
  582. %macro RET 0
  583. %if stack_size_padded > 0
  584. %if mmsize == 32 || HAVE_ALIGNED_STACK == 0
  585.     mov rsp, rstkm
  586. %else
  587.     add rsp, stack_size_padded
  588. %endif
  589. %endif
  590.     POP_IF_USED 6, 5, 4, 3
  591. %if mmsize == 32
  592.     vzeroupper
  593. %endif
  594.     AUTO_REP_RET
  595. %endmacro
  596.  
  597. %endif ;======================================================================
  598.  
  599. %if WIN64 == 0
  600. %macro WIN64_SPILL_XMM 1
  601. %endmacro
  602. %macro WIN64_RESTORE_XMM 1
  603. %endmacro
  604. %macro WIN64_PUSH_XMM 0
  605. %endmacro
  606. %endif
  607.  
  608. ; On AMD cpus <=K10, an ordinary ret is slow if it immediately follows either
  609. ; a branch or a branch target. So switch to a 2-byte form of ret in that case.
  610. ; We can automatically detect "follows a branch", but not a branch target.
  611. ; (SSSE3 is a sufficient condition to know that your cpu doesn't have this problem.)
  612. %macro REP_RET 0
  613.     %if has_epilogue
  614.         RET
  615.     %else
  616.         rep ret
  617.     %endif
  618. %endmacro
  619.  
  620. %define last_branch_adr $$
  621. %macro AUTO_REP_RET 0
  622.     %ifndef cpuflags
  623.         times ((last_branch_adr-$)>>31)+1 rep ; times 1 iff $ != last_branch_adr.
  624.     %elif notcpuflag(ssse3)
  625.         times ((last_branch_adr-$)>>31)+1 rep
  626.     %endif
  627.     ret
  628. %endmacro
  629.  
  630. %macro BRANCH_INSTR 0-*
  631.     %rep %0
  632.         %macro %1 1-2 %1
  633.             %2 %1
  634.             %%branch_instr:
  635.             %xdefine last_branch_adr %%branch_instr
  636.         %endmacro
  637.         %rotate 1
  638.     %endrep
  639. %endmacro
  640.  
  641. BRANCH_INSTR jz, je, jnz, jne, jl, jle, jnl, jnle, jg, jge, jng, jnge, ja, jae, jna, jnae, jb, jbe, jnb, jnbe, jc, jnc, js, jns, jo, jno, jp, jnp
  642.  
  643. %macro TAIL_CALL 2 ; callee, is_nonadjacent
  644.     %if has_epilogue
  645.         call %1
  646.         RET
  647.     %elif %2
  648.         jmp %1
  649.     %endif
  650. %endmacro
  651.  
  652. ;=============================================================================
  653. ; arch-independent part
  654. ;=============================================================================
  655.  
  656. %assign function_align 16
  657.  
  658. ; Begin a function.
  659. ; Applies any symbol mangling needed for C linkage, and sets up a define such that
  660. ; subsequent uses of the function name automatically refer to the mangled version.
  661. ; Appends cpuflags to the function name if cpuflags has been specified.
  662. ; The "" empty default parameter is a workaround for nasm, which fails if SUFFIX
  663. ; is empty and we call cglobal_internal with just %1 %+ SUFFIX (without %2).
  664. %macro cglobal 1-2+ "" ; name, [PROLOGUE args]
  665.     cglobal_internal 1, %1 %+ SUFFIX, %2
  666. %endmacro
  667. %macro cvisible 1-2+ "" ; name, [PROLOGUE args]
  668.     cglobal_internal 0, %1 %+ SUFFIX, %2
  669. %endmacro
  670. %macro cglobal_internal 2-3+
  671.     %if %1
  672.         %xdefine %%FUNCTION_PREFIX private_prefix
  673.         %xdefine %%VISIBILITY hidden
  674.     %else
  675.         %xdefine %%FUNCTION_PREFIX public_prefix
  676.         %xdefine %%VISIBILITY
  677.     %endif
  678.     %ifndef cglobaled_%2
  679.         %xdefine %2 mangle(%%FUNCTION_PREFIX %+ _ %+ %2)
  680.         %xdefine %2.skip_prologue %2 %+ .skip_prologue
  681.         CAT_XDEFINE cglobaled_, %2, 1
  682.     %endif
  683.     %xdefine current_function %2
  684.     %ifidn __OUTPUT_FORMAT__,elf
  685.         global %2:function %%VISIBILITY
  686.     %else
  687.         global %2
  688.     %endif
  689.     align function_align
  690.     %2:
  691.     RESET_MM_PERMUTATION        ; needed for x86-64, also makes disassembly somewhat nicer
  692.     %xdefine rstk rsp           ; copy of the original stack pointer, used when greater alignment than the known stack alignment is required
  693.     %assign stack_offset 0      ; stack pointer offset relative to the return address
  694.     %assign stack_size 0        ; amount of stack space that can be freely used inside a function
  695.     %assign stack_size_padded 0 ; total amount of allocated stack space, including space for callee-saved xmm registers on WIN64 and alignment padding
  696.     %assign xmm_regs_used 0     ; number of XMM registers requested, used for dealing with callee-saved registers on WIN64
  697.     %ifnidn %3, ""
  698.         PROLOGUE %3
  699.     %endif
  700. %endmacro
  701.  
  702. %macro cextern 1
  703.     %xdefine %1 mangle(private_prefix %+ _ %+ %1)
  704.     CAT_XDEFINE cglobaled_, %1, 1
  705.     extern %1
  706. %endmacro
  707.  
  708. ; like cextern, but without the prefix
  709. %macro cextern_naked 1
  710.     %xdefine %1 mangle(%1)
  711.     CAT_XDEFINE cglobaled_, %1, 1
  712.     extern %1
  713. %endmacro
  714.  
  715. %macro const 1-2+
  716.     %xdefine %1 mangle(private_prefix %+ _ %+ %1)
  717.     %ifidn __OUTPUT_FORMAT__,elf
  718.         global %1:data hidden
  719.     %else
  720.         global %1
  721.     %endif
  722.     %1: %2
  723. %endmacro
  724.  
  725. ; This is needed for ELF, otherwise the GNU linker assumes the stack is
  726. ; executable by default.
  727. %ifidn __OUTPUT_FORMAT__,elf
  728. SECTION .note.GNU-stack noalloc noexec nowrite progbits
  729. %endif
  730.  
  731. ; cpuflags
  732.  
  733. %assign cpuflags_mmx      (1<<0)
  734. %assign cpuflags_mmx2     (1<<1) | cpuflags_mmx
  735. %assign cpuflags_3dnow    (1<<2) | cpuflags_mmx
  736. %assign cpuflags_3dnowext (1<<3) | cpuflags_3dnow
  737. %assign cpuflags_sse      (1<<4) | cpuflags_mmx2
  738. %assign cpuflags_sse2     (1<<5) | cpuflags_sse
  739. %assign cpuflags_sse2slow (1<<6) | cpuflags_sse2
  740. %assign cpuflags_sse3     (1<<7) | cpuflags_sse2
  741. %assign cpuflags_ssse3    (1<<8) | cpuflags_sse3
  742. %assign cpuflags_sse4     (1<<9) | cpuflags_ssse3
  743. %assign cpuflags_sse42    (1<<10)| cpuflags_sse4
  744. %assign cpuflags_avx      (1<<11)| cpuflags_sse42
  745. %assign cpuflags_xop      (1<<12)| cpuflags_avx
  746. %assign cpuflags_fma4     (1<<13)| cpuflags_avx
  747. %assign cpuflags_avx2     (1<<14)| cpuflags_avx
  748. %assign cpuflags_fma3     (1<<15)| cpuflags_avx
  749.  
  750. %assign cpuflags_cache32  (1<<16)
  751. %assign cpuflags_cache64  (1<<17)
  752. %assign cpuflags_slowctz  (1<<18)
  753. %assign cpuflags_lzcnt    (1<<19)
  754. %assign cpuflags_aligned  (1<<20) ; not a cpu feature, but a function variant
  755. %assign cpuflags_atom     (1<<21)
  756. %assign cpuflags_bmi1     (1<<22)|cpuflags_lzcnt
  757. %assign cpuflags_bmi2     (1<<23)|cpuflags_bmi1
  758.  
  759. %define    cpuflag(x) ((cpuflags & (cpuflags_ %+ x)) == (cpuflags_ %+ x))
  760. %define notcpuflag(x) ((cpuflags & (cpuflags_ %+ x)) != (cpuflags_ %+ x))
  761.  
  762. ; Takes up to 2 cpuflags from the above list.
  763. ; All subsequent functions (up to the next INIT_CPUFLAGS) is built for the specified cpu.
  764. ; You shouldn't need to invoke this macro directly, it's a subroutine for INIT_MMX &co.
  765. %macro INIT_CPUFLAGS 0-2
  766.     CPUNOP amdnop
  767.     %if %0 >= 1
  768.         %xdefine cpuname %1
  769.         %assign cpuflags cpuflags_%1
  770.         %if %0 >= 2
  771.             %xdefine cpuname %1_%2
  772.             %assign cpuflags cpuflags | cpuflags_%2
  773.         %endif
  774.         %xdefine SUFFIX _ %+ cpuname
  775.         %if cpuflag(avx)
  776.             %assign avx_enabled 1
  777.         %endif
  778.         %if (mmsize == 16 && notcpuflag(sse2)) || (mmsize == 32 && notcpuflag(avx2))
  779.             %define mova movaps
  780.             %define movu movups
  781.             %define movnta movntps
  782.         %endif
  783.         %if cpuflag(aligned)
  784.             %define movu mova
  785.         %elifidn %1, sse3
  786.             %define movu lddqu
  787.         %endif
  788.         %if notcpuflag(sse2)
  789.             CPUNOP basicnop
  790.         %endif
  791.     %else
  792.         %xdefine SUFFIX
  793.         %undef cpuname
  794.         %undef cpuflags
  795.     %endif
  796. %endmacro
  797.  
  798. ; Merge mmx and sse*
  799. ; m# is a simd regsiter of the currently selected size
  800. ; xm# is the corresponding xmmreg (if selcted xmm or ymm size), or mmreg (if selected mmx)
  801. ; ym# is the corresponding ymmreg (if selcted xmm or ymm size), or mmreg (if selected mmx)
  802. ; (All 3 remain in sync through SWAP.)
  803.  
  804. %macro CAT_XDEFINE 3
  805.     %xdefine %1%2 %3
  806. %endmacro
  807.  
  808. %macro CAT_UNDEF 2
  809.     %undef %1%2
  810. %endmacro
  811.  
  812. %macro INIT_MMX 0-1+
  813.     %assign avx_enabled 0
  814.     %define RESET_MM_PERMUTATION INIT_MMX %1
  815.     %define mmsize 8
  816.     %define num_mmregs 8
  817.     %define mova movq
  818.     %define movu movq
  819.     %define movh movd
  820.     %define movnta movntq
  821.     %assign %%i 0
  822.     %rep 8
  823.     CAT_XDEFINE m, %%i, mm %+ %%i
  824.     CAT_XDEFINE nmm, %%i, %%i
  825.     %assign %%i %%i+1
  826.     %endrep
  827.     %rep 8
  828.     CAT_UNDEF m, %%i
  829.     CAT_UNDEF nmm, %%i
  830.     %assign %%i %%i+1
  831.     %endrep
  832.     INIT_CPUFLAGS %1
  833. %endmacro
  834.  
  835. %macro INIT_XMM 0-1+
  836.     %assign avx_enabled 0
  837.     %define RESET_MM_PERMUTATION INIT_XMM %1
  838.     %define mmsize 16
  839.     %define num_mmregs 8
  840.     %if ARCH_X86_64
  841.     %define num_mmregs 16
  842.     %endif
  843.     %define mova movdqa
  844.     %define movu movdqu
  845.     %define movh movq
  846.     %define movnta movntdq
  847.     %assign %%i 0
  848.     %rep num_mmregs
  849.     CAT_XDEFINE m, %%i, xmm %+ %%i
  850.     CAT_XDEFINE nxmm, %%i, %%i
  851.     %assign %%i %%i+1
  852.     %endrep
  853.     INIT_CPUFLAGS %1
  854. %endmacro
  855.  
  856. ; FIXME: INIT_AVX can be replaced by INIT_XMM avx
  857. %macro INIT_AVX 0
  858.     INIT_XMM
  859.     %assign avx_enabled 1
  860.     %define PALIGNR PALIGNR_SSSE3
  861.     %define RESET_MM_PERMUTATION INIT_AVX
  862. %endmacro
  863.  
  864. %macro INIT_YMM 0-1+
  865.     %assign avx_enabled 1
  866.     %define RESET_MM_PERMUTATION INIT_YMM %1
  867.     %define mmsize 32
  868.     %define num_mmregs 8
  869.     %if ARCH_X86_64
  870.     %define num_mmregs 16
  871.     %endif
  872.     %define mova movdqa
  873.     %define movu movdqu
  874.     %undef movh
  875.     %define movnta movntdq
  876.     %assign %%i 0
  877.     %rep num_mmregs
  878.     CAT_XDEFINE m, %%i, ymm %+ %%i
  879.     CAT_XDEFINE nymm, %%i, %%i
  880.     %assign %%i %%i+1
  881.     %endrep
  882.     INIT_CPUFLAGS %1
  883. %endmacro
  884.  
  885. INIT_XMM
  886.  
  887. %macro DECLARE_MMCAST 1
  888.     %define  mmmm%1   mm%1
  889.     %define  mmxmm%1  mm%1
  890.     %define  mmymm%1  mm%1
  891.     %define xmmmm%1   mm%1
  892.     %define xmmxmm%1 xmm%1
  893.     %define xmmymm%1 xmm%1
  894.     %define ymmmm%1   mm%1
  895.     %define ymmxmm%1 ymm%1
  896.     %define ymmymm%1 ymm%1
  897.     %define xm%1 xmm %+ m%1
  898.     %define ym%1 ymm %+ m%1
  899. %endmacro
  900.  
  901. %assign i 0
  902. %rep 16
  903.     DECLARE_MMCAST i
  904. %assign i i+1
  905. %endrep
  906.  
  907. ; I often want to use macros that permute their arguments. e.g. there's no
  908. ; efficient way to implement butterfly or transpose or dct without swapping some
  909. ; arguments.
  910. ;
  911. ; I would like to not have to manually keep track of the permutations:
  912. ; If I insert a permutation in the middle of a function, it should automatically
  913. ; change everything that follows. For more complex macros I may also have multiple
  914. ; implementations, e.g. the SSE2 and SSSE3 versions may have different permutations.
  915. ;
  916. ; Hence these macros. Insert a PERMUTE or some SWAPs at the end of a macro that
  917. ; permutes its arguments. It's equivalent to exchanging the contents of the
  918. ; registers, except that this way you exchange the register names instead, so it
  919. ; doesn't cost any cycles.
  920.  
  921. %macro PERMUTE 2-* ; takes a list of pairs to swap
  922. %rep %0/2
  923.     %xdefine %%tmp%2 m%2
  924.     %rotate 2
  925. %endrep
  926. %rep %0/2
  927.     %xdefine m%1 %%tmp%2
  928.     CAT_XDEFINE n, m%1, %1
  929.     %rotate 2
  930. %endrep
  931. %endmacro
  932.  
  933. %macro SWAP 2+ ; swaps a single chain (sometimes more concise than pairs)
  934. %ifnum %1 ; SWAP 0, 1, ...
  935.     SWAP_INTERNAL_NUM %1, %2
  936. %else ; SWAP m0, m1, ...
  937.     SWAP_INTERNAL_NAME %1, %2
  938. %endif
  939. %endmacro
  940.  
  941. %macro SWAP_INTERNAL_NUM 2-*
  942.     %rep %0-1
  943.         %xdefine %%tmp m%1
  944.         %xdefine m%1 m%2
  945.         %xdefine m%2 %%tmp
  946.         CAT_XDEFINE n, m%1, %1
  947.         CAT_XDEFINE n, m%2, %2
  948.     %rotate 1
  949.     %endrep
  950. %endmacro
  951.  
  952. %macro SWAP_INTERNAL_NAME 2-*
  953.     %xdefine %%args n %+ %1
  954.     %rep %0-1
  955.         %xdefine %%args %%args, n %+ %2
  956.     %rotate 1
  957.     %endrep
  958.     SWAP_INTERNAL_NUM %%args
  959. %endmacro
  960.  
  961. ; If SAVE_MM_PERMUTATION is placed at the end of a function, then any later
  962. ; calls to that function will automatically load the permutation, so values can
  963. ; be returned in mmregs.
  964. %macro SAVE_MM_PERMUTATION 0-1
  965.     %if %0
  966.         %xdefine %%f %1_m
  967.     %else
  968.         %xdefine %%f current_function %+ _m
  969.     %endif
  970.     %assign %%i 0
  971.     %rep num_mmregs
  972.         CAT_XDEFINE %%f, %%i, m %+ %%i
  973.     %assign %%i %%i+1
  974.     %endrep
  975. %endmacro
  976.  
  977. %macro LOAD_MM_PERMUTATION 1 ; name to load from
  978.     %ifdef %1_m0
  979.         %assign %%i 0
  980.         %rep num_mmregs
  981.             CAT_XDEFINE m, %%i, %1_m %+ %%i
  982.             CAT_XDEFINE n, m %+ %%i, %%i
  983.         %assign %%i %%i+1
  984.         %endrep
  985.     %endif
  986. %endmacro
  987.  
  988. ; Append cpuflags to the callee's name iff the appended name is known and the plain name isn't
  989. %macro call 1
  990.     call_internal %1 %+ SUFFIX, %1
  991. %endmacro
  992. %macro call_internal 2
  993.     %xdefine %%i %2
  994.     %ifndef cglobaled_%2
  995.         %ifdef cglobaled_%1
  996.             %xdefine %%i %1
  997.         %endif
  998.     %endif
  999.     call %%i
  1000.     LOAD_MM_PERMUTATION %%i
  1001. %endmacro
  1002.  
  1003. ; Substitutions that reduce instruction size but are functionally equivalent
  1004. %macro add 2
  1005.     %ifnum %2
  1006.         %if %2==128
  1007.             sub %1, -128
  1008.         %else
  1009.             add %1, %2
  1010.         %endif
  1011.     %else
  1012.         add %1, %2
  1013.     %endif
  1014. %endmacro
  1015.  
  1016. %macro sub 2
  1017.     %ifnum %2
  1018.         %if %2==128
  1019.             add %1, -128
  1020.         %else
  1021.             sub %1, %2
  1022.         %endif
  1023.     %else
  1024.         sub %1, %2
  1025.     %endif
  1026. %endmacro
  1027.  
  1028. ;=============================================================================
  1029. ; AVX abstraction layer
  1030. ;=============================================================================
  1031.  
  1032. %assign i 0
  1033. %rep 16
  1034.     %if i < 8
  1035.         CAT_XDEFINE sizeofmm, i, 8
  1036.     %endif
  1037.     CAT_XDEFINE sizeofxmm, i, 16
  1038.     CAT_XDEFINE sizeofymm, i, 32
  1039. %assign i i+1
  1040. %endrep
  1041. %undef i
  1042.  
  1043. %macro CHECK_AVX_INSTR_EMU 3-*
  1044.     %xdefine %%opcode %1
  1045.     %xdefine %%dst %2
  1046.     %rep %0-2
  1047.         %ifidn %%dst, %3
  1048.             %error non-avx emulation of ``%%opcode'' is not supported
  1049.         %endif
  1050.         %rotate 1
  1051.     %endrep
  1052. %endmacro
  1053.  
  1054. ;%1 == instruction
  1055. ;%2 == 1 if float, 0 if int
  1056. ;%3 == 1 if non-destructive or 4-operand (xmm, xmm, xmm, imm), 0 otherwise
  1057. ;%4 == 1 if commutative (i.e. doesn't matter which src arg is which), 0 if not
  1058. ;%5+: operands
  1059. %macro RUN_AVX_INSTR 5-8+
  1060.     %ifnum sizeof%6
  1061.         %assign %%sizeofreg sizeof%6
  1062.     %elifnum sizeof%5
  1063.         %assign %%sizeofreg sizeof%5
  1064.     %else
  1065.         %assign %%sizeofreg mmsize
  1066.     %endif
  1067.     %assign %%emulate_avx 0
  1068.     %if avx_enabled && %%sizeofreg >= 16
  1069.         %xdefine %%instr v%1
  1070.     %else
  1071.         %xdefine %%instr %1
  1072.         %if %0 >= 7+%3
  1073.             %assign %%emulate_avx 1
  1074.         %endif
  1075.     %endif
  1076.  
  1077.     %if %%emulate_avx
  1078.         %xdefine %%src1 %6
  1079.         %xdefine %%src2 %7
  1080.         %ifnidn %5, %6
  1081.             %if %0 >= 8
  1082.                 CHECK_AVX_INSTR_EMU {%1 %5, %6, %7, %8}, %5, %7, %8
  1083.             %else
  1084.                 CHECK_AVX_INSTR_EMU {%1 %5, %6, %7}, %5, %7
  1085.             %endif
  1086.             %if %4 && %3 == 0
  1087.                 %ifnid %7
  1088.                     ; 3-operand AVX instructions with a memory arg can only have it in src2,
  1089.                     ; whereas SSE emulation prefers to have it in src1 (i.e. the mov).
  1090.                     ; So, if the instruction is commutative with a memory arg, swap them.
  1091.                     %xdefine %%src1 %7
  1092.                     %xdefine %%src2 %6
  1093.                 %endif
  1094.             %endif
  1095.             %if %%sizeofreg == 8
  1096.                 MOVQ %5, %%src1
  1097.             %elif %2
  1098.                 MOVAPS %5, %%src1
  1099.             %else
  1100.                 MOVDQA %5, %%src1
  1101.             %endif
  1102.         %endif
  1103.         %if %0 >= 8
  1104.             %1 %5, %%src2, %8
  1105.         %else
  1106.             %1 %5, %%src2
  1107.         %endif
  1108.     %elif %0 >= 8
  1109.         %%instr %5, %6, %7, %8
  1110.     %elif %0 == 7
  1111.         %%instr %5, %6, %7
  1112.     %elif %0 == 6
  1113.         %%instr %5, %6
  1114.     %else
  1115.         %%instr %5
  1116.     %endif
  1117. %endmacro
  1118.  
  1119. ;%1 == instruction
  1120. ;%2 == 1 if float, 0 if int
  1121. ;%3 == 1 if non-destructive or 4-operand (xmm, xmm, xmm, imm), 0 otherwise
  1122. ;%4 == 1 if commutative (i.e. doesn't matter which src arg is which), 0 if not
  1123. %macro AVX_INSTR 1-4 0, 1, 0
  1124.     %macro %1 1-9 fnord, fnord, fnord, fnord, %1, %2, %3, %4
  1125.         %ifidn %2, fnord
  1126.             RUN_AVX_INSTR %6, %7, %8, %9, %1
  1127.         %elifidn %3, fnord
  1128.             RUN_AVX_INSTR %6, %7, %8, %9, %1, %2
  1129.         %elifidn %4, fnord
  1130.             RUN_AVX_INSTR %6, %7, %8, %9, %1, %2, %3
  1131.         %elifidn %5, fnord
  1132.             RUN_AVX_INSTR %6, %7, %8, %9, %1, %2, %3, %4
  1133.         %else
  1134.             RUN_AVX_INSTR %6, %7, %8, %9, %1, %2, %3, %4, %5
  1135.         %endif
  1136.     %endmacro
  1137. %endmacro
  1138.  
  1139. ; Instructions with both VEX and non-VEX encodings
  1140. ; Non-destructive instructions are written without parameters
  1141. AVX_INSTR addpd, 1, 0, 1
  1142. AVX_INSTR addps, 1, 0, 1
  1143. AVX_INSTR addsd, 1, 0, 1
  1144. AVX_INSTR addss, 1, 0, 1
  1145. AVX_INSTR addsubpd, 1, 0, 0
  1146. AVX_INSTR addsubps, 1, 0, 0
  1147. AVX_INSTR aesdec, 0, 0, 0
  1148. AVX_INSTR aesdeclast, 0, 0, 0
  1149. AVX_INSTR aesenc, 0, 0, 0
  1150. AVX_INSTR aesenclast, 0, 0, 0
  1151. AVX_INSTR aesimc
  1152. AVX_INSTR aeskeygenassist
  1153. AVX_INSTR andnpd, 1, 0, 0
  1154. AVX_INSTR andnps, 1, 0, 0
  1155. AVX_INSTR andpd, 1, 0, 1
  1156. AVX_INSTR andps, 1, 0, 1
  1157. AVX_INSTR blendpd, 1, 0, 0
  1158. AVX_INSTR blendps, 1, 0, 0
  1159. AVX_INSTR blendvpd, 1, 0, 0
  1160. AVX_INSTR blendvps, 1, 0, 0
  1161. AVX_INSTR cmppd, 1, 1, 0
  1162. AVX_INSTR cmpps, 1, 1, 0
  1163. AVX_INSTR cmpsd, 1, 1, 0
  1164. AVX_INSTR cmpss, 1, 1, 0
  1165. AVX_INSTR comisd
  1166. AVX_INSTR comiss
  1167. AVX_INSTR cvtdq2pd
  1168. AVX_INSTR cvtdq2ps
  1169. AVX_INSTR cvtpd2dq
  1170. AVX_INSTR cvtpd2ps
  1171. AVX_INSTR cvtps2dq
  1172. AVX_INSTR cvtps2pd
  1173. AVX_INSTR cvtsd2si
  1174. AVX_INSTR cvtsd2ss
  1175. AVX_INSTR cvtsi2sd
  1176. AVX_INSTR cvtsi2ss
  1177. AVX_INSTR cvtss2sd
  1178. AVX_INSTR cvtss2si
  1179. AVX_INSTR cvttpd2dq
  1180. AVX_INSTR cvttps2dq
  1181. AVX_INSTR cvttsd2si
  1182. AVX_INSTR cvttss2si
  1183. AVX_INSTR divpd, 1, 0, 0
  1184. AVX_INSTR divps, 1, 0, 0
  1185. AVX_INSTR divsd, 1, 0, 0
  1186. AVX_INSTR divss, 1, 0, 0
  1187. AVX_INSTR dppd, 1, 1, 0
  1188. AVX_INSTR dpps, 1, 1, 0
  1189. AVX_INSTR extractps
  1190. AVX_INSTR haddpd, 1, 0, 0
  1191. AVX_INSTR haddps, 1, 0, 0
  1192. AVX_INSTR hsubpd, 1, 0, 0
  1193. AVX_INSTR hsubps, 1, 0, 0
  1194. AVX_INSTR insertps, 1, 1, 0
  1195. AVX_INSTR lddqu
  1196. AVX_INSTR ldmxcsr
  1197. AVX_INSTR maskmovdqu
  1198. AVX_INSTR maxpd, 1, 0, 1
  1199. AVX_INSTR maxps, 1, 0, 1
  1200. AVX_INSTR maxsd, 1, 0, 1
  1201. AVX_INSTR maxss, 1, 0, 1
  1202. AVX_INSTR minpd, 1, 0, 1
  1203. AVX_INSTR minps, 1, 0, 1
  1204. AVX_INSTR minsd, 1, 0, 1
  1205. AVX_INSTR minss, 1, 0, 1
  1206. AVX_INSTR movapd
  1207. AVX_INSTR movaps
  1208. AVX_INSTR movd
  1209. AVX_INSTR movddup
  1210. AVX_INSTR movdqa
  1211. AVX_INSTR movdqu
  1212. AVX_INSTR movhlps, 1, 0, 0
  1213. AVX_INSTR movhpd, 1, 0, 0
  1214. AVX_INSTR movhps, 1, 0, 0
  1215. AVX_INSTR movlhps, 1, 0, 0
  1216. AVX_INSTR movlpd, 1, 0, 0
  1217. AVX_INSTR movlps, 1, 0, 0
  1218. AVX_INSTR movmskpd
  1219. AVX_INSTR movmskps
  1220. AVX_INSTR movntdq
  1221. AVX_INSTR movntdqa
  1222. AVX_INSTR movntpd
  1223. AVX_INSTR movntps
  1224. AVX_INSTR movq
  1225. AVX_INSTR movsd, 1, 0, 0
  1226. AVX_INSTR movshdup
  1227. AVX_INSTR movsldup
  1228. AVX_INSTR movss, 1, 0, 0
  1229. AVX_INSTR movupd
  1230. AVX_INSTR movups
  1231. AVX_INSTR mpsadbw, 0, 1, 0
  1232. AVX_INSTR mulpd, 1, 0, 1
  1233. AVX_INSTR mulps, 1, 0, 1
  1234. AVX_INSTR mulsd, 1, 0, 1
  1235. AVX_INSTR mulss, 1, 0, 1
  1236. AVX_INSTR orpd, 1, 0, 1
  1237. AVX_INSTR orps, 1, 0, 1
  1238. AVX_INSTR pabsb
  1239. AVX_INSTR pabsd
  1240. AVX_INSTR pabsw
  1241. AVX_INSTR packsswb, 0, 0, 0
  1242. AVX_INSTR packssdw, 0, 0, 0
  1243. AVX_INSTR packuswb, 0, 0, 0
  1244. AVX_INSTR packusdw, 0, 0, 0
  1245. AVX_INSTR paddb, 0, 0, 1
  1246. AVX_INSTR paddw, 0, 0, 1
  1247. AVX_INSTR paddd, 0, 0, 1
  1248. AVX_INSTR paddq, 0, 0, 1
  1249. AVX_INSTR paddsb, 0, 0, 1
  1250. AVX_INSTR paddsw, 0, 0, 1
  1251. AVX_INSTR paddusb, 0, 0, 1
  1252. AVX_INSTR paddusw, 0, 0, 1
  1253. AVX_INSTR palignr, 0, 1, 0
  1254. AVX_INSTR pand, 0, 0, 1
  1255. AVX_INSTR pandn, 0, 0, 0
  1256. AVX_INSTR pavgb, 0, 0, 1
  1257. AVX_INSTR pavgw, 0, 0, 1
  1258. AVX_INSTR pblendvb, 0, 0, 0
  1259. AVX_INSTR pblendw, 0, 1, 0
  1260. AVX_INSTR pclmulqdq, 0, 1, 0
  1261. AVX_INSTR pcmpestri
  1262. AVX_INSTR pcmpestrm
  1263. AVX_INSTR pcmpistri
  1264. AVX_INSTR pcmpistrm
  1265. AVX_INSTR pcmpeqb, 0, 0, 1
  1266. AVX_INSTR pcmpeqw, 0, 0, 1
  1267. AVX_INSTR pcmpeqd, 0, 0, 1
  1268. AVX_INSTR pcmpeqq, 0, 0, 1
  1269. AVX_INSTR pcmpgtb, 0, 0, 0
  1270. AVX_INSTR pcmpgtw, 0, 0, 0
  1271. AVX_INSTR pcmpgtd, 0, 0, 0
  1272. AVX_INSTR pcmpgtq, 0, 0, 0
  1273. AVX_INSTR pextrb
  1274. AVX_INSTR pextrd
  1275. AVX_INSTR pextrq
  1276. AVX_INSTR pextrw
  1277. AVX_INSTR phaddw, 0, 0, 0
  1278. AVX_INSTR phaddd, 0, 0, 0
  1279. AVX_INSTR phaddsw, 0, 0, 0
  1280. AVX_INSTR phminposuw
  1281. AVX_INSTR phsubw, 0, 0, 0
  1282. AVX_INSTR phsubd, 0, 0, 0
  1283. AVX_INSTR phsubsw, 0, 0, 0
  1284. AVX_INSTR pinsrb, 0, 1, 0
  1285. AVX_INSTR pinsrd, 0, 1, 0
  1286. AVX_INSTR pinsrq, 0, 1, 0
  1287. AVX_INSTR pinsrw, 0, 1, 0
  1288. AVX_INSTR pmaddwd, 0, 0, 1
  1289. AVX_INSTR pmaddubsw, 0, 0, 0
  1290. AVX_INSTR pmaxsb, 0, 0, 1
  1291. AVX_INSTR pmaxsw, 0, 0, 1
  1292. AVX_INSTR pmaxsd, 0, 0, 1
  1293. AVX_INSTR pmaxub, 0, 0, 1
  1294. AVX_INSTR pmaxuw, 0, 0, 1
  1295. AVX_INSTR pmaxud, 0, 0, 1
  1296. AVX_INSTR pminsb, 0, 0, 1
  1297. AVX_INSTR pminsw, 0, 0, 1
  1298. AVX_INSTR pminsd, 0, 0, 1
  1299. AVX_INSTR pminub, 0, 0, 1
  1300. AVX_INSTR pminuw, 0, 0, 1
  1301. AVX_INSTR pminud, 0, 0, 1
  1302. AVX_INSTR pmovmskb
  1303. AVX_INSTR pmovsxbw
  1304. AVX_INSTR pmovsxbd
  1305. AVX_INSTR pmovsxbq
  1306. AVX_INSTR pmovsxwd
  1307. AVX_INSTR pmovsxwq
  1308. AVX_INSTR pmovsxdq
  1309. AVX_INSTR pmovzxbw
  1310. AVX_INSTR pmovzxbd
  1311. AVX_INSTR pmovzxbq
  1312. AVX_INSTR pmovzxwd
  1313. AVX_INSTR pmovzxwq
  1314. AVX_INSTR pmovzxdq
  1315. AVX_INSTR pmuldq, 0, 0, 1
  1316. AVX_INSTR pmulhrsw, 0, 0, 1
  1317. AVX_INSTR pmulhuw, 0, 0, 1
  1318. AVX_INSTR pmulhw, 0, 0, 1
  1319. AVX_INSTR pmullw, 0, 0, 1
  1320. AVX_INSTR pmulld, 0, 0, 1
  1321. AVX_INSTR pmuludq, 0, 0, 1
  1322. AVX_INSTR por, 0, 0, 1
  1323. AVX_INSTR psadbw, 0, 0, 1
  1324. AVX_INSTR pshufb, 0, 0, 0
  1325. AVX_INSTR pshufd
  1326. AVX_INSTR pshufhw
  1327. AVX_INSTR pshuflw
  1328. AVX_INSTR psignb, 0, 0, 0
  1329. AVX_INSTR psignw, 0, 0, 0
  1330. AVX_INSTR psignd, 0, 0, 0
  1331. AVX_INSTR psllw, 0, 0, 0
  1332. AVX_INSTR pslld, 0, 0, 0
  1333. AVX_INSTR psllq, 0, 0, 0
  1334. AVX_INSTR pslldq, 0, 0, 0
  1335. AVX_INSTR psraw, 0, 0, 0
  1336. AVX_INSTR psrad, 0, 0, 0
  1337. AVX_INSTR psrlw, 0, 0, 0
  1338. AVX_INSTR psrld, 0, 0, 0
  1339. AVX_INSTR psrlq, 0, 0, 0
  1340. AVX_INSTR psrldq, 0, 0, 0
  1341. AVX_INSTR psubb, 0, 0, 0
  1342. AVX_INSTR psubw, 0, 0, 0
  1343. AVX_INSTR psubd, 0, 0, 0
  1344. AVX_INSTR psubq, 0, 0, 0
  1345. AVX_INSTR psubsb, 0, 0, 0
  1346. AVX_INSTR psubsw, 0, 0, 0
  1347. AVX_INSTR psubusb, 0, 0, 0
  1348. AVX_INSTR psubusw, 0, 0, 0
  1349. AVX_INSTR ptest
  1350. AVX_INSTR punpckhbw, 0, 0, 0
  1351. AVX_INSTR punpckhwd, 0, 0, 0
  1352. AVX_INSTR punpckhdq, 0, 0, 0
  1353. AVX_INSTR punpckhqdq, 0, 0, 0
  1354. AVX_INSTR punpcklbw, 0, 0, 0
  1355. AVX_INSTR punpcklwd, 0, 0, 0
  1356. AVX_INSTR punpckldq, 0, 0, 0
  1357. AVX_INSTR punpcklqdq, 0, 0, 0
  1358. AVX_INSTR pxor, 0, 0, 1
  1359. AVX_INSTR rcpps, 1, 0, 0
  1360. AVX_INSTR rcpss, 1, 0, 0
  1361. AVX_INSTR roundpd
  1362. AVX_INSTR roundps
  1363. AVX_INSTR roundsd
  1364. AVX_INSTR roundss
  1365. AVX_INSTR rsqrtps, 1, 0, 0
  1366. AVX_INSTR rsqrtss, 1, 0, 0
  1367. AVX_INSTR shufpd, 1, 1, 0
  1368. AVX_INSTR shufps, 1, 1, 0
  1369. AVX_INSTR sqrtpd, 1, 0, 0
  1370. AVX_INSTR sqrtps, 1, 0, 0
  1371. AVX_INSTR sqrtsd, 1, 0, 0
  1372. AVX_INSTR sqrtss, 1, 0, 0
  1373. AVX_INSTR stmxcsr
  1374. AVX_INSTR subpd, 1, 0, 0
  1375. AVX_INSTR subps, 1, 0, 0
  1376. AVX_INSTR subsd, 1, 0, 0
  1377. AVX_INSTR subss, 1, 0, 0
  1378. AVX_INSTR ucomisd
  1379. AVX_INSTR ucomiss
  1380. AVX_INSTR unpckhpd, 1, 0, 0
  1381. AVX_INSTR unpckhps, 1, 0, 0
  1382. AVX_INSTR unpcklpd, 1, 0, 0
  1383. AVX_INSTR unpcklps, 1, 0, 0
  1384. AVX_INSTR xorpd, 1, 0, 1
  1385. AVX_INSTR xorps, 1, 0, 1
  1386.  
  1387. ; 3DNow instructions, for sharing code between AVX, SSE and 3DN
  1388. AVX_INSTR pfadd, 1, 0, 1
  1389. AVX_INSTR pfsub, 1, 0, 0
  1390. AVX_INSTR pfmul, 1, 0, 1
  1391.  
  1392. ; base-4 constants for shuffles
  1393. %assign i 0
  1394. %rep 256
  1395.     %assign j ((i>>6)&3)*1000 + ((i>>4)&3)*100 + ((i>>2)&3)*10 + (i&3)
  1396.     %if j < 10
  1397.         CAT_XDEFINE q000, j, i
  1398.     %elif j < 100
  1399.         CAT_XDEFINE q00, j, i
  1400.     %elif j < 1000
  1401.         CAT_XDEFINE q0, j, i
  1402.     %else
  1403.         CAT_XDEFINE q, j, i
  1404.     %endif
  1405. %assign i i+1
  1406. %endrep
  1407. %undef i
  1408. %undef j
  1409.  
  1410. %macro FMA_INSTR 3
  1411.     %macro %1 4-7 %1, %2, %3
  1412.         %if cpuflag(xop)
  1413.             v%5 %1, %2, %3, %4
  1414.         %else
  1415.             %6 %1, %2, %3
  1416.             %7 %1, %4
  1417.         %endif
  1418.     %endmacro
  1419. %endmacro
  1420.  
  1421. FMA_INSTR  pmacsdd,  pmulld, paddd
  1422. FMA_INSTR  pmacsww,  pmullw, paddw
  1423. FMA_INSTR pmadcswd, pmaddwd, paddd
  1424.  
  1425. ; tzcnt is equivalent to "rep bsf" and is backwards-compatible with bsf.
  1426. ; This lets us use tzcnt without bumping the yasm version requirement yet.
  1427. %define tzcnt rep bsf
  1428.  
  1429. ; convert FMA4 to FMA3 if possible
  1430. %macro FMA4_INSTR 4
  1431.     %macro %1 4-8 %1, %2, %3, %4
  1432.         %if cpuflag(fma4)
  1433.             v%5 %1, %2, %3, %4
  1434.         %elifidn %1, %2
  1435.             v%6 %1, %4, %3 ; %1 = %1 * %3 + %4
  1436.         %elifidn %1, %3
  1437.             v%7 %1, %2, %4 ; %1 = %2 * %1 + %4
  1438.         %elifidn %1, %4
  1439.             v%8 %1, %2, %3 ; %1 = %2 * %3 + %1
  1440.         %else
  1441.             %error fma3 emulation of ``%5 %1, %2, %3, %4'' is not supported
  1442.         %endif
  1443.     %endmacro
  1444. %endmacro
  1445.  
  1446. FMA4_INSTR fmaddpd, fmadd132pd, fmadd213pd, fmadd231pd
  1447. FMA4_INSTR fmaddps, fmadd132ps, fmadd213ps, fmadd231ps
  1448. FMA4_INSTR fmaddsd, fmadd132sd, fmadd213sd, fmadd231sd
  1449. FMA4_INSTR fmaddss, fmadd132ss, fmadd213ss, fmadd231ss
  1450.  
  1451. FMA4_INSTR fmaddsubpd, fmaddsub132pd, fmaddsub213pd, fmaddsub231pd
  1452. FMA4_INSTR fmaddsubps, fmaddsub132ps, fmaddsub213ps, fmaddsub231ps
  1453. FMA4_INSTR fmsubaddpd, fmsubadd132pd, fmsubadd213pd, fmsubadd231pd
  1454. FMA4_INSTR fmsubaddps, fmsubadd132ps, fmsubadd213ps, fmsubadd231ps
  1455.  
  1456. FMA4_INSTR fmsubpd, fmsub132pd, fmsub213pd, fmsub231pd
  1457. FMA4_INSTR fmsubps, fmsub132ps, fmsub213ps, fmsub231ps
  1458. FMA4_INSTR fmsubsd, fmsub132sd, fmsub213sd, fmsub231sd
  1459. FMA4_INSTR fmsubss, fmsub132ss, fmsub213ss, fmsub231ss
  1460.  
  1461. FMA4_INSTR fnmaddpd, fnmadd132pd, fnmadd213pd, fnmadd231pd
  1462. FMA4_INSTR fnmaddps, fnmadd132ps, fnmadd213ps, fnmadd231ps
  1463. FMA4_INSTR fnmaddsd, fnmadd132sd, fnmadd213sd, fnmadd231sd
  1464. FMA4_INSTR fnmaddss, fnmadd132ss, fnmadd213ss, fnmadd231ss
  1465.  
  1466. FMA4_INSTR fnmsubpd, fnmsub132pd, fnmsub213pd, fnmsub231pd
  1467. FMA4_INSTR fnmsubps, fnmsub132ps, fnmsub213ps, fnmsub231ps
  1468. FMA4_INSTR fnmsubsd, fnmsub132sd, fnmsub213sd, fnmsub231sd
  1469. FMA4_INSTR fnmsubss, fnmsub132ss, fnmsub213ss, fnmsub231ss
  1470.  
  1471. ; workaround: vpbroadcastq is broken in x86_32 due to a yasm bug
  1472. %if ARCH_X86_64 == 0
  1473. %macro vpbroadcastq 2
  1474. %if sizeof%1 == 16
  1475.     movddup %1, %2
  1476. %else
  1477.     vbroadcastsd %1, %2
  1478. %endif
  1479. %endmacro
  1480. %endif
  1481.