summaryrefslogtreecommitdiff
path: root/sysdeps/x86_64/fpu/svml_d_sincos8_core.S
diff options
context:
space:
mode:
authorZack Weinberg <zackw@panix.com>2017-06-08 15:39:03 -0400
committerZack Weinberg <zackw@panix.com>2017-06-08 15:39:03 -0400
commit5046dbb4a7eba5eccfd258f92f4735c9ffc8d069 (patch)
tree4470480d904b65cf14ca524f96f79eca818c3eaf /sysdeps/x86_64/fpu/svml_d_sincos8_core.S
parent199fc19d3aaaf57944ef036e15904febe877fc93 (diff)
downloadglibc-zack/build-layout-experiment.tar.gz
Prepare for radical source tree reorganization.zack/build-layout-experiment
All top-level files and directories are moved into a temporary storage directory, REORG.TODO, except for files that will certainly still exist in their current form at top level when we're done (COPYING, COPYING.LIB, LICENSES, NEWS, README), all old ChangeLog files (which are moved to the new directory OldChangeLogs, instead), and the generated file INSTALL (which is just deleted; in the new order, there will be no generated files checked into version control).
Diffstat (limited to 'sysdeps/x86_64/fpu/svml_d_sincos8_core.S')
-rw-r--r--sysdeps/x86_64/fpu/svml_d_sincos8_core.S224
1 files changed, 0 insertions, 224 deletions
diff --git a/sysdeps/x86_64/fpu/svml_d_sincos8_core.S b/sysdeps/x86_64/fpu/svml_d_sincos8_core.S
deleted file mode 100644
index c104539821..0000000000
--- a/sysdeps/x86_64/fpu/svml_d_sincos8_core.S
+++ /dev/null
@@ -1,224 +0,0 @@
-/* Function sincos vectorized with AVX-512. Wrapper to AVX2 version.
- Copyright (C) 2014-2017 Free Software Foundation, Inc.
- This file is part of the GNU C Library.
-
- The GNU C Library is free software; you can redistribute it and/or
- modify it under the terms of the GNU Lesser General Public
- License as published by the Free Software Foundation; either
- version 2.1 of the License, or (at your option) any later version.
-
- The GNU C Library is distributed in the hope that it will be useful,
- but WITHOUT ANY WARRANTY; without even the implied warranty of
- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- Lesser General Public License for more details.
-
- You should have received a copy of the GNU Lesser General Public
- License along with the GNU C Library; if not, see
- <http://www.gnu.org/licenses/>. */
-
-#include <sysdep.h>
-#include "svml_d_wrapper_impl.h"
-
- .text
-ENTRY (_ZGVeN8vl8l8_sincos)
-WRAPPER_IMPL_AVX512_fFF _ZGVdN4vl8l8_sincos
-END (_ZGVeN8vl8l8_sincos)
-
-/* AVX512 ISA version as wrapper to AVX2 ISA version (for vector
- function declared with #pragma omp declare simd notinbranch). */
-.macro WRAPPER_IMPL_AVX512_fFF_vvv callee
-#ifndef __ILP32__
- pushq %rbp
- cfi_adjust_cfa_offset (8)
- cfi_rel_offset (%rbp, 0)
- movq %rsp, %rbp
- cfi_def_cfa_register (%rbp)
- andq $-64, %rsp
- subq $320, %rsp
- /* Encoding for vmovups %zmm0, 256(%rsp). */
- .byte 0x62
- .byte 0xf1
- .byte 0x7c
- .byte 0x48
- .byte 0x11
- .byte 0x44
- .byte 0x24
- .byte 0x04
- lea (%rsp), %rdi
- /* Encoding for vmovups %zmm1, 128(%rdi). */
- .byte 0x62
- .byte 0xf1
- .byte 0x7c
- .byte 0x48
- .byte 0x11
- .byte 0x4f
- .byte 0x02
- /* Encoding for vmovups %zmm2, 192(%rdi). */
- .byte 0x62
- .byte 0xf1
- .byte 0x7c
- .byte 0x48
- .byte 0x11
- .byte 0x57
- .byte 0x03
- lea 64(%rsp), %rsi
- call HIDDEN_JUMPTARGET(\callee)
- vmovdqu 288(%rsp), %ymm0
- lea 32(%rsp), %rdi
- lea 96(%rsp), %rsi
- call HIDDEN_JUMPTARGET(\callee)
- movq 128(%rsp), %rdx
- movq 192(%rsp), %rsi
- movq 136(%rsp), %r8
- movq 200(%rsp), %r10
- movq (%rsp), %rax
- movq 64(%rsp), %rcx
- movq 8(%rsp), %rdi
- movq 72(%rsp), %r9
- movq %rax, (%rdx)
- movq %rcx, (%rsi)
- movq 144(%rsp), %rax
- movq 208(%rsp), %rcx
- movq %rdi, (%r8)
- movq %r9, (%r10)
- movq 152(%rsp), %rdi
- movq 216(%rsp), %r9
- movq 16(%rsp), %r11
- movq 80(%rsp), %rdx
- movq 24(%rsp), %rsi
- movq 88(%rsp), %r8
- movq %r11, (%rax)
- movq %rdx, (%rcx)
- movq 160(%rsp), %r11
- movq 224(%rsp), %rdx
- movq %rsi, (%rdi)
- movq %r8, (%r9)
- movq 168(%rsp), %rsi
- movq 232(%rsp), %r8
- movq 32(%rsp), %r10
- movq 96(%rsp), %rax
- movq 40(%rsp), %rcx
- movq 104(%rsp), %rdi
- movq %r10, (%r11)
- movq %rax, (%rdx)
- movq 176(%rsp), %r10
- movq 240(%rsp), %rax
- movq %rcx, (%rsi)
- movq %rdi, (%r8)
- movq 184(%rsp), %rcx
- movq 248(%rsp), %rdi
- movq 48(%rsp), %r9
- movq 112(%rsp), %r11
- movq 56(%rsp), %rdx
- movq 120(%rsp), %rsi
- movq %r9, (%r10)
- movq %r11, (%rax)
- movq %rdx, (%rcx)
- movq %rsi, (%rdi)
- movq %rbp, %rsp
- cfi_def_cfa_register (%rsp)
- popq %rbp
- cfi_adjust_cfa_offset (-8)
- cfi_restore (%rbp)
- ret
-#else
- leal 8(%rsp), %r10d
- .cfi_def_cfa 10, 0
- andl $-64, %esp
- pushq -8(%r10d)
- pushq %rbp
- .cfi_escape 0x10,0x6,0x2,0x76,0
- movl %esp, %ebp
- pushq %r12
- leal -112(%rbp), %esi
- pushq %r10
- .cfi_escape 0xf,0x3,0x76,0x70,0x6
- .cfi_escape 0x10,0xc,0x2,0x76,0x78
- leal -176(%rbp), %edi
- movq %rsi, %r12
- pushq %rbx
- .cfi_escape 0x10,0x3,0x2,0x76,0x68
- movq %rdi, %rbx
- subl $280, %esp
- vmovdqa %ymm1, -208(%ebp)
- vmovdqa %ymm2, -240(%ebp)
- /* Encoding for vmovapd %zmm0, -304(%ebp). */
- .byte 0x67
- .byte 0x62
- .byte 0xf1
- .byte 0xfd
- .byte 0x48
- .byte 0x29
- .byte 0x85
- .byte 0xd0
- .byte 0xfe
- .byte 0xff
- .byte 0xff
- call HIDDEN_JUMPTARGET(\callee)
- leal 32(%r12), %esi
- vmovupd -272(%ebp), %ymm0
- leal 32(%rbx), %edi
- call HIDDEN_JUMPTARGET(\callee)
- movl -208(%ebp), %eax
- vmovsd -176(%ebp), %xmm0
- vmovsd %xmm0, (%eax)
- movl -204(%ebp), %eax
- vmovsd -168(%ebp), %xmm0
- vmovsd %xmm0, (%eax)
- movl -200(%ebp), %eax
- vmovsd -160(%ebp), %xmm0
- vmovsd %xmm0, (%eax)
- movl -196(%ebp), %eax
- vmovsd -152(%ebp), %xmm0
- vmovsd %xmm0, (%eax)
- movl -192(%ebp), %eax
- vmovsd -144(%ebp), %xmm0
- vmovsd %xmm0, (%eax)
- movl -188(%ebp), %eax
- vmovsd -136(%ebp), %xmm0
- vmovsd %xmm0, (%eax)
- movl -184(%ebp), %eax
- vmovsd -128(%ebp), %xmm0
- vmovsd %xmm0, (%eax)
- movl -180(%ebp), %eax
- vmovsd -120(%ebp), %xmm0
- vmovsd %xmm0, (%eax)
- movl -240(%ebp), %eax
- vmovsd -112(%ebp), %xmm0
- vmovsd %xmm0, (%eax)
- movl -236(%ebp), %eax
- vmovsd -104(%ebp), %xmm0
- vmovsd %xmm0, (%eax)
- movl -232(%ebp), %eax
- vmovsd -96(%ebp), %xmm0
- vmovsd %xmm0, (%eax)
- movl -228(%ebp), %eax
- vmovsd -88(%ebp), %xmm0
- vmovsd %xmm0, (%eax)
- movl -224(%ebp), %eax
- vmovsd -80(%ebp), %xmm0
- vmovsd %xmm0, (%eax)
- movl -220(%ebp), %eax
- vmovsd -72(%ebp), %xmm0
- vmovsd %xmm0, (%eax)
- movl -216(%ebp), %eax
- vmovsd -64(%ebp), %xmm0
- vmovsd %xmm0, (%eax)
- movl -212(%ebp), %eax
- vmovsd -56(%ebp), %xmm0
- vmovsd %xmm0, (%eax)
- addl $280, %esp
- popq %rbx
- popq %r10
- .cfi_def_cfa 10, 0
- popq %r12
- popq %rbp
- leal -8(%r10), %esp
- .cfi_def_cfa 7, 8
- ret
-#endif
-.endm
-
-ENTRY (_ZGVeN8vvv_sincos)
-WRAPPER_IMPL_AVX512_fFF_vvv _ZGVdN4vl8l8_sincos
-END (_ZGVeN8vvv_sincos)