/dports/lang/gcc6-aux/gcc-6-20180516/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 27 movdqu %xmm0, xmm_regs+0(%rip) 28 movdqu %xmm1, xmm_regs+16(%rip) 29 movdqu %xmm2, xmm_regs+32(%rip) 30 movdqu %xmm3, xmm_regs+48(%rip) 31 movdqu %xmm4, xmm_regs+64(%rip) 32 movdqu %xmm5, xmm_regs+80(%rip) 33 movdqu %xmm6, xmm_regs+96(%rip) 34 movdqu %xmm7, xmm_regs+112(%rip) 35 movdqu %xmm8, xmm_regs+128(%rip) 57 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/lang/gcc12-devel/gcc-12-20211205/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 26 movdqu %xmm0, xmm_regs+0(%rip) 27 movdqu %xmm1, xmm_regs+16(%rip) 28 movdqu %xmm2, xmm_regs+32(%rip) 29 movdqu %xmm3, xmm_regs+48(%rip) 30 movdqu %xmm4, xmm_regs+64(%rip) 31 movdqu %xmm5, xmm_regs+80(%rip) 32 movdqu %xmm6, xmm_regs+96(%rip) 33 movdqu %xmm7, xmm_regs+112(%rip) 34 movdqu %xmm8, xmm_regs+128(%rip) 56 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/lang/gcc12-devel/gcc-12-20211205/gcc/testsuite/gcc.target/x86_64/abi/avx512fp16/ |
H A D | asm-support.S | 23 vmovdqu %xmm0, xmm_regs+0(%rip) 24 vmovdqu %xmm1, xmm_regs+16(%rip) 25 vmovdqu %xmm2, xmm_regs+32(%rip) 26 vmovdqu %xmm3, xmm_regs+48(%rip) 27 vmovdqu %xmm4, xmm_regs+64(%rip) 28 vmovdqu %xmm5, xmm_regs+80(%rip) 29 vmovdqu %xmm6, xmm_regs+96(%rip) 30 vmovdqu %xmm7, xmm_regs+112(%rip) 31 vmovdqu %xmm8, xmm_regs+128(%rip) 53 vmovdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/devel/riscv64-none-elf-gcc/gcc-8.4.0/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 27 movdqu %xmm0, xmm_regs+0(%rip) 28 movdqu %xmm1, xmm_regs+16(%rip) 29 movdqu %xmm2, xmm_regs+32(%rip) 30 movdqu %xmm3, xmm_regs+48(%rip) 31 movdqu %xmm4, xmm_regs+64(%rip) 32 movdqu %xmm5, xmm_regs+80(%rip) 33 movdqu %xmm6, xmm_regs+96(%rip) 34 movdqu %xmm7, xmm_regs+112(%rip) 35 movdqu %xmm8, xmm_regs+128(%rip) 57 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/lang/gcc10/gcc-10.3.0/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 26 movdqu %xmm0, xmm_regs+0(%rip) 27 movdqu %xmm1, xmm_regs+16(%rip) 28 movdqu %xmm2, xmm_regs+32(%rip) 29 movdqu %xmm3, xmm_regs+48(%rip) 30 movdqu %xmm4, xmm_regs+64(%rip) 31 movdqu %xmm5, xmm_regs+80(%rip) 32 movdqu %xmm6, xmm_regs+96(%rip) 33 movdqu %xmm7, xmm_regs+112(%rip) 34 movdqu %xmm8, xmm_regs+128(%rip) 56 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/lang/gcc11/gcc-11.2.0/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 26 movdqu %xmm0, xmm_regs+0(%rip) 27 movdqu %xmm1, xmm_regs+16(%rip) 28 movdqu %xmm2, xmm_regs+32(%rip) 29 movdqu %xmm3, xmm_regs+48(%rip) 30 movdqu %xmm4, xmm_regs+64(%rip) 31 movdqu %xmm5, xmm_regs+80(%rip) 32 movdqu %xmm6, xmm_regs+96(%rip) 33 movdqu %xmm7, xmm_regs+112(%rip) 34 movdqu %xmm8, xmm_regs+128(%rip) 56 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/devel/arm-none-eabi-gcc/gcc-8.4.0/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 27 movdqu %xmm0, xmm_regs+0(%rip) 28 movdqu %xmm1, xmm_regs+16(%rip) 29 movdqu %xmm2, xmm_regs+32(%rip) 30 movdqu %xmm3, xmm_regs+48(%rip) 31 movdqu %xmm4, xmm_regs+64(%rip) 32 movdqu %xmm5, xmm_regs+80(%rip) 33 movdqu %xmm6, xmm_regs+96(%rip) 34 movdqu %xmm7, xmm_regs+112(%rip) 35 movdqu %xmm8, xmm_regs+128(%rip) 57 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/devel/riscv32-unknown-elf-gcc/gcc-8.4.0/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 27 movdqu %xmm0, xmm_regs+0(%rip) 28 movdqu %xmm1, xmm_regs+16(%rip) 29 movdqu %xmm2, xmm_regs+32(%rip) 30 movdqu %xmm3, xmm_regs+48(%rip) 31 movdqu %xmm4, xmm_regs+64(%rip) 32 movdqu %xmm5, xmm_regs+80(%rip) 33 movdqu %xmm6, xmm_regs+96(%rip) 34 movdqu %xmm7, xmm_regs+112(%rip) 35 movdqu %xmm8, xmm_regs+128(%rip) 57 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/devel/aarch64-none-elf-gcc/gcc-8.4.0/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 27 movdqu %xmm0, xmm_regs+0(%rip) 28 movdqu %xmm1, xmm_regs+16(%rip) 29 movdqu %xmm2, xmm_regs+32(%rip) 30 movdqu %xmm3, xmm_regs+48(%rip) 31 movdqu %xmm4, xmm_regs+64(%rip) 32 movdqu %xmm5, xmm_regs+80(%rip) 33 movdqu %xmm6, xmm_regs+96(%rip) 34 movdqu %xmm7, xmm_regs+112(%rip) 35 movdqu %xmm8, xmm_regs+128(%rip) 57 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/lang/gcc9-aux/gcc-9.1.0/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 26 movdqu %xmm0, xmm_regs+0(%rip) 27 movdqu %xmm1, xmm_regs+16(%rip) 28 movdqu %xmm2, xmm_regs+32(%rip) 29 movdqu %xmm3, xmm_regs+48(%rip) 30 movdqu %xmm4, xmm_regs+64(%rip) 31 movdqu %xmm5, xmm_regs+80(%rip) 32 movdqu %xmm6, xmm_regs+96(%rip) 33 movdqu %xmm7, xmm_regs+112(%rip) 34 movdqu %xmm8, xmm_regs+128(%rip) 56 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/lang/gcc48/gcc-4.8.5/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 27 movdqu %xmm0, xmm_regs+0(%rip) 28 movdqu %xmm1, xmm_regs+16(%rip) 29 movdqu %xmm2, xmm_regs+32(%rip) 30 movdqu %xmm3, xmm_regs+48(%rip) 31 movdqu %xmm4, xmm_regs+64(%rip) 32 movdqu %xmm5, xmm_regs+80(%rip) 33 movdqu %xmm6, xmm_regs+96(%rip) 34 movdqu %xmm7, xmm_regs+112(%rip) 35 movdqu %xmm8, xmm_regs+128(%rip) 57 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/devel/arm-none-eabi-gcc492/gcc-4.9.2/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 27 movdqu %xmm0, xmm_regs+0(%rip) 28 movdqu %xmm1, xmm_regs+16(%rip) 29 movdqu %xmm2, xmm_regs+32(%rip) 30 movdqu %xmm3, xmm_regs+48(%rip) 31 movdqu %xmm4, xmm_regs+64(%rip) 32 movdqu %xmm5, xmm_regs+80(%rip) 33 movdqu %xmm6, xmm_regs+96(%rip) 34 movdqu %xmm7, xmm_regs+112(%rip) 35 movdqu %xmm8, xmm_regs+128(%rip) 57 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/lang/gcc9-devel/gcc-9-20211007/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 26 movdqu %xmm0, xmm_regs+0(%rip) 27 movdqu %xmm1, xmm_regs+16(%rip) 28 movdqu %xmm2, xmm_regs+32(%rip) 29 movdqu %xmm3, xmm_regs+48(%rip) 30 movdqu %xmm4, xmm_regs+64(%rip) 31 movdqu %xmm5, xmm_regs+80(%rip) 32 movdqu %xmm6, xmm_regs+96(%rip) 33 movdqu %xmm7, xmm_regs+112(%rip) 34 movdqu %xmm8, xmm_regs+128(%rip) 56 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/lang/gcc11-devel/gcc-11-20211009/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 26 movdqu %xmm0, xmm_regs+0(%rip) 27 movdqu %xmm1, xmm_regs+16(%rip) 28 movdqu %xmm2, xmm_regs+32(%rip) 29 movdqu %xmm3, xmm_regs+48(%rip) 30 movdqu %xmm4, xmm_regs+64(%rip) 31 movdqu %xmm5, xmm_regs+80(%rip) 32 movdqu %xmm6, xmm_regs+96(%rip) 33 movdqu %xmm7, xmm_regs+112(%rip) 34 movdqu %xmm8, xmm_regs+128(%rip) 56 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/misc/cxx_atomics_pic/gcc-11.2.0/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 26 movdqu %xmm0, xmm_regs+0(%rip) 27 movdqu %xmm1, xmm_regs+16(%rip) 28 movdqu %xmm2, xmm_regs+32(%rip) 29 movdqu %xmm3, xmm_regs+48(%rip) 30 movdqu %xmm4, xmm_regs+64(%rip) 31 movdqu %xmm5, xmm_regs+80(%rip) 32 movdqu %xmm6, xmm_regs+96(%rip) 33 movdqu %xmm7, xmm_regs+112(%rip) 34 movdqu %xmm8, xmm_regs+128(%rip) 56 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/lang/gcc8/gcc-8.5.0/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 27 movdqu %xmm0, xmm_regs+0(%rip) 28 movdqu %xmm1, xmm_regs+16(%rip) 29 movdqu %xmm2, xmm_regs+32(%rip) 30 movdqu %xmm3, xmm_regs+48(%rip) 31 movdqu %xmm4, xmm_regs+64(%rip) 32 movdqu %xmm5, xmm_regs+80(%rip) 33 movdqu %xmm6, xmm_regs+96(%rip) 34 movdqu %xmm7, xmm_regs+112(%rip) 35 movdqu %xmm8, xmm_regs+128(%rip) 57 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/lang/gnat_util/gcc-6-20180516/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 27 movdqu %xmm0, xmm_regs+0(%rip) 28 movdqu %xmm1, xmm_regs+16(%rip) 29 movdqu %xmm2, xmm_regs+32(%rip) 30 movdqu %xmm3, xmm_regs+48(%rip) 31 movdqu %xmm4, xmm_regs+64(%rip) 32 movdqu %xmm5, xmm_regs+80(%rip) 33 movdqu %xmm6, xmm_regs+96(%rip) 34 movdqu %xmm7, xmm_regs+112(%rip) 35 movdqu %xmm8, xmm_regs+128(%rip) 57 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/devel/avr-gcc/gcc-10.2.0/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 26 movdqu %xmm0, xmm_regs+0(%rip) 27 movdqu %xmm1, xmm_regs+16(%rip) 28 movdqu %xmm2, xmm_regs+32(%rip) 29 movdqu %xmm3, xmm_regs+48(%rip) 30 movdqu %xmm4, xmm_regs+64(%rip) 31 movdqu %xmm5, xmm_regs+80(%rip) 32 movdqu %xmm6, xmm_regs+96(%rip) 33 movdqu %xmm7, xmm_regs+112(%rip) 34 movdqu %xmm8, xmm_regs+128(%rip) 56 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/devel/riscv64-gcc/gcc-8.3.0/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 27 movdqu %xmm0, xmm_regs+0(%rip) 28 movdqu %xmm1, xmm_regs+16(%rip) 29 movdqu %xmm2, xmm_regs+32(%rip) 30 movdqu %xmm3, xmm_regs+48(%rip) 31 movdqu %xmm4, xmm_regs+64(%rip) 32 movdqu %xmm5, xmm_regs+80(%rip) 33 movdqu %xmm6, xmm_regs+96(%rip) 34 movdqu %xmm7, xmm_regs+112(%rip) 35 movdqu %xmm8, xmm_regs+128(%rip) 57 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/lang/gcc9/gcc-9.4.0/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 26 movdqu %xmm0, xmm_regs+0(%rip) 27 movdqu %xmm1, xmm_regs+16(%rip) 28 movdqu %xmm2, xmm_regs+32(%rip) 29 movdqu %xmm3, xmm_regs+48(%rip) 30 movdqu %xmm4, xmm_regs+64(%rip) 31 movdqu %xmm5, xmm_regs+80(%rip) 32 movdqu %xmm6, xmm_regs+96(%rip) 33 movdqu %xmm7, xmm_regs+112(%rip) 34 movdqu %xmm8, xmm_regs+128(%rip) 56 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/lang/gcc10-devel/gcc-10-20211008/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 26 movdqu %xmm0, xmm_regs+0(%rip) 27 movdqu %xmm1, xmm_regs+16(%rip) 28 movdqu %xmm2, xmm_regs+32(%rip) 29 movdqu %xmm3, xmm_regs+48(%rip) 30 movdqu %xmm4, xmm_regs+64(%rip) 31 movdqu %xmm5, xmm_regs+80(%rip) 32 movdqu %xmm6, xmm_regs+96(%rip) 33 movdqu %xmm7, xmm_regs+112(%rip) 34 movdqu %xmm8, xmm_regs+128(%rip) 56 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/devel/mingw32-gcc/gcc-4.8.1/gcc/testsuite/gcc.target/x86_64/abi/ |
H A D | asm-support.S | 27 movdqu %xmm0, xmm_regs+0(%rip) 28 movdqu %xmm1, xmm_regs+16(%rip) 29 movdqu %xmm2, xmm_regs+32(%rip) 30 movdqu %xmm3, xmm_regs+48(%rip) 31 movdqu %xmm4, xmm_regs+64(%rip) 32 movdqu %xmm5, xmm_regs+80(%rip) 33 movdqu %xmm6, xmm_regs+96(%rip) 34 movdqu %xmm7, xmm_regs+112(%rip) 35 movdqu %xmm8, xmm_regs+128(%rip) 57 movdqu %xmm0, xmm_regs+0(%rip) [all …]
|
/dports/emulators/qemu-utils/qemu-4.2.1/target/i386/ |
H A D | xsave_helper.c | 41 uint8_t *xmm = xsave->legacy.xmm_regs[i]; in x86_cpu_xsave_all_areas() 44 stq_p(xmm, env->xmm_regs[i].ZMM_Q(0)); in x86_cpu_xsave_all_areas() 45 stq_p(xmm+8, env->xmm_regs[i].ZMM_Q(1)); in x86_cpu_xsave_all_areas() 46 stq_p(ymmh, env->xmm_regs[i].ZMM_Q(2)); in x86_cpu_xsave_all_areas() 47 stq_p(ymmh+8, env->xmm_regs[i].ZMM_Q(3)); in x86_cpu_xsave_all_areas() 48 stq_p(zmmh, env->xmm_regs[i].ZMM_Q(4)); in x86_cpu_xsave_all_areas() 56 16 * sizeof env->xmm_regs[16]); in x86_cpu_xsave_all_areas() 96 env->xmm_regs[i].ZMM_Q(0) = ldq_p(xmm); in x86_cpu_xrstor_all_areas() 98 env->xmm_regs[i].ZMM_Q(2) = ldq_p(ymmh); in x86_cpu_xrstor_all_areas() 100 env->xmm_regs[i].ZMM_Q(4) = ldq_p(zmmh); in x86_cpu_xrstor_all_areas() [all …]
|
/dports/emulators/qemu5/qemu-5.2.0/target/i386/ |
H A D | xsave_helper.c | 41 uint8_t *xmm = xsave->legacy.xmm_regs[i]; in x86_cpu_xsave_all_areas() 44 stq_p(xmm, env->xmm_regs[i].ZMM_Q(0)); in x86_cpu_xsave_all_areas() 45 stq_p(xmm+8, env->xmm_regs[i].ZMM_Q(1)); in x86_cpu_xsave_all_areas() 46 stq_p(ymmh, env->xmm_regs[i].ZMM_Q(2)); in x86_cpu_xsave_all_areas() 47 stq_p(ymmh+8, env->xmm_regs[i].ZMM_Q(3)); in x86_cpu_xsave_all_areas() 48 stq_p(zmmh, env->xmm_regs[i].ZMM_Q(4)); in x86_cpu_xsave_all_areas() 56 16 * sizeof env->xmm_regs[16]); in x86_cpu_xsave_all_areas() 96 env->xmm_regs[i].ZMM_Q(0) = ldq_p(xmm); in x86_cpu_xrstor_all_areas() 98 env->xmm_regs[i].ZMM_Q(2) = ldq_p(ymmh); in x86_cpu_xrstor_all_areas() 100 env->xmm_regs[i].ZMM_Q(4) = ldq_p(zmmh); in x86_cpu_xrstor_all_areas() [all …]
|
/dports/emulators/qemu-guest-agent/qemu-5.0.1/target/i386/ |
H A D | xsave_helper.c | 41 uint8_t *xmm = xsave->legacy.xmm_regs[i]; in x86_cpu_xsave_all_areas() 44 stq_p(xmm, env->xmm_regs[i].ZMM_Q(0)); in x86_cpu_xsave_all_areas() 45 stq_p(xmm+8, env->xmm_regs[i].ZMM_Q(1)); in x86_cpu_xsave_all_areas() 46 stq_p(ymmh, env->xmm_regs[i].ZMM_Q(2)); in x86_cpu_xsave_all_areas() 47 stq_p(ymmh+8, env->xmm_regs[i].ZMM_Q(3)); in x86_cpu_xsave_all_areas() 48 stq_p(zmmh, env->xmm_regs[i].ZMM_Q(4)); in x86_cpu_xsave_all_areas() 56 16 * sizeof env->xmm_regs[16]); in x86_cpu_xsave_all_areas() 96 env->xmm_regs[i].ZMM_Q(0) = ldq_p(xmm); in x86_cpu_xrstor_all_areas() 98 env->xmm_regs[i].ZMM_Q(2) = ldq_p(ymmh); in x86_cpu_xrstor_all_areas() 100 env->xmm_regs[i].ZMM_Q(4) = ldq_p(zmmh); in x86_cpu_xrstor_all_areas() [all …]
|