Home
last modified time | relevance | path

Searched refs:__m128 (Results 1 – 25 of 29) sorted by relevance

12

/freebsd/contrib/llvm-project/clang/lib/Headers/
H A Dxmmintrin.h58 _mm_add_ss(__m128 __a, __m128 __b) in _mm_add_ss()
78 _mm_add_ps(__m128 __a, __m128 __b) in _mm_add_ps()
100 _mm_sub_ss(__m128 __a, __m128 __b) in _mm_sub_ss()
121 _mm_sub_ps(__m128 __a, __m128 __b) in _mm_sub_ps()
143 _mm_mul_ss(__m128 __a, __m128 __b) in _mm_mul_ss()
163 _mm_mul_ps(__m128 __a, __m128 __b) in _mm_mul_ps()
185 _mm_div_ss(__m128 __a, __m128 __b) in _mm_div_ss()
204 _mm_div_ps(__m128 __a, __m128 __b) in _mm_div_ps()
333 _mm_min_ss(__m128 __a, __m128 __b) in _mm_min_ss()
352 _mm_min_ps(__m128 __a, __m128 __b) in _mm_min_ps()
[all …]
H A Dfma4intrin.h23 static __inline__ __m128 __DEFAULT_FN_ATTRS128
24 _mm_macc_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_macc_ps()
36 _mm_macc_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_macc_ss()
48 _mm_msub_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_msub_ps()
60 _mm_msub_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_msub_ss()
72 _mm_nmacc_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_nmacc_ps()
84 _mm_nmacc_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_nmacc_ss()
96 _mm_nmsub_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_nmsub_ps()
108 _mm_nmsub_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_nmsub_ss()
120 _mm_maddsub_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_maddsub_ps()
[all …]
H A Dfmaintrin.h35 static __inline__ __m128 __DEFAULT_FN_ATTRS128
36 _mm_fmadd_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_fmadd_ps()
84 _mm_fmadd_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_fmadd_ss()
132 _mm_fmsub_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_fmsub_ps()
180 _mm_fmsub_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_fmsub_ss()
228 _mm_fnmadd_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_fnmadd_ps()
276 _mm_fnmadd_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_fnmadd_ss()
324 _mm_fnmsub_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_fnmsub_ps()
372 _mm_fnmsub_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_fnmsub_ss()
426 _mm_fmaddsub_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_fmaddsub_ps()
[all …]
H A Davx512vldqintrin.h108 _mm_mask_andnot_ps(__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) { in _mm_mask_andnot_ps()
115 _mm_maskz_andnot_ps(__mmask8 __U, __m128 __A, __m128 __B) { in _mm_maskz_andnot_ps()
164 _mm_mask_and_ps(__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) { in _mm_mask_and_ps()
171 _mm_maskz_and_ps(__mmask8 __U, __m128 __A, __m128 __B) { in _mm_maskz_and_ps()
220 _mm_mask_xor_ps(__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) { in _mm_mask_xor_ps()
227 _mm_maskz_xor_ps(__mmask8 __U, __m128 __A, __m128 __B) { in _mm_maskz_xor_ps()
276 _mm_mask_or_ps(__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) { in _mm_mask_or_ps()
283 _mm_maskz_or_ps(__mmask8 __U, __m128 __A, __m128 __B) { in _mm_maskz_or_ps()
818 ((__m128)__builtin_ia32_rangeps128_mask((__v4sf)(__m128)(A), \
824 ((__m128)__builtin_ia32_rangeps128_mask((__v4sf)(__m128)(A), \
[all …]
H A Davxneconvertintrin.h56 static __inline__ __m128 __DEFAULT_FN_ATTRS128
58 return (__m128)__builtin_ia32_vbcstnebf162ps128((const __bf16 *)__A); in _mm_bcstnebf16_ps()
122 static __inline__ __m128 __DEFAULT_FN_ATTRS128
124 return (__m128)__builtin_ia32_vbcstnesh2ps128((const _Float16 *)__A); in _mm_bcstnesh_ps()
188 static __inline__ __m128 __DEFAULT_FN_ATTRS128
254 static __inline__ __m128 __DEFAULT_FN_ATTRS128
256 return (__m128)__builtin_ia32_vcvtneeph2ps128((const __v8hf *)__A); in _mm_cvtneeph_ps()
320 static __inline__ __m128 __DEFAULT_FN_ATTRS128
386 static __inline__ __m128 __DEFAULT_FN_ATTRS128
388 return (__m128)__builtin_ia32_vcvtneoph2ps128((const __v8hf *)__A); in _mm_cvtneoph_ps()
[all …]
H A Dpmmintrin.h57 static __inline__ __m128 __DEFAULT_FN_ATTRS
58 _mm_addsub_ps(__m128 __a, __m128 __b) in _mm_addsub_ps()
80 static __inline__ __m128 __DEFAULT_FN_ATTRS
81 _mm_hadd_ps(__m128 __a, __m128 __b) in _mm_hadd_ps()
103 static __inline__ __m128 __DEFAULT_FN_ATTRS
104 _mm_hsub_ps(__m128 __a, __m128 __b) in _mm_hsub_ps()
125 static __inline__ __m128 __DEFAULT_FN_ATTRS
126 _mm_movehdup_ps(__m128 __a) in _mm_movehdup_ps()
146 static __inline__ __m128 __DEFAULT_FN_ATTRS
147 _mm_moveldup_ps(__m128 __a) in _mm_moveldup_ps()
H A Davx512vlbf16intrin.h40 _mm_cvtne2ps_pbh(__m128 __A, __m128 __B) { in _mm_cvtne2ps_pbh()
63 _mm_mask_cvtne2ps_pbh(__m128bh __W, __mmask8 __U, __m128 __A, __m128 __B) { in _mm_mask_cvtne2ps_pbh()
85 _mm_maskz_cvtne2ps_pbh(__mmask8 __U, __m128 __A, __m128 __B) { in _mm_maskz_cvtne2ps_pbh()
204 _mm_maskz_cvtneps_pbh(__mmask8 __U, __m128 __A) { in _mm_maskz_cvtneps_pbh()
276 static __inline__ __m128 __DEFAULT_FN_ATTRS128
277 _mm_dpbf16_ps(__m128 __D, __m128bh __A, __m128bh __B) { in _mm_dpbf16_ps()
278 return (__m128)__builtin_ia32_dpbf16ps_128((__v4sf)__D, in _mm_dpbf16_ps()
300 static __inline__ __m128 __DEFAULT_FN_ATTRS128
324 static __inline__ __m128 __DEFAULT_FN_ATTRS128
451 static __inline__ __m128 __DEFAULT_FN_ATTRS128
[all …]
H A Davx512erintrin.h115 ((__m128)__builtin_ia32_rsqrt28ss_round_mask((__v4sf)(__m128)(A), \
116 (__v4sf)(__m128)(B), \
121 ((__m128)__builtin_ia32_rsqrt28ss_round_mask((__v4sf)(__m128)(A), \
127 ((__m128)__builtin_ia32_rsqrt28ss_round_mask((__v4sf)(__m128)(A), \
218 ((__m128)__builtin_ia32_rcp28ss_round_mask((__v4sf)(__m128)(A), \
219 (__v4sf)(__m128)(B), \
224 ((__m128)__builtin_ia32_rcp28ss_round_mask((__v4sf)(__m128)(A), \
225 (__v4sf)(__m128)(B), \
226 (__v4sf)(__m128)(S), \
230 ((__m128)__builtin_ia32_rcp28ss_round_mask((__v4sf)(__m128)(A), \
[all …]
H A Davx512fintrin.h1032 _mm_mask_max_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { in _mm_mask_max_ss()
1267 _mm_mask_min_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { in _mm_mask_min_ss()
1639 _mm_mask_rsqrt14_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) in _mm_mask_rsqrt14_ss()
1737 _mm_rcp14_ss(__m128 __A, __m128 __B) in _mm_rcp14_ss()
1747 _mm_mask_rcp14_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) in _mm_mask_rcp14_ss()
1909 _mm_mask_add_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { in _mm_mask_add_ss()
2024 _mm_mask_sub_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { in _mm_mask_sub_ss()
2139 _mm_mask_mul_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { in _mm_mask_mul_ss()
2254 _mm_mask_div_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { in _mm_mask_div_ss()
6794 _mm_mask_sqrt_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) in _mm_mask_sqrt_ss()
[all …]
H A Davx512vlintrin.h1060 _mm_mask_fmadd_ps(__m128 __A, __mmask8 __U, __m128 __B, __m128 __C) in _mm_mask_fmadd_ps()
1070 _mm_mask3_fmadd_ps(__m128 __A, __m128 __B, __m128 __C, __mmask8 __U) in _mm_mask3_fmadd_ps()
1080 _mm_maskz_fmadd_ps(__mmask8 __U, __m128 __A, __m128 __B, __m128 __C) in _mm_maskz_fmadd_ps()
1090 _mm_mask_fmsub_ps(__m128 __A, __mmask8 __U, __m128 __B, __m128 __C) in _mm_mask_fmsub_ps()
1100 _mm_maskz_fmsub_ps(__mmask8 __U, __m128 __A, __m128 __B, __m128 __C) in _mm_maskz_fmsub_ps()
1110 _mm_mask3_fnmadd_ps(__m128 __A, __m128 __B, __m128 __C, __mmask8 __U) in _mm_mask3_fnmadd_ps()
1120 _mm_maskz_fnmadd_ps(__mmask8 __U, __m128 __A, __m128 __B, __m128 __C) in _mm_maskz_fnmadd_ps()
1130 _mm_maskz_fnmsub_ps(__mmask8 __U, __m128 __A, __m128 __B, __m128 __C) in _mm_maskz_fnmsub_ps()
1441 _mm_mask3_fmsub_ps(__m128 __A, __m128 __B, __m128 __C, __mmask8 __U) in _mm_mask3_fmsub_ps()
1521 _mm_mask_fnmadd_ps(__m128 __A, __mmask8 __U, __m128 __B, __m128 __C) in _mm_mask_fnmadd_ps()
[all …]
H A Dlsxintrin.h1637 __lsx_vfadd_s(__m128 _1, __m128 _2) { in __lsx_vfadd_s()
1649 __lsx_vfsub_s(__m128 _1, __m128 _2) { in __lsx_vfsub_s()
1661 __lsx_vfmul_s(__m128 _1, __m128 _2) { in __lsx_vfmul_s()
1673 __lsx_vfdiv_s(__m128 _1, __m128 _2) { in __lsx_vfdiv_s()
1697 __lsx_vfmin_s(__m128 _1, __m128 _2) { in __lsx_vfmin_s()
1709 __lsx_vfmina_s(__m128 _1, __m128 _2) { in __lsx_vfmina_s()
1721 __lsx_vfmax_s(__m128 _1, __m128 _2) { in __lsx_vfmax_s()
2300 __lsx_vfmadd_s(__m128 _1, __m128 _2, __m128 _3) { in __lsx_vfmadd_s()
2312 __lsx_vfmsub_s(__m128 _1, __m128 _2, __m128 _3) { in __lsx_vfmsub_s()
2324 __lsx_vfnmadd_s(__m128 _1, __m128 _2, __m128 _3) { in __lsx_vfnmadd_s()
[all …]
H A Dsmmintrin.h240 ((__m128)__builtin_ia32_roundps((__v4sf)(__m128)(X), (M)))
281 ((__m128)__builtin_ia32_roundss((__v4sf)(__m128)(X), (__v4sf)(__m128)(Y), \
413 ((__m128)__builtin_ia32_blendps((__v4sf)(__m128)(V1), (__v4sf)(__m128)(V2), \
463 static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_blendv_ps(__m128 __V1, in _mm_blendv_ps()
464 __m128 __V2, in _mm_blendv_ps()
465 __m128 __M) { in _mm_blendv_ps()
466 return (__m128)__builtin_ia32_blendvps((__v4sf)__V1, (__v4sf)__V2, in _mm_blendv_ps()
597 ((__m128)__builtin_ia32_dpps((__v4sf)(__m128)(X), (__v4sf)(__m128)(Y), (M)))
864 int, __builtin_ia32_vec_ext_v4sf((__v4sf)(__m128)(X), (int)(N)))
870 (D) = __builtin_ia32_vec_ext_v4sf((__v4sf)(__m128)(X), (int)(N)); \
H A Davxintrin.h1102 ((__m128)__builtin_ia32_vpermilps((__v4sf)(__m128)(A), (int)(C)))
1727 ((__m128)__builtin_ia32_cmpps((__v4sf)(__m128)(a), \
1965 ((__m128)__builtin_ia32_cmpss((__v4sf)(__m128)(a), \
2240 _mm256_cvtps_pd(__m128 __a) in _mm256_cvtps_pd()
2629 _mm_testz_ps(__m128 __a, __m128 __b) in _mm_testz_ps()
2658 _mm_testc_ps(__m128 __a, __m128 __b) in _mm_testc_ps()
2688 _mm_testnzc_ps(__m128 __a, __m128 __b) in _mm_testnzc_ps()
4535 _mm256_castps128_ps256(__m128 __a) in _mm256_castps128_ps256()
4829 _mm256_set_m128 (__m128 __hi, __m128 __lo) in _mm256_set_m128()
4893 _mm256_setr_m128 (__m128 __lo, __m128 __hi) in _mm256_setr_m128()
[all …]
H A Davx512dqintrin.h861 ((__m128)__builtin_ia32_rangess128_round_mask((__v4sf)(__m128)(A), \
870 ((__m128)__builtin_ia32_rangess128_round_mask((__v4sf)(__m128)(A), \
879 ((__m128)__builtin_ia32_rangess128_round_mask((__v4sf)(__m128)(A), \
981 ((__m128)__builtin_ia32_reducess_mask((__v4sf)(__m128)(A), \
987 ((__m128)__builtin_ia32_reducess_mask((__v4sf)(__m128)(A), \
993 ((__m128)__builtin_ia32_reducess_mask((__v4sf)(__m128)(A), \
1000 ((__m128)__builtin_ia32_reducess_mask((__v4sf)(__m128)(A), \
1006 ((__m128)__builtin_ia32_reducess_mask((__v4sf)(__m128)(A), \
1012 ((__m128)__builtin_ia32_reducess_mask((__v4sf)(__m128)(A), \
1081 _mm512_broadcast_f32x2 (__m128 __A) in _mm512_broadcast_f32x2()
[all …]
H A Df16cintrin.h96 ((__m128i)__builtin_ia32_vcvtps2ph((__v4sf)(__m128)(a), (imm)))
109 static __inline __m128 __DEFAULT_FN_ATTRS128
112 return (__m128)__builtin_ia32_vcvtph2ps((__v8hi)__a); in _mm_cvtph_ps()
H A Davx2intrin.h3003 static __inline__ __m128 __DEFAULT_FN_ATTRS128
3004 _mm_broadcastss_ps(__m128 __X) in _mm_broadcastss_ps()
3038 _mm256_broadcastss_ps(__m128 __X) in _mm256_broadcastss_ps()
4192 ((__m128)__builtin_ia32_gatherd_ps((__v4sf)(__m128)(a), \
4195 (__v4sf)(__m128)(mask), (s)))
4291 ((__m128)__builtin_ia32_gatherq_ps((__v4sf)(__m128)(a), \
4294 (__v4sf)(__m128)(mask), (s)))
4339 ((__m128)__builtin_ia32_gatherq_ps256((__v4sf)(__m128)(a), \
4342 (__v4sf)(__m128)(mask), (s)))
4903 ((__m128)__builtin_ia32_gatherd_ps((__v4sf)_mm_undefined_ps(), \
[all …]
H A Demmintrin.h1258 static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_cvtpd_ps(__m128d __a) { in _mm_cvtpd_ps()
1276 static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_cvtps_pd(__m128 __a) { in _mm_cvtps_pd()
1353 static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_cvtsd_ss(__m128 __a, in _mm_cvtsd_ss()
1355 return (__m128)__builtin_ia32_cvtsd2ss((__v4sf)__a, (__v2df)__b); in _mm_cvtsd_ss()
1401 __m128 __b) { in _mm_cvtss_sd()
3250 static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_cvtepi32_ps(__m128i __a) { in _mm_cvtepi32_ps()
3251 return (__m128) __builtin_convertvector((__v4si)__a, __v4sf); in _mm_cvtepi32_ps()
4666 static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_castpd_ps(__m128d __a) { in _mm_castpd_ps()
4667 return (__m128)__a; in _mm_castpd_ps()
4696 static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_castps_pd(__m128 __a) { in _mm_castps_pd()
[all …]
H A Dxopintrin.h723 ((__m128)__builtin_ia32_vpermil2ps((__v4sf)(__m128)(X), (__v4sf)(__m128)(Y), \
731 static __inline__ __m128 __DEFAULT_FN_ATTRS
732 _mm_frcz_ss(__m128 __A) in _mm_frcz_ss()
734 return (__m128)__builtin_ia32_vfrczss((__v4sf)__A); in _mm_frcz_ss()
743 static __inline__ __m128 __DEFAULT_FN_ATTRS
744 _mm_frcz_ps(__m128 __A) in _mm_frcz_ps()
746 return (__m128)__builtin_ia32_vfrczps((__v4sf)__A); in _mm_frcz_ps()
H A Dammintrin.h176 _mm_stream_ss(void *__p, __m128 __a) in _mm_stream_ss()
H A Davx512fp16intrin.h103 static __inline__ __m128 __DEFAULT_FN_ATTRS128 _mm_castph_ps(__m128h __a) { in _mm_castph_ps()
104 return (__m128)__a; in _mm_castph_ps()
141 static __inline__ __m128h __DEFAULT_FN_ATTRS128 _mm_castps_ph(__m128 __a) { in _mm_castps_ph()
1536 static __inline__ __m128 __DEFAULT_FN_ATTRS128 _mm_cvtsh_ss(__m128 __A, in _mm_cvtsh_ss()
1538 return (__m128)__builtin_ia32_vcvtsh2ss_round_mask( in _mm_cvtsh_ss()
1543 static __inline__ __m128 __DEFAULT_FN_ATTRS128 _mm_mask_cvtsh_ss(__m128 __W, in _mm_mask_cvtsh_ss()
1545 __m128 __A, in _mm_mask_cvtsh_ss()
1547 return (__m128)__builtin_ia32_vcvtsh2ss_round_mask((__v4sf)__A, (__v8hf)__B, in _mm_mask_cvtsh_ss()
1553 __m128 __A, in _mm_maskz_cvtsh_ss()
1555 return (__m128)__builtin_ia32_vcvtsh2ss_round_mask( in _mm_maskz_cvtsh_ss()
[all …]
H A Davx512vlfp16intrin.h343 return (__m128h)_mm_xor_ps((__m128)__A, _mm_set1_ps(-0.0f)); in _mm_conj_pch()
1349 static __inline__ __m128 __DEFAULT_FN_ATTRS128 _mm_cvtxph_ps(__m128h __A) { in _mm_cvtxph_ps()
1350 return (__m128)__builtin_ia32_vcvtph2psx128_mask( in _mm_cvtxph_ps()
1354 static __inline__ __m128 __DEFAULT_FN_ATTRS128 _mm_mask_cvtxph_ps(__m128 __W, in _mm_mask_cvtxph_ps()
1357 return (__m128)__builtin_ia32_vcvtph2psx128_mask((__v8hf)__A, (__v4sf)__W, in _mm_mask_cvtxph_ps()
1361 static __inline__ __m128 __DEFAULT_FN_ATTRS128
1363 return (__m128)__builtin_ia32_vcvtph2psx128_mask( in _mm_maskz_cvtxph_ps()
1384 static __inline__ __m128h __DEFAULT_FN_ATTRS128 _mm_cvtxps_ph(__m128 __A) { in _mm_cvtxps_ph()
1391 __m128 __A) { in _mm_mask_cvtxps_ph()
1397 _mm_maskz_cvtxps_ph(__mmask8 __U, __m128 __A) { in _mm_maskz_cvtxps_ph()
/freebsd/contrib/llvm-project/clang/lib/Headers/ppc_wrappers/
H A Dxmmintrin.h198 _mm_move_ss(__m128 __A, __m128 __B) { in _mm_move_ss()
224 _mm_add_ss(__m128 __A, __m128 __B) { in _mm_add_ss()
246 _mm_sub_ss(__m128 __A, __m128 __B) { in _mm_sub_ss()
268 _mm_mul_ss(__m128 __A, __m128 __B) { in _mm_mul_ss()
290 _mm_div_ss(__m128 __A, __m128 __B) { in _mm_div_ss()
329 _mm_add_ps(__m128 __A, __m128 __B) { in _mm_add_ps()
335 _mm_sub_ps(__m128 __A, __m128 __B) { in _mm_sub_ps()
341 _mm_mul_ps(__m128 __A, __m128 __B) { in _mm_mul_ps()
347 _mm_div_ps(__m128 __A, __m128 __B) { in _mm_div_ps()
403 _mm_min_ss(__m128 __A, __m128 __B) { in _mm_min_ss()
[all …]
H A Dpmmintrin.h48 extern __inline __m128
50 _mm_addsub_ps(__m128 __X, __m128 __Y) { in _mm_addsub_ps()
53 return (__m128)vec_add(__X, __even_neg_Y); in _mm_addsub_ps()
64 extern __inline __m128
66 _mm_hadd_ps(__m128 __X, __m128 __Y) { in _mm_hadd_ps()
77 extern __inline __m128
79 _mm_hsub_ps(__m128 __X, __m128 __Y) { in _mm_hsub_ps()
105 extern __inline __m128
107 _mm_movehdup_ps(__m128 __X) { in _mm_movehdup_ps()
113 extern __inline __m128
[all …]
H A Dsmmintrin.h156 extern __inline __m128
158 _mm_round_ps(__m128 __A, int __rounding) { in _mm_round_ps()
246 return (__m128)__r; in _mm_round_ps()
249 extern __inline __m128
251 _mm_round_ss(__m128 __A, __m128 __B, int __rounding) { in _mm_round_ss()
255 return (__m128)__r; in _mm_round_ss()
350 extern __inline __m128
352 _mm_blend_ps(__m128 __A, __m128 __B, const int __imm8) { in _mm_blend_ps()
372 return (__m128)__r; in _mm_blend_ps()
375 extern __inline __m128
[all …]
H A Demmintrin.h835 extern __inline __m128
878 extern __inline __m128
902 return ((__m128)__result); in _mm_cvtpd_ps()
986 _mm_cvtps_pd(__m128 __A) { in _mm_cvtps_pd()
1064 extern __inline __m128
1083 return (__m128)__result; in _mm_cvtsd_ss()
2228 extern __inline __m128
2231 return (__m128)__A; in _mm_castpd_ps()
2242 _mm_castps_pd(__m128 __A) { in _mm_castps_pd()
2252 extern __inline __m128
[all …]

12