/freebsd/contrib/llvm-project/clang/lib/Headers/ |
H A D | xmmintrin.h | 58 _mm_add_ss(__m128 __a, __m128 __b) in _mm_add_ss() 78 _mm_add_ps(__m128 __a, __m128 __b) in _mm_add_ps() 100 _mm_sub_ss(__m128 __a, __m128 __b) in _mm_sub_ss() 121 _mm_sub_ps(__m128 __a, __m128 __b) in _mm_sub_ps() 143 _mm_mul_ss(__m128 __a, __m128 __b) in _mm_mul_ss() 163 _mm_mul_ps(__m128 __a, __m128 __b) in _mm_mul_ps() 185 _mm_div_ss(__m128 __a, __m128 __b) in _mm_div_ss() 204 _mm_div_ps(__m128 __a, __m128 __b) in _mm_div_ps() 333 _mm_min_ss(__m128 __a, __m128 __b) in _mm_min_ss() 352 _mm_min_ps(__m128 __a, __m128 __b) in _mm_min_ps() [all …]
|
H A D | fma4intrin.h | 23 static __inline__ __m128 __DEFAULT_FN_ATTRS128 24 _mm_macc_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_macc_ps() 36 _mm_macc_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_macc_ss() 48 _mm_msub_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_msub_ps() 60 _mm_msub_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_msub_ss() 72 _mm_nmacc_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_nmacc_ps() 84 _mm_nmacc_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_nmacc_ss() 96 _mm_nmsub_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_nmsub_ps() 108 _mm_nmsub_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_nmsub_ss() 120 _mm_maddsub_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_maddsub_ps() [all …]
|
H A D | fmaintrin.h | 35 static __inline__ __m128 __DEFAULT_FN_ATTRS128 36 _mm_fmadd_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_fmadd_ps() 84 _mm_fmadd_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_fmadd_ss() 132 _mm_fmsub_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_fmsub_ps() 180 _mm_fmsub_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_fmsub_ss() 228 _mm_fnmadd_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_fnmadd_ps() 276 _mm_fnmadd_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_fnmadd_ss() 324 _mm_fnmsub_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_fnmsub_ps() 372 _mm_fnmsub_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_fnmsub_ss() 426 _mm_fmaddsub_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_fmaddsub_ps() [all …]
|
H A D | avx512vldqintrin.h | 108 _mm_mask_andnot_ps(__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) { in _mm_mask_andnot_ps() 115 _mm_maskz_andnot_ps(__mmask8 __U, __m128 __A, __m128 __B) { in _mm_maskz_andnot_ps() 164 _mm_mask_and_ps(__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) { in _mm_mask_and_ps() 171 _mm_maskz_and_ps(__mmask8 __U, __m128 __A, __m128 __B) { in _mm_maskz_and_ps() 220 _mm_mask_xor_ps(__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) { in _mm_mask_xor_ps() 227 _mm_maskz_xor_ps(__mmask8 __U, __m128 __A, __m128 __B) { in _mm_maskz_xor_ps() 276 _mm_mask_or_ps(__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) { in _mm_mask_or_ps() 283 _mm_maskz_or_ps(__mmask8 __U, __m128 __A, __m128 __B) { in _mm_maskz_or_ps() 818 ((__m128)__builtin_ia32_rangeps128_mask((__v4sf)(__m128)(A), \ 824 ((__m128)__builtin_ia32_rangeps128_mask((__v4sf)(__m128)(A), \ [all …]
|
H A D | avxneconvertintrin.h | 56 static __inline__ __m128 __DEFAULT_FN_ATTRS128 58 return (__m128)__builtin_ia32_vbcstnebf162ps128((const __bf16 *)__A); in _mm_bcstnebf16_ps() 122 static __inline__ __m128 __DEFAULT_FN_ATTRS128 124 return (__m128)__builtin_ia32_vbcstnesh2ps128((const _Float16 *)__A); in _mm_bcstnesh_ps() 188 static __inline__ __m128 __DEFAULT_FN_ATTRS128 254 static __inline__ __m128 __DEFAULT_FN_ATTRS128 256 return (__m128)__builtin_ia32_vcvtneeph2ps128((const __v8hf *)__A); in _mm_cvtneeph_ps() 320 static __inline__ __m128 __DEFAULT_FN_ATTRS128 386 static __inline__ __m128 __DEFAULT_FN_ATTRS128 388 return (__m128)__builtin_ia32_vcvtneoph2ps128((const __v8hf *)__A); in _mm_cvtneoph_ps() [all …]
|
H A D | pmmintrin.h | 57 static __inline__ __m128 __DEFAULT_FN_ATTRS 58 _mm_addsub_ps(__m128 __a, __m128 __b) in _mm_addsub_ps() 80 static __inline__ __m128 __DEFAULT_FN_ATTRS 81 _mm_hadd_ps(__m128 __a, __m128 __b) in _mm_hadd_ps() 103 static __inline__ __m128 __DEFAULT_FN_ATTRS 104 _mm_hsub_ps(__m128 __a, __m128 __b) in _mm_hsub_ps() 125 static __inline__ __m128 __DEFAULT_FN_ATTRS 126 _mm_movehdup_ps(__m128 __a) in _mm_movehdup_ps() 146 static __inline__ __m128 __DEFAULT_FN_ATTRS 147 _mm_moveldup_ps(__m128 __a) in _mm_moveldup_ps()
|
H A D | avx512vlbf16intrin.h | 40 _mm_cvtne2ps_pbh(__m128 __A, __m128 __B) { in _mm_cvtne2ps_pbh() 63 _mm_mask_cvtne2ps_pbh(__m128bh __W, __mmask8 __U, __m128 __A, __m128 __B) { in _mm_mask_cvtne2ps_pbh() 85 _mm_maskz_cvtne2ps_pbh(__mmask8 __U, __m128 __A, __m128 __B) { in _mm_maskz_cvtne2ps_pbh() 204 _mm_maskz_cvtneps_pbh(__mmask8 __U, __m128 __A) { in _mm_maskz_cvtneps_pbh() 276 static __inline__ __m128 __DEFAULT_FN_ATTRS128 277 _mm_dpbf16_ps(__m128 __D, __m128bh __A, __m128bh __B) { in _mm_dpbf16_ps() 278 return (__m128)__builtin_ia32_dpbf16ps_128((__v4sf)__D, in _mm_dpbf16_ps() 300 static __inline__ __m128 __DEFAULT_FN_ATTRS128 324 static __inline__ __m128 __DEFAULT_FN_ATTRS128 451 static __inline__ __m128 __DEFAULT_FN_ATTRS128 [all …]
|
H A D | avx512erintrin.h | 115 ((__m128)__builtin_ia32_rsqrt28ss_round_mask((__v4sf)(__m128)(A), \ 116 (__v4sf)(__m128)(B), \ 121 ((__m128)__builtin_ia32_rsqrt28ss_round_mask((__v4sf)(__m128)(A), \ 127 ((__m128)__builtin_ia32_rsqrt28ss_round_mask((__v4sf)(__m128)(A), \ 218 ((__m128)__builtin_ia32_rcp28ss_round_mask((__v4sf)(__m128)(A), \ 219 (__v4sf)(__m128)(B), \ 224 ((__m128)__builtin_ia32_rcp28ss_round_mask((__v4sf)(__m128)(A), \ 225 (__v4sf)(__m128)(B), \ 226 (__v4sf)(__m128)(S), \ 230 ((__m128)__builtin_ia32_rcp28ss_round_mask((__v4sf)(__m128)(A), \ [all …]
|
H A D | avx512fintrin.h | 1032 _mm_mask_max_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { in _mm_mask_max_ss() 1267 _mm_mask_min_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { in _mm_mask_min_ss() 1639 _mm_mask_rsqrt14_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) in _mm_mask_rsqrt14_ss() 1737 _mm_rcp14_ss(__m128 __A, __m128 __B) in _mm_rcp14_ss() 1747 _mm_mask_rcp14_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) in _mm_mask_rcp14_ss() 1909 _mm_mask_add_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { in _mm_mask_add_ss() 2024 _mm_mask_sub_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { in _mm_mask_sub_ss() 2139 _mm_mask_mul_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { in _mm_mask_mul_ss() 2254 _mm_mask_div_ss(__m128 __W, __mmask8 __U,__m128 __A, __m128 __B) { in _mm_mask_div_ss() 6794 _mm_mask_sqrt_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) in _mm_mask_sqrt_ss() [all …]
|
H A D | avx512vlintrin.h | 1060 _mm_mask_fmadd_ps(__m128 __A, __mmask8 __U, __m128 __B, __m128 __C) in _mm_mask_fmadd_ps() 1070 _mm_mask3_fmadd_ps(__m128 __A, __m128 __B, __m128 __C, __mmask8 __U) in _mm_mask3_fmadd_ps() 1080 _mm_maskz_fmadd_ps(__mmask8 __U, __m128 __A, __m128 __B, __m128 __C) in _mm_maskz_fmadd_ps() 1090 _mm_mask_fmsub_ps(__m128 __A, __mmask8 __U, __m128 __B, __m128 __C) in _mm_mask_fmsub_ps() 1100 _mm_maskz_fmsub_ps(__mmask8 __U, __m128 __A, __m128 __B, __m128 __C) in _mm_maskz_fmsub_ps() 1110 _mm_mask3_fnmadd_ps(__m128 __A, __m128 __B, __m128 __C, __mmask8 __U) in _mm_mask3_fnmadd_ps() 1120 _mm_maskz_fnmadd_ps(__mmask8 __U, __m128 __A, __m128 __B, __m128 __C) in _mm_maskz_fnmadd_ps() 1130 _mm_maskz_fnmsub_ps(__mmask8 __U, __m128 __A, __m128 __B, __m128 __C) in _mm_maskz_fnmsub_ps() 1441 _mm_mask3_fmsub_ps(__m128 __A, __m128 __B, __m128 __C, __mmask8 __U) in _mm_mask3_fmsub_ps() 1521 _mm_mask_fnmadd_ps(__m128 __A, __mmask8 __U, __m128 __B, __m128 __C) in _mm_mask_fnmadd_ps() [all …]
|
H A D | lsxintrin.h | 1637 __lsx_vfadd_s(__m128 _1, __m128 _2) { in __lsx_vfadd_s() 1649 __lsx_vfsub_s(__m128 _1, __m128 _2) { in __lsx_vfsub_s() 1661 __lsx_vfmul_s(__m128 _1, __m128 _2) { in __lsx_vfmul_s() 1673 __lsx_vfdiv_s(__m128 _1, __m128 _2) { in __lsx_vfdiv_s() 1697 __lsx_vfmin_s(__m128 _1, __m128 _2) { in __lsx_vfmin_s() 1709 __lsx_vfmina_s(__m128 _1, __m128 _2) { in __lsx_vfmina_s() 1721 __lsx_vfmax_s(__m128 _1, __m128 _2) { in __lsx_vfmax_s() 2300 __lsx_vfmadd_s(__m128 _1, __m128 _2, __m128 _3) { in __lsx_vfmadd_s() 2312 __lsx_vfmsub_s(__m128 _1, __m128 _2, __m128 _3) { in __lsx_vfmsub_s() 2324 __lsx_vfnmadd_s(__m128 _1, __m128 _2, __m128 _3) { in __lsx_vfnmadd_s() [all …]
|
H A D | smmintrin.h | 240 ((__m128)__builtin_ia32_roundps((__v4sf)(__m128)(X), (M))) 281 ((__m128)__builtin_ia32_roundss((__v4sf)(__m128)(X), (__v4sf)(__m128)(Y), \ 413 ((__m128)__builtin_ia32_blendps((__v4sf)(__m128)(V1), (__v4sf)(__m128)(V2), \ 463 static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_blendv_ps(__m128 __V1, in _mm_blendv_ps() 464 __m128 __V2, in _mm_blendv_ps() 465 __m128 __M) { in _mm_blendv_ps() 466 return (__m128)__builtin_ia32_blendvps((__v4sf)__V1, (__v4sf)__V2, in _mm_blendv_ps() 597 ((__m128)__builtin_ia32_dpps((__v4sf)(__m128)(X), (__v4sf)(__m128)(Y), (M))) 864 int, __builtin_ia32_vec_ext_v4sf((__v4sf)(__m128)(X), (int)(N))) 870 (D) = __builtin_ia32_vec_ext_v4sf((__v4sf)(__m128)(X), (int)(N)); \
|
H A D | avxintrin.h | 1102 ((__m128)__builtin_ia32_vpermilps((__v4sf)(__m128)(A), (int)(C))) 1727 ((__m128)__builtin_ia32_cmpps((__v4sf)(__m128)(a), \ 1965 ((__m128)__builtin_ia32_cmpss((__v4sf)(__m128)(a), \ 2240 _mm256_cvtps_pd(__m128 __a) in _mm256_cvtps_pd() 2629 _mm_testz_ps(__m128 __a, __m128 __b) in _mm_testz_ps() 2658 _mm_testc_ps(__m128 __a, __m128 __b) in _mm_testc_ps() 2688 _mm_testnzc_ps(__m128 __a, __m128 __b) in _mm_testnzc_ps() 4535 _mm256_castps128_ps256(__m128 __a) in _mm256_castps128_ps256() 4829 _mm256_set_m128 (__m128 __hi, __m128 __lo) in _mm256_set_m128() 4893 _mm256_setr_m128 (__m128 __lo, __m128 __hi) in _mm256_setr_m128() [all …]
|
H A D | avx512dqintrin.h | 861 ((__m128)__builtin_ia32_rangess128_round_mask((__v4sf)(__m128)(A), \ 870 ((__m128)__builtin_ia32_rangess128_round_mask((__v4sf)(__m128)(A), \ 879 ((__m128)__builtin_ia32_rangess128_round_mask((__v4sf)(__m128)(A), \ 981 ((__m128)__builtin_ia32_reducess_mask((__v4sf)(__m128)(A), \ 987 ((__m128)__builtin_ia32_reducess_mask((__v4sf)(__m128)(A), \ 993 ((__m128)__builtin_ia32_reducess_mask((__v4sf)(__m128)(A), \ 1000 ((__m128)__builtin_ia32_reducess_mask((__v4sf)(__m128)(A), \ 1006 ((__m128)__builtin_ia32_reducess_mask((__v4sf)(__m128)(A), \ 1012 ((__m128)__builtin_ia32_reducess_mask((__v4sf)(__m128)(A), \ 1081 _mm512_broadcast_f32x2 (__m128 __A) in _mm512_broadcast_f32x2() [all …]
|
H A D | f16cintrin.h | 96 ((__m128i)__builtin_ia32_vcvtps2ph((__v4sf)(__m128)(a), (imm))) 109 static __inline __m128 __DEFAULT_FN_ATTRS128 112 return (__m128)__builtin_ia32_vcvtph2ps((__v8hi)__a); in _mm_cvtph_ps()
|
H A D | avx2intrin.h | 3003 static __inline__ __m128 __DEFAULT_FN_ATTRS128 3004 _mm_broadcastss_ps(__m128 __X) in _mm_broadcastss_ps() 3038 _mm256_broadcastss_ps(__m128 __X) in _mm256_broadcastss_ps() 4192 ((__m128)__builtin_ia32_gatherd_ps((__v4sf)(__m128)(a), \ 4195 (__v4sf)(__m128)(mask), (s))) 4291 ((__m128)__builtin_ia32_gatherq_ps((__v4sf)(__m128)(a), \ 4294 (__v4sf)(__m128)(mask), (s))) 4339 ((__m128)__builtin_ia32_gatherq_ps256((__v4sf)(__m128)(a), \ 4342 (__v4sf)(__m128)(mask), (s))) 4903 ((__m128)__builtin_ia32_gatherd_ps((__v4sf)_mm_undefined_ps(), \ [all …]
|
H A D | emmintrin.h | 1258 static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_cvtpd_ps(__m128d __a) { in _mm_cvtpd_ps() 1276 static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_cvtps_pd(__m128 __a) { in _mm_cvtps_pd() 1353 static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_cvtsd_ss(__m128 __a, in _mm_cvtsd_ss() 1355 return (__m128)__builtin_ia32_cvtsd2ss((__v4sf)__a, (__v2df)__b); in _mm_cvtsd_ss() 1401 __m128 __b) { in _mm_cvtss_sd() 3250 static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_cvtepi32_ps(__m128i __a) { in _mm_cvtepi32_ps() 3251 return (__m128) __builtin_convertvector((__v4si)__a, __v4sf); in _mm_cvtepi32_ps() 4666 static __inline__ __m128 __DEFAULT_FN_ATTRS _mm_castpd_ps(__m128d __a) { in _mm_castpd_ps() 4667 return (__m128)__a; in _mm_castpd_ps() 4696 static __inline__ __m128d __DEFAULT_FN_ATTRS _mm_castps_pd(__m128 __a) { in _mm_castps_pd() [all …]
|
H A D | xopintrin.h | 723 ((__m128)__builtin_ia32_vpermil2ps((__v4sf)(__m128)(X), (__v4sf)(__m128)(Y), \ 731 static __inline__ __m128 __DEFAULT_FN_ATTRS 732 _mm_frcz_ss(__m128 __A) in _mm_frcz_ss() 734 return (__m128)__builtin_ia32_vfrczss((__v4sf)__A); in _mm_frcz_ss() 743 static __inline__ __m128 __DEFAULT_FN_ATTRS 744 _mm_frcz_ps(__m128 __A) in _mm_frcz_ps() 746 return (__m128)__builtin_ia32_vfrczps((__v4sf)__A); in _mm_frcz_ps()
|
H A D | ammintrin.h | 176 _mm_stream_ss(void *__p, __m128 __a) in _mm_stream_ss()
|
H A D | avx512fp16intrin.h | 103 static __inline__ __m128 __DEFAULT_FN_ATTRS128 _mm_castph_ps(__m128h __a) { in _mm_castph_ps() 104 return (__m128)__a; in _mm_castph_ps() 141 static __inline__ __m128h __DEFAULT_FN_ATTRS128 _mm_castps_ph(__m128 __a) { in _mm_castps_ph() 1536 static __inline__ __m128 __DEFAULT_FN_ATTRS128 _mm_cvtsh_ss(__m128 __A, in _mm_cvtsh_ss() 1538 return (__m128)__builtin_ia32_vcvtsh2ss_round_mask( in _mm_cvtsh_ss() 1543 static __inline__ __m128 __DEFAULT_FN_ATTRS128 _mm_mask_cvtsh_ss(__m128 __W, in _mm_mask_cvtsh_ss() 1545 __m128 __A, in _mm_mask_cvtsh_ss() 1547 return (__m128)__builtin_ia32_vcvtsh2ss_round_mask((__v4sf)__A, (__v8hf)__B, in _mm_mask_cvtsh_ss() 1553 __m128 __A, in _mm_maskz_cvtsh_ss() 1555 return (__m128)__builtin_ia32_vcvtsh2ss_round_mask( in _mm_maskz_cvtsh_ss() [all …]
|
H A D | avx512vlfp16intrin.h | 343 return (__m128h)_mm_xor_ps((__m128)__A, _mm_set1_ps(-0.0f)); in _mm_conj_pch() 1349 static __inline__ __m128 __DEFAULT_FN_ATTRS128 _mm_cvtxph_ps(__m128h __A) { in _mm_cvtxph_ps() 1350 return (__m128)__builtin_ia32_vcvtph2psx128_mask( in _mm_cvtxph_ps() 1354 static __inline__ __m128 __DEFAULT_FN_ATTRS128 _mm_mask_cvtxph_ps(__m128 __W, in _mm_mask_cvtxph_ps() 1357 return (__m128)__builtin_ia32_vcvtph2psx128_mask((__v8hf)__A, (__v4sf)__W, in _mm_mask_cvtxph_ps() 1361 static __inline__ __m128 __DEFAULT_FN_ATTRS128 1363 return (__m128)__builtin_ia32_vcvtph2psx128_mask( in _mm_maskz_cvtxph_ps() 1384 static __inline__ __m128h __DEFAULT_FN_ATTRS128 _mm_cvtxps_ph(__m128 __A) { in _mm_cvtxps_ph() 1391 __m128 __A) { in _mm_mask_cvtxps_ph() 1397 _mm_maskz_cvtxps_ph(__mmask8 __U, __m128 __A) { in _mm_maskz_cvtxps_ph()
|
/freebsd/contrib/llvm-project/clang/lib/Headers/ppc_wrappers/ |
H A D | xmmintrin.h | 198 _mm_move_ss(__m128 __A, __m128 __B) { in _mm_move_ss() 224 _mm_add_ss(__m128 __A, __m128 __B) { in _mm_add_ss() 246 _mm_sub_ss(__m128 __A, __m128 __B) { in _mm_sub_ss() 268 _mm_mul_ss(__m128 __A, __m128 __B) { in _mm_mul_ss() 290 _mm_div_ss(__m128 __A, __m128 __B) { in _mm_div_ss() 329 _mm_add_ps(__m128 __A, __m128 __B) { in _mm_add_ps() 335 _mm_sub_ps(__m128 __A, __m128 __B) { in _mm_sub_ps() 341 _mm_mul_ps(__m128 __A, __m128 __B) { in _mm_mul_ps() 347 _mm_div_ps(__m128 __A, __m128 __B) { in _mm_div_ps() 403 _mm_min_ss(__m128 __A, __m128 __B) { in _mm_min_ss() [all …]
|
H A D | pmmintrin.h | 48 extern __inline __m128 50 _mm_addsub_ps(__m128 __X, __m128 __Y) { in _mm_addsub_ps() 53 return (__m128)vec_add(__X, __even_neg_Y); in _mm_addsub_ps() 64 extern __inline __m128 66 _mm_hadd_ps(__m128 __X, __m128 __Y) { in _mm_hadd_ps() 77 extern __inline __m128 79 _mm_hsub_ps(__m128 __X, __m128 __Y) { in _mm_hsub_ps() 105 extern __inline __m128 107 _mm_movehdup_ps(__m128 __X) { in _mm_movehdup_ps() 113 extern __inline __m128 [all …]
|
H A D | smmintrin.h | 156 extern __inline __m128 158 _mm_round_ps(__m128 __A, int __rounding) { in _mm_round_ps() 246 return (__m128)__r; in _mm_round_ps() 249 extern __inline __m128 251 _mm_round_ss(__m128 __A, __m128 __B, int __rounding) { in _mm_round_ss() 255 return (__m128)__r; in _mm_round_ss() 350 extern __inline __m128 352 _mm_blend_ps(__m128 __A, __m128 __B, const int __imm8) { in _mm_blend_ps() 372 return (__m128)__r; in _mm_blend_ps() 375 extern __inline __m128 [all …]
|
H A D | emmintrin.h | 835 extern __inline __m128 878 extern __inline __m128 902 return ((__m128)__result); in _mm_cvtpd_ps() 986 _mm_cvtps_pd(__m128 __A) { in _mm_cvtps_pd() 1064 extern __inline __m128 1083 return (__m128)__result; in _mm_cvtsd_ss() 2228 extern __inline __m128 2231 return (__m128)__A; in _mm_castpd_ps() 2242 _mm_castps_pd(__m128 __A) { in _mm_castps_pd() 2252 extern __inline __m128 [all …]
|