Home
last modified time | relevance | path

Searched refs:__b (Results 1 – 25 of 87) sorted by relevance

1234

/minix/external/bsd/llvm/dist/clang/lib/Headers/
H A Daltivec.h6276 (__b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1)); in vec_splat()
6285 (__b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1)); in vec_splat()
6294 (__b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1)); in vec_splat()
6303 (__b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1)); in vec_splat()
6310 unsigned char b1=__b+1, b2=__b+2, b3=__b+3; in vec_splat()
6319 unsigned char b1=__b+1, b2=__b+2, b3=__b+3; in vec_splat()
6328 unsigned char b1=__b+1, b2=__b+2, b3=__b+3; in vec_splat()
6374 (__b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1)); in vec_vsplth()
6383 (__b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1)); in vec_vsplth()
6392 (__b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1, __b, b1)); in vec_vsplth()
[all …]
H A Dxmmintrin.h46 __a[0] += __b[0]; in _mm_add_ss()
53 return __a + __b; in _mm_add_ps()
59 __a[0] -= __b[0]; in _mm_sub_ss()
473 __a[0] = __b; in _mm_cvtsi32_ss()
488 __a[0] = __b; in _mm_cvtsi64_ss()
836 __b = _mm_cmpgt_pi16(__b, __a); in _mm_cvtpi16_ps()
867 __m64 __b; in _mm_cvtpi8_ps() local
870 __b = _mm_cmpgt_pi8(__b, __a); in _mm_cvtpi8_ps()
871 __b = _mm_unpacklo_pi8(__a, __b); in _mm_cvtpi8_ps()
879 __m64 __b; in _mm_cvtpu8_ps() local
[all …]
H A Demmintrin.h45 __a[0] += __b[0]; in _mm_add_sd()
52 return __a + __b; in _mm_add_pd()
58 __a[0] -= __b[0]; in _mm_sub_sd()
65 return __a - __b; in _mm_sub_pd()
71 __a[0] *= __b[0]; in _mm_mul_sd()
78 return __a * __b; in _mm_mul_pd()
84 __a[0] /= __b[0]; in _mm_div_sd()
415 __a[0] = __b; in _mm_cvtsi32_sd()
998 __a[0] = __b; in _mm_cvtsi64_sd()
1143 …return (__m128i)(__v16qi){ __b, __b, __b, __b, __b, __b, __b, __b, __b, __b, __b, __b, __b, __b, _… in _mm_set1_epi8()
[all …]
H A Dtmmintrin.h71 __m128i __b = (b); \
76 __m64 __b = (b); \
92 _mm_hadd_pi16(__m64 __a, __m64 __b) in _mm_hadd_pi16() argument
98 _mm_hadd_pi32(__m64 __a, __m64 __b) in _mm_hadd_pi32() argument
110 _mm_hadds_pi16(__m64 __a, __m64 __b) in _mm_hadds_pi16() argument
128 _mm_hsub_pi16(__m64 __a, __m64 __b) in _mm_hsub_pi16() argument
134 _mm_hsub_pi32(__m64 __a, __m64 __b) in _mm_hsub_pi32() argument
146 _mm_hsubs_pi16(__m64 __a, __m64 __b) in _mm_hsubs_pi16() argument
206 _mm_sign_pi8(__m64 __a, __m64 __b) in _mm_sign_pi8() argument
212 _mm_sign_pi16(__m64 __a, __m64 __b) in _mm_sign_pi16() argument
[all …]
H A Davxintrin.h46 return __a+__b; in _mm256_add_pd()
52 return __a+__b; in _mm256_add_ps()
58 return __a-__b; in _mm256_sub_pd()
64 return __a-__b; in _mm256_sub_ps()
82 return __a / __b; in _mm256_div_pd()
88 return __a / __b; in _mm256_div_ps()
118 return __a * __b; in _mm256_mul_pd()
124 return __a * __b; in _mm256_mul_ps()
1067 return (__m256i)(__v32qi){ __b, __b, __b, __b, __b, __b, __b, __b, __b, __b, in _mm256_set1_epi8()
1068 __b, __b, __b, __b, __b, __b, __b, __b, __b, __b, __b, __b, __b, __b, __b, in _mm256_set1_epi8()
[all …]
H A Davx512vlbwintrin.h34 _mm_cmpeq_epi8_mask(__m128i __a, __m128i __b) { in _mm_cmpeq_epi8_mask() argument
35 return (__mmask16)__builtin_ia32_pcmpeqb128_mask((__v16qi)__a, (__v16qi)__b, in _mm_cmpeq_epi8_mask()
40 _mm_mask_cmpeq_epi8_mask(__mmask16 __u, __m128i __a, __m128i __b) { in _mm_mask_cmpeq_epi8_mask() argument
47 _mm256_cmpeq_epi8_mask(__m256i __a, __m256i __b) { in _mm256_cmpeq_epi8_mask() argument
53 _mm256_mask_cmpeq_epi8_mask(__mmask32 __u, __m256i __a, __m256i __b) { in _mm256_mask_cmpeq_epi8_mask() argument
59 _mm_cmpeq_epi16_mask(__m128i __a, __m128i __b) { in _mm_cmpeq_epi16_mask() argument
60 return (__mmask8)__builtin_ia32_pcmpeqw128_mask((__v8hi)__a, (__v8hi)__b, in _mm_cmpeq_epi16_mask()
65 _mm_mask_cmpeq_epi16_mask(__mmask8 __u, __m128i __a, __m128i __b) { in _mm_mask_cmpeq_epi16_mask() argument
66 return (__mmask8)__builtin_ia32_pcmpeqw128_mask((__v8hi)__a, (__v8hi)__b, in _mm_mask_cmpeq_epi16_mask()
72 _mm256_cmpeq_epi16_mask(__m256i __a, __m256i __b) { in _mm256_cmpeq_epi16_mask() argument
[all …]
H A Davx512vlintrin.h34 _mm_cmpeq_epi32_mask(__m128i __a, __m128i __b) { in _mm_cmpeq_epi32_mask() argument
35 return (__mmask8)__builtin_ia32_pcmpeqd128_mask((__v4si)__a, (__v4si)__b, in _mm_cmpeq_epi32_mask()
40 _mm_mask_cmpeq_epi32_mask(__mmask8 __u, __m128i __a, __m128i __b) { in _mm_mask_cmpeq_epi32_mask() argument
41 return (__mmask8)__builtin_ia32_pcmpeqd128_mask((__v4si)__a, (__v4si)__b, in _mm_mask_cmpeq_epi32_mask()
47 _mm256_cmpeq_epi32_mask(__m256i __a, __m256i __b) { in _mm256_cmpeq_epi32_mask() argument
48 return (__mmask8)__builtin_ia32_pcmpeqd256_mask((__v8si)__a, (__v8si)__b, in _mm256_cmpeq_epi32_mask()
53 _mm256_mask_cmpeq_epi32_mask(__mmask8 __u, __m256i __a, __m256i __b) { in _mm256_mask_cmpeq_epi32_mask() argument
59 _mm_cmpeq_epi64_mask(__m128i __a, __m128i __b) { in _mm_cmpeq_epi64_mask() argument
65 _mm_mask_cmpeq_epi64_mask(__mmask8 __u, __m128i __a, __m128i __b) { in _mm_mask_cmpeq_epi64_mask() argument
72 _mm256_cmpeq_epi64_mask(__m256i __a, __m256i __b) { in _mm256_cmpeq_epi64_mask() argument
[all …]
H A Davx2intrin.h77 _mm256_add_epi8(__m256i __a, __m256i __b) in _mm256_add_epi8() argument
97 return __a + __b; in _mm256_add_epi64()
126 __m256i __b = (b); \
132 return __a & __b; in _mm256_and_si256()
138 return ~__a & __b; in _mm256_andnot_si256()
142 _mm256_avg_epu8(__m256i __a, __m256i __b) in _mm256_avg_epu8() argument
202 return (__m256i)(__a == __b); in _mm256_cmpeq_epi64()
226 return (__m256i)(__a > __b); in _mm256_cmpgt_epi64()
472 return __a | __b; in _mm256_or_si256()
666 return __a - __b; in _mm256_sub_epi64()
[all …]
H A Dpmmintrin.h40 _mm_addsub_ps(__m128 __a, __m128 __b) in _mm_addsub_ps() argument
42 return __builtin_ia32_addsubps(__a, __b); in _mm_addsub_ps()
46 _mm_hadd_ps(__m128 __a, __m128 __b) in _mm_hadd_ps() argument
48 return __builtin_ia32_haddps(__a, __b); in _mm_hadd_ps()
52 _mm_hsub_ps(__m128 __a, __m128 __b) in _mm_hsub_ps() argument
54 return __builtin_ia32_hsubps(__a, __b); in _mm_hsub_ps()
70 _mm_addsub_pd(__m128d __a, __m128d __b) in _mm_addsub_pd() argument
76 _mm_hadd_pd(__m128d __a, __m128d __b) in _mm_hadd_pd() argument
78 return __builtin_ia32_haddpd(__a, __b); in _mm_hadd_pd()
82 _mm_hsub_pd(__m128d __a, __m128d __b) in _mm_hsub_pd() argument
[all …]
H A Davx512bwintrin.h37 _mm512_cmpeq_epi8_mask(__m512i __a, __m512i __b) { in _mm512_cmpeq_epi8_mask() argument
38 return (__mmask64)__builtin_ia32_pcmpeqb512_mask((__v64qi)__a, (__v64qi)__b, in _mm512_cmpeq_epi8_mask()
43 _mm512_mask_cmpeq_epi8_mask(__mmask64 __u, __m512i __a, __m512i __b) { in _mm512_mask_cmpeq_epi8_mask() argument
44 return (__mmask64)__builtin_ia32_pcmpeqb512_mask((__v64qi)__a, (__v64qi)__b, in _mm512_mask_cmpeq_epi8_mask()
49 _mm512_cmpeq_epi16_mask(__m512i __a, __m512i __b) { in _mm512_cmpeq_epi16_mask() argument
50 return (__mmask32)__builtin_ia32_pcmpeqw512_mask((__v32hi)__a, (__v32hi)__b, in _mm512_cmpeq_epi16_mask()
55 _mm512_mask_cmpeq_epi16_mask(__mmask32 __u, __m512i __a, __m512i __b) { in _mm512_mask_cmpeq_epi16_mask() argument
56 return (__mmask32)__builtin_ia32_pcmpeqw512_mask((__v32hi)__a, (__v32hi)__b, in _mm512_mask_cmpeq_epi16_mask()
/minix/external/bsd/llvm/dist/clang/test/CodeGen/
H A Dsse-builtins.c262 return _mm_cmpeq_ss(__a, __b); in test_mm_cmpeq_ss()
268 return _mm_cmplt_ss(__a, __b); in test_mm_cmplt_ss()
274 return _mm_cmple_ss(__a, __b); in test_mm_cmple_ss()
310 return _mm_cmpgt_ss(__a, __b); in test_mm_cmpgt_ss()
316 return _mm_cmpge_ss(__a, __b); in test_mm_cmpge_ss()
334 return _mm_cmpeq_ps(__a, __b); in test_mm_cmpeq_ps()
340 return _mm_cmplt_ps(__a, __b); in test_mm_cmplt_ps()
346 return _mm_cmple_ps(__a, __b); in test_mm_cmple_ps()
382 return _mm_cmpgt_ps(__a, __b); in test_mm_cmpgt_ps()
388 return _mm_cmpge_ps(__a, __b); in test_mm_cmpge_ps()
[all …]
H A Davx512vl-builtins.c5 __mmask8 test_mm256_cmpeq_epi32_mask(__m256i __a, __m256i __b) { in test_mm256_cmpeq_epi32_mask() argument
8 return (__mmask8)_mm256_cmpeq_epi32_mask(__a, __b); in test_mm256_cmpeq_epi32_mask()
14 return (__mmask8)_mm256_mask_cmpeq_epi32_mask(__u, __a, __b); in test_mm256_mask_cmpeq_epi32_mask()
17 __mmask8 test_mm_cmpeq_epi32_mask(__m128i __a, __m128i __b) { in test_mm_cmpeq_epi32_mask() argument
20 return (__mmask8)_mm_cmpeq_epi32_mask(__a, __b); in test_mm_cmpeq_epi32_mask()
26 return (__mmask8)_mm_mask_cmpeq_epi32_mask(__u, __a, __b); in test_mm_mask_cmpeq_epi32_mask()
32 return (__mmask8)_mm256_cmpeq_epi64_mask(__a, __b); in test_mm256_cmpeq_epi64_mask()
38 return (__mmask8)_mm256_mask_cmpeq_epi64_mask(__u, __a, __b); in test_mm256_mask_cmpeq_epi64_mask()
41 __mmask8 test_mm_cmpeq_epi64_mask(__m128i __a, __m128i __b) { in test_mm_cmpeq_epi64_mask() argument
44 return (__mmask8)_mm_cmpeq_epi64_mask(__a, __b); in test_mm_cmpeq_epi64_mask()
[all …]
H A Davx512vlbw-builtins.c5 __mmask32 test_mm256_cmpeq_epi8_mask(__m256i __a, __m256i __b) { in test_mm256_cmpeq_epi8_mask() argument
8 return (__mmask32)_mm256_cmpeq_epi8_mask(__a, __b); in test_mm256_cmpeq_epi8_mask()
14 return (__mmask32)_mm256_mask_cmpeq_epi8_mask(__u, __a, __b); in test_mm256_mask_cmpeq_epi8_mask()
17 __mmask16 test_mm_cmpeq_epi8_mask(__m128i __a, __m128i __b) { in test_mm_cmpeq_epi8_mask() argument
20 return (__mmask16)_mm_cmpeq_epi8_mask(__a, __b); in test_mm_cmpeq_epi8_mask()
26 return (__mmask16)_mm_mask_cmpeq_epi8_mask(__u, __a, __b); in test_mm_mask_cmpeq_epi8_mask()
32 return (__mmask16)_mm256_cmpeq_epi16_mask(__a, __b); in test_mm256_cmpeq_epi16_mask()
38 return (__mmask16)_mm256_mask_cmpeq_epi16_mask(__u, __a, __b); in test_mm256_mask_cmpeq_epi16_mask()
41 __mmask8 test_mm_cmpeq_epi16_mask(__m128i __a, __m128i __b) { in test_mm_cmpeq_epi16_mask() argument
44 return (__mmask8)_mm_cmpeq_epi16_mask(__a, __b); in test_mm_cmpeq_epi16_mask()
[all …]
H A Davx512bw-builtins.c5 __mmask64 test_mm512_cmpeq_epi8_mask(__m512i __a, __m512i __b) { in test_mm512_cmpeq_epi8_mask() argument
8 return (__mmask64)_mm512_cmpeq_epi8_mask(__a, __b); in test_mm512_cmpeq_epi8_mask()
11 __mmask64 test_mm512_mask_cmpeq_epi8_mask(__mmask64 __u, __m512i __a, __m512i __b) { in test_mm512_mask_cmpeq_epi8_mask() argument
14 return (__mmask64)_mm512_mask_cmpeq_epi8_mask(__u, __a, __b); in test_mm512_mask_cmpeq_epi8_mask()
17 __mmask32 test_mm512_cmpeq_epi16_mask(__m512i __a, __m512i __b) { in test_mm512_cmpeq_epi16_mask() argument
20 return (__mmask32)_mm512_cmpeq_epi16_mask(__a, __b); in test_mm512_cmpeq_epi16_mask()
23 __mmask32 test_mm512_mask_cmpeq_epi16_mask(__mmask32 __u, __m512i __a, __m512i __b) { in test_mm512_mask_cmpeq_epi16_mask() argument
26 return (__mmask32)_mm512_mask_cmpeq_epi16_mask(__u, __a, __b); in test_mm512_mask_cmpeq_epi16_mask()
H A Davx512f-builtins.c162 __mmask16 test_mm512_cmpeq_epi32_mask(__m512i __a, __m512i __b) { in test_mm512_cmpeq_epi32_mask() argument
165 return (__mmask16)_mm512_cmpeq_epi32_mask(__a, __b); in test_mm512_cmpeq_epi32_mask()
168 __mmask16 test_mm512_mask_cmpeq_epi32_mask(__mmask16 __u, __m512i __a, __m512i __b) { in test_mm512_mask_cmpeq_epi32_mask() argument
171 return (__mmask16)_mm512_mask_cmpeq_epi32_mask(__u, __a, __b); in test_mm512_mask_cmpeq_epi32_mask()
174 __mmask8 test_mm512_mask_cmpeq_epi64_mask(__mmask8 __u, __m512i __a, __m512i __b) { in test_mm512_mask_cmpeq_epi64_mask() argument
177 return (__mmask8)_mm512_mask_cmpeq_epi64_mask(__u, __a, __b); in test_mm512_mask_cmpeq_epi64_mask()
180 __mmask8 test_mm512_cmpeq_epi64_mask(__m512i __a, __m512i __b) { in test_mm512_cmpeq_epi64_mask() argument
183 return (__mmask8)_mm512_cmpeq_epi64_mask(__a, __b); in test_mm512_cmpeq_epi64_mask()
/minix/sys/external/bsd/compiler_rt/dist/lib/builtins/
H A Dmuldc3.c21 __muldc3(double __a, double __b, double __c, double __d) in __muldc3() argument
24 double __bd = __b * __d; in __muldc3()
26 double __bc = __b * __c; in __muldc3()
33 if (crt_isinf(__a) || crt_isinf(__b)) in __muldc3()
36 __b = crt_copysign(crt_isinf(__b) ? 1 : 0, __b); in __muldc3()
49 if (crt_isnan(__b)) in __muldc3()
50 __b = crt_copysign(0, __b); in __muldc3()
58 if (crt_isnan(__b)) in __muldc3()
59 __b = crt_copysign(0, __b); in __muldc3()
68 __real__ z = CRT_INFINITY * (__a * __c - __b * __d); in __muldc3()
[all …]
H A Dmulxc3.c23 __mulxc3(long double __a, long double __b, long double __c, long double __d) in __mulxc3() argument
26 long double __bd = __b * __d; in __mulxc3()
28 long double __bc = __b * __c; in __mulxc3()
35 if (crt_isinf(__a) || crt_isinf(__b)) in __mulxc3()
38 __b = crt_copysignl(crt_isinf(__b) ? 1 : 0, __b); in __mulxc3()
51 if (crt_isnan(__b)) in __mulxc3()
52 __b = crt_copysignl(0, __b); in __mulxc3()
60 if (crt_isnan(__b)) in __mulxc3()
61 __b = crt_copysignl(0, __b); in __mulxc3()
70 __real__ z = CRT_INFINITY * (__a * __c - __b * __d); in __mulxc3()
[all …]
H A Dmulsc3.c21 __mulsc3(float __a, float __b, float __c, float __d) in __mulsc3() argument
24 float __bd = __b * __d; in __mulsc3()
26 float __bc = __b * __c; in __mulsc3()
33 if (crt_isinf(__a) || crt_isinf(__b)) in __mulsc3()
36 __b = crt_copysignf(crt_isinf(__b) ? 1 : 0, __b); in __mulsc3()
49 if (crt_isnan(__b)) in __mulsc3()
50 __b = crt_copysignf(0, __b); in __mulsc3()
58 if (crt_isnan(__b)) in __mulsc3()
59 __b = crt_copysignf(0, __b); in __mulsc3()
68 __real__ z = CRT_INFINITY * (__a * __c - __b * __d); in __mulsc3()
[all …]
H A Ddivdc3.c21 __divdc3(double __a, double __b, double __c, double __d) in __divdc3() argument
33 __real__ z = crt_scalbn((__a * __c + __b * __d) / __denom, -__ilogbw); in __divdc3()
37 if ((__denom == 0.0) && (!crt_isnan(__a) || !crt_isnan(__b))) in __divdc3()
40 __imag__ z = crt_copysign(CRT_INFINITY, __c) * __b; in __divdc3()
42 else if ((crt_isinf(__a) || crt_isinf(__b)) && in __divdc3()
46 __b = crt_copysign(crt_isinf(__b) ? 1.0 : 0.0, __b); in __divdc3()
47 __real__ z = CRT_INFINITY * (__a * __c + __b * __d); in __divdc3()
48 __imag__ z = CRT_INFINITY * (__b * __c - __a * __d); in __divdc3()
51 crt_isfinite(__a) && crt_isfinite(__b)) in __divdc3()
55 __real__ z = 0.0 * (__a * __c + __b * __d); in __divdc3()
[all …]
H A Ddivsc3.c21 __divsc3(float __a, float __b, float __c, float __d) in __divsc3() argument
33 __real__ z = crt_scalbnf((__a * __c + __b * __d) / __denom, -__ilogbw); in __divsc3()
37 if ((__denom == 0) && (!crt_isnan(__a) || !crt_isnan(__b))) in __divsc3()
40 __imag__ z = crt_copysignf(CRT_INFINITY, __c) * __b; in __divsc3()
42 else if ((crt_isinf(__a) || crt_isinf(__b)) && in __divsc3()
46 __b = crt_copysignf(crt_isinf(__b) ? 1 : 0, __b); in __divsc3()
47 __real__ z = CRT_INFINITY * (__a * __c + __b * __d); in __divsc3()
48 __imag__ z = CRT_INFINITY * (__b * __c - __a * __d); in __divsc3()
51 crt_isfinite(__a) && crt_isfinite(__b)) in __divsc3()
55 __real__ z = 0 * (__a * __c + __b * __d); in __divsc3()
[all …]
H A Ddivxc3.c22 __divxc3(long double __a, long double __b, long double __c, long double __d) in __divxc3() argument
34 __real__ z = crt_scalbnl((__a * __c + __b * __d) / __denom, -__ilogbw); in __divxc3()
38 if ((__denom == 0) && (!crt_isnan(__a) || !crt_isnan(__b))) in __divxc3()
41 __imag__ z = crt_copysignl(CRT_INFINITY, __c) * __b; in __divxc3()
43 else if ((crt_isinf(__a) || crt_isinf(__b)) && in __divxc3()
47 __b = crt_copysignl(crt_isinf(__b) ? 1 : 0, __b); in __divxc3()
48 __real__ z = CRT_INFINITY * (__a * __c + __b * __d); in __divxc3()
49 __imag__ z = CRT_INFINITY * (__b * __c - __a * __d); in __divxc3()
52 crt_isfinite(__a) && crt_isfinite(__b)) in __divxc3()
56 __real__ z = 0 * (__a * __c + __b * __d); in __divxc3()
[all …]
/minix/external/bsd/llvm/dist/clang/test/Sema/
H A Dbuiltins-x86.c6 __m128 test__builtin_ia32_cmpps(__m128 __a, __m128 __b) { in test__builtin_ia32_cmpps() argument
7 __builtin_ia32_cmpps(__a, __b, 32); // expected-error {{argument should be a value from 0 to 31}} in test__builtin_ia32_cmpps()
10 __m128d test__builtin_ia32_cmppd(__m128d __a, __m128d __b) { in test__builtin_ia32_cmppd() argument
11 __builtin_ia32_cmppd(__a, __b, 32); // expected-error {{argument should be a value from 0 to 31}} in test__builtin_ia32_cmppd()
14 __m128 test__builtin_ia32_cmpss(__m128 __a, __m128 __b) { in test__builtin_ia32_cmpss() argument
15 __builtin_ia32_cmpss(__a, __b, 32); // expected-error {{argument should be a value from 0 to 31}} in test__builtin_ia32_cmpss()
18 __m128d test__builtin_ia32_cmpsd(__m128d __a, __m128d __b) { in test__builtin_ia32_cmpsd() argument
19 __builtin_ia32_cmpsd(__a, __b, 32); // expected-error {{argument should be a value from 0 to 31}} in test__builtin_ia32_cmpsd()
/minix/external/bsd/llvm/dist/llvm/test/CodeGen/PowerPC/
H A Dvsx-ldst-builtin-le.ll33 %__b.addr.i21 = alloca i32, align 4
36 %__b.addr.i18 = alloca i32, align 4
39 %__b.addr.i15 = alloca i32, align 4
42 %__b.addr.i12 = alloca i32, align 4
45 %__b.addr.i9 = alloca i32, align 4
48 %__b.addr.i7 = alloca i32, align 4
51 %__b.addr.i = alloca <4 x i32>*, align 8
108 %35 = load i32* %__b.addr.i21, align 4
118 %41 = load i32* %__b.addr.i18, align 4
151 %62 = load i32* %__b.addr.i9, align 4
[all …]
/minix/external/bsd/llvm/dist/llvm/test/CodeGen/AArch64/
H A Darm64-vext.ll8 %__b = alloca <8 x i8>, align 8
14 %tmp3 = load <8 x i8>* %__b, align 8
25 %__b = alloca <8 x i8>, align 8
31 %tmp3 = load <8 x i8>* %__b, align 8
42 %__b = alloca <8 x i8>, align 8
59 %__b = alloca <4 x i16>, align 8
80 %__b = alloca <4 x i16>, align 8
101 %__b = alloca <4 x i16>, align 8
122 %__b = alloca <2 x i32>, align 8
143 %__b = alloca <2 x i32>, align 8
[all …]
/minix/external/bsd/libc++/dist/libcxx/include/
H A Dlocale971 __b = do_get(__b, __e, __iob, __err, __lv);
1020 for (; __b != __e; ++__b)
1070 for (; __b != __e; ++__b)
1124 for (; __b != __e; ++__b)
1174 for (; __b != __e; ++__b)
1859 for (++__b, (void) --__n; __b != __e && __n > 0; ++__b, (void) --__n)
2227 for (; __b != __e && __ct.is(ctype_base::space, *__b); ++__b)
2314 for ( ; __b != __e && __ct.is(ctype_base::space, *__b); ++__b)
3134 for (; __b != __e; ++__b)
3162 if (__b == __e || *__b != __dp)
[all …]

1234