/freebsd/contrib/arm-optimized-routines/math/test/ |
H A D | mathtest.c | 722 memset(&ret, 0, sizeof(ret)); in parsetest() 728 ret.comment = 1; in parsetest() 730 return ret; in parsetest() 732 ret.comment = 0; in parsetest() 745 ret.in_err = 0; in parsetest() 758 ret.comment = 1; in parsetest() 915 ret.in_err_limit = ret.in_err + 1; in parsetest() 928 ret.valid = 1; in parsetest() 929 return ret; in parsetest() 933 ret.valid = 0; in parsetest() [all …]
|
H A D | ulp.h | 40 static double RT (ulperr) (RT (float) got, const struct RT (ret) * p, int r, in RT() 135 int r, struct RT(ret) * p, in T() 168 int r, struct RT(ret) * p, in T() 199 int r_fenv, struct RT(ret) * p, in T() 274 struct RT(ret) want; in T()
|
/freebsd/contrib/arm-optimized-routines/math/test/rtest/ |
H A D | dotest.c | 368 mpc_set(ret, retbig, rnd); in test_cpow() 816 void float32_case(uint32 *ret) { in float32_case() argument 822 if (!ret) { in float32_case() 906 *ret = t; in float32_case() 940 *ret = f; in float32_case() 943 static void float64_case(uint32 *ret) { in float64_case() argument 949 if (!ret) { in float64_case() 1069 ret[0] = t; in float64_case() 1073 ret[1] = t; in float64_case() 1133 ret[0] = g; in float64_case() [all …]
|
H A D | random.c | 70 uint32 ret; in random_upto_internal() local 74 return ret; in random_upto_internal() 88 tmp = random_upto_internal(limit, mask); if (tmp < ret) ret = tmp; in random_upto_biased() 89 tmp = random_upto_internal(limit, mask); if (tmp < ret) ret = tmp; in random_upto_biased() 90 tmp = random_upto_internal(limit, mask); if (tmp < ret) ret = tmp; in random_upto_biased() 91 tmp = random_upto_internal(limit, mask); if (tmp < ret) ret = tmp; in random_upto_biased() 92 tmp = random_upto_internal(limit, mask); if (tmp < ret) ret = tmp; in random_upto_biased() 93 tmp = random_upto_internal(limit, mask); if (tmp < ret) ret = tmp; in random_upto_biased() 94 tmp = random_upto_internal(limit, mask); if (tmp < ret) ret = tmp; in random_upto_biased() 95 tmp = random_upto_internal(limit, mask); if (tmp < ret) ret = tmp; in random_upto_biased() [all …]
|
/freebsd/contrib/arm-optimized-routines/math/ |
H A D | tgamma128.c | 111 long double ret = 1 / ((expret * powhalf) * (x * negadjust) * p); in tgamma_large() local 112 return ret / powhalf; in tgamma_large()
|
/freebsd/contrib/arm-optimized-routines/math/tools/ |
H A D | remez.jl | 587 ret = best[n,1][2] 589 ret = best[n,2][2] 592 @assert(length(ret) == n) 593 return ret 940 ret = m \ v 941 println(" ",repr(ret)) 942 test(approx_eq(ret[1], 109/26)) 943 test(approx_eq(ret[2], -105/130)) 944 test(approx_eq(ret[3], -31/26))
|
/freebsd/contrib/arm-optimized-routines/pl/math/ |
H A D | atan2_2u5.c | 131 double ret; in atan2() local 136 ret = z; in atan2() 145 ret = eval_poly (z, z, shift); in atan2() 149 return asdouble (asuint64 (ret) ^ sign_xy); in atan2()
|
H A D | atan2f_3u.c | 139 float ret; in atan2f() local 144 ret = z; in atan2f() 153 ret = eval_poly (z, z, shift); in atan2f() 157 return asfloat (asuint (ret) ^ sign_xy); in atan2f()
|
H A D | sv_atan2_2u5.c | 35 special_case (svfloat64_t y, svfloat64_t x, svfloat64_t ret, in special_case() argument 38 return sv_call2_f64 (atan2, y, x, ret, cmp); in special_case() 91 svfloat64_t ret = svmla_x ( in SV_NAME_D2() local 97 ret = svmla_x (pg, z, z3, ret); in SV_NAME_D2() 99 ret = svadd_m (pg, ret, shift); in SV_NAME_D2() 102 ret = svreinterpret_f64 (sveor_x (pg, svreinterpret_u64 (ret), sign_xy)); in SV_NAME_D2() 105 return special_case (y, x, ret, cmp_xy); in SV_NAME_D2() 107 return ret; in SV_NAME_D2()
|
H A D | sv_atan2f_3u.c | 29 special_case (svfloat32_t y, svfloat32_t x, svfloat32_t ret, in special_case() argument 32 return sv_call2_f32 (atan2f, y, x, ret, cmp); in special_case() 85 svfloat32_t ret = sv_estrin_7_f32_x (pg, z2, z4, z8, data_ptr->poly); in SV_NAME_F2() local 89 ret = svmla_x (pg, z, z3, ret); in SV_NAME_F2() 91 ret = svadd_m (pg, ret, shift); in SV_NAME_F2() 94 ret = svreinterpret_f32 (sveor_x (pg, svreinterpret_u32 (ret), sign_xy)); in SV_NAME_F2() 97 return special_case (y, x, ret, cmp_xy); in SV_NAME_F2() 99 return ret; in SV_NAME_F2()
|
H A D | sv_math.h | 66 double ret = (*f) (elem1, elem2); in sv_call2_f64() local 67 svfloat64_t y2 = sv_f64 (ret); in sv_call2_f64() 124 float ret = (*f) (elem1, elem2); in sv_call2_f32() local 125 svfloat32_t y2 = sv_f32 (ret); in sv_call2_f32()
|
H A D | sv_powf_2u6.c | 307 svfloat32_t ret = sv_powf_core (pg, i, iz, k, y, sign_bias, &ylogx, d); in SV_NAME_F2() local 314 ret = svsel (svcmple (pg, ylogx, d->uflow_bound), ret_uflow, ret); in SV_NAME_F2() 315 ret = svsel (svcmpgt (pg, ylogx, d->oflow_bound), ret_oflow, ret); in SV_NAME_F2() 318 ret = svsel (yisnotint_xisneg, sv_f32 (__builtin_nanf ("")), ret); in SV_NAME_F2() 321 return sv_call_powf_sc (x, y, ret, cmp); in SV_NAME_F2() 323 return ret; in SV_NAME_F2()
|
H A D | v_atan2_3u.c | 37 special_case (float64x2_t y, float64x2_t x, float64x2_t ret, uint64x2_t cmp) in special_case() argument 39 return v_call2_f64 (atan2, y, x, ret, cmp); in special_case() 95 float64x2_t ret in V_NAME_D2() local 100 ret = vfmaq_f64 (z, ret, vmulq_f64 (z2, z)); in V_NAME_D2() 101 ret = vaddq_f64 (ret, shift); in V_NAME_D2() 104 ret = vreinterpretq_f64_u64 ( in V_NAME_D2() 105 veorq_u64 (vreinterpretq_u64_f64 (ret), sign_xy)); in V_NAME_D2() 108 return special_case (y, x, ret, special_cases); in V_NAME_D2() 110 return ret; in V_NAME_D2()
|
H A D | v_atan2f_3u.c | 31 special_case (float32x4_t y, float32x4_t x, float32x4_t ret, uint32x4_t cmp) in special_case() argument 33 return v_call2_f32 (atan2f, y, x, ret, cmp); in special_case() 89 float32x4_t ret = vfmaq_f32 ( in V_NAME_F2() local 94 ret = vaddq_f32 (vfmaq_f32 (z, ret, vmulq_f32 (z2, z)), shift); in V_NAME_F2() 97 ret = vreinterpretq_f32_u32 ( in V_NAME_F2() 98 veorq_u32 (vreinterpretq_u32_f32 (ret), sign_xy)); in V_NAME_F2() 102 return special_case (y, x, ret, special_cases); in V_NAME_F2() 105 return ret; in V_NAME_F2()
|
/freebsd/contrib/arm-optimized-routines/pl/math/test/ |
H A D | ulp_wrappers.h | 22 static int wrap_mpfr_powi(mpfr_t ret, const mpfr_t x, const mpfr_t y, mpfr_rnd_t rnd) { in wrap_mpfr_powi() argument 26 return mpfr_pow(ret, x, y2, rnd); in wrap_mpfr_powi()
|
/freebsd/contrib/arm-optimized-routines/string/aarch64/ |
H A D | __mtag_tag_region.S | 48 ret 57 ret 85 ret 97 ret
|
H A D | __mtag_tag_zero_region.S | 48 ret 57 ret 85 ret 97 ret
|
H A D | memchr-mte.S | 61 ret 103 ret 107 ret
|
H A D | memchr-sve.S | 45 ret 59 ret
|
H A D | memchr.S | 139 ret 143 ret
|
H A D | memcmp-sve.S | 42 ret 46 ret
|
H A D | memcmp.S | 92 ret 127 ret 188 ret
|
H A D | memcpy-advsimd.S | 73 ret 84 ret 94 ret 103 ret 116 ret 129 ret 170 ret 209 ret
|
H A D | memcpy-mops.S | 19 ret
|
H A D | memcpy-sve.S | 70 ret 82 ret 95 ret 136 ret 173 ret
|