/dports/graphics/librsvg2-rust/librsvg-2.52.8/vendor/wide/src/ |
H A D | u32x8_.rs | 53 Self { sse0: add_i32_m128i(self.sse0, rhs.sse0), sse1: add_i32_m128i(self.sse1, rhs.sse1)} in add() 81 Self { sse0: sub_i32_m128i(self.sse0, rhs.sse0), sse1: sub_i32_m128i(self.sse1, rhs.sse1)} in sub() 109 … sse0: mul_i32_keep_low_m128i(self.sse0, rhs.sse0), sse1: mul_i32_keep_low_m128i(self.sse1, rhs.ss… in mul() 139 Self { sse0: bitand_m128i(self.sse0, rhs.sse0), sse1: bitand_m128i(self.sse1, rhs.sse1)} in bitand() 167 Self { sse0: bitor_m128i(self.sse0, rhs.sse0), sse1: bitor_m128i(self.sse1, rhs.sse1)} in bitor() 195 Self { sse0: bitxor_m128i(self.sse0, rhs.sse0), sse1: bitxor_m128i(self.sse1, rhs.sse1)} in bitxor() 304 …f { sse0: cmp_eq_mask_i32_m128i(self.sse0,rhs.sse0), sse1: cmp_eq_mask_i32_m128i(self.sse1,rhs.sse… in cmp_eq() 376 …end_varying_i8_m128i(f.sse0, t.sse0, self.sse0), sse1: blend_varying_i8_m128i(f.sse1, t.sse1, self… in blend() 392 Self { sse0: max_i32_m128i(self.sse0, rhs.sse0), sse1: max_i32_m128i(self.sse1, rhs.sse1) } in max() 407 Self { sse0: max_i32_m128i(self.sse0, rhs.sse0), sse1: max_i32_m128i(self.sse1, rhs.sse1) } in min() [all …]
|
H A D | i32x8_.rs | 53 Self { sse0: add_i32_m128i(self.sse0, rhs.sse0), sse1: add_i32_m128i(self.sse1, rhs.sse1)} in add() 81 Self { sse0: sub_i32_m128i(self.sse0, rhs.sse0), sse1: sub_i32_m128i(self.sse1, rhs.sse1)} in sub() 109 … sse0: mul_i32_keep_low_m128i(self.sse0, rhs.sse0), sse1: mul_i32_keep_low_m128i(self.sse1, rhs.ss… in mul() 193 Self { sse0: bitand_m128i(self.sse0, rhs.sse0), sse1: bitand_m128i(self.sse1, rhs.sse1)} in bitand() 221 Self { sse0: bitor_m128i(self.sse0, rhs.sse0), sse1: bitor_m128i(self.sse1, rhs.sse1)} in bitor() 249 Self { sse0: bitxor_m128i(self.sse0, rhs.sse0), sse1: bitxor_m128i(self.sse1, rhs.sse1)} in bitxor() 354 …f { sse0: cmp_eq_mask_i32_m128i(self.sse0,rhs.sse0), sse1: cmp_eq_mask_i32_m128i(self.sse1,rhs.sse… in cmp_eq() 441 …end_varying_i8_m128i(f.sse0, t.sse0, self.sse0), sse1: blend_varying_i8_m128i(f.sse1, t.sse1, self… in blend() 481 Self { sse0: max_i32_m128i(self.sse0, rhs.sse0), sse1: max_i32_m128i(self.sse1, rhs.sse1) } in max() 496 Self { sse0: min_i32_m128i(self.sse0, rhs.sse0), sse1: min_i32_m128i(self.sse1, rhs.sse1) } in min() [all …]
|
H A D | u64x4_.rs | 53 Self { sse0: add_i64_m128i(self.sse0, rhs.sse0), sse1: add_i64_m128i(self.sse1, rhs.sse1) } in add() 77 Self { sse0: sub_i64_m128i(self.sse0, rhs.sse0), sse1: sub_i64_m128i(self.sse1, rhs.sse1) } in sub() 177 Self { sse0: bitand_m128i(self.sse0, rhs.sse0), sse1: bitand_m128i(self.sse1, rhs.sse1) } in bitand() 201 Self { sse0: bitor_m128i(self.sse0, rhs.sse0) , sse1: bitor_m128i(self.sse1, rhs.sse1)} in bitor() 225 Self { sse0: bitxor_m128i(self.sse0, rhs.sse0), sse1: bitxor_m128i(self.sse1, rhs.sse1) } in bitxor() 254 … Self { sse0: shl_all_u64_m128i(self.sse0, shift), sse1: shl_all_u64_m128i(self.sse1, shift) } 288 … Self { sse0: shr_all_u64_m128i(self.sse0, shift), sse1: shr_all_u64_m128i(self.sse1, shift) } 321 …f { sse0: cmp_eq_mask_i64_m128i(self.sse0, rhs.sse0),sse1: cmp_eq_mask_i64_m128i(self.sse1, rhs.ss… in cmp_eq() 343 … { sse0: cmp_gt_mask_i64_m128i(self.sse0, rhs.sse0), sse1: cmp_gt_mask_i64_m128i(self.sse1, rhs.ss… in cmp_gt() 365 …end_varying_i8_m128i(f.sse0, t.sse0, self.sse0), sse1: blend_varying_i8_m128i(f.sse1, t.sse1, self… in blend() [all …]
|
H A D | i64x4_.rs | 53 Self { sse0: add_i64_m128i(self.sse0, rhs.sse0), sse1: add_i64_m128i(self.sse1, rhs.sse1) } in add() 77 Self { sse0: sub_i64_m128i(self.sse0, rhs.sse0), sse1: sub_i64_m128i(self.sse1, rhs.sse1) } in sub() 177 Self { sse0: bitand_m128i(self.sse0, rhs.sse0), sse1: bitand_m128i(self.sse1, rhs.sse1) } in bitand() 201 Self { sse0: bitor_m128i(self.sse0, rhs.sse0) , sse1: bitor_m128i(self.sse1, rhs.sse1)} in bitor() 225 Self { sse0: bitxor_m128i(self.sse0, rhs.sse0), sse1: bitxor_m128i(self.sse1, rhs.sse1) } in bitxor() 254 … Self { sse0: shl_all_u64_m128i(self.sse0, shift), sse1: shl_all_u64_m128i(self.sse1, shift) } 313 …f { sse0: cmp_eq_mask_i64_m128i(self.sse0, rhs.sse0),sse1: cmp_eq_mask_i64_m128i(self.sse1, rhs.ss… in cmp_eq() 339 … { sse0: cmp_gt_mask_i64_m128i(self.sse0, rhs.sse0), sse1: cmp_gt_mask_i64_m128i(self.sse1, rhs.ss… in cmp_gt() 366 … sse1: !cmp_gt_mask_i64_m128i(self.sse1, rhs.sse1) ^ cmp_eq_mask_i64_m128i(self.sse1, rhs.sse1)} in cmp_lt() 396 …end_varying_i8_m128i(f.sse0, t.sse0, self.sse0), sse1: blend_varying_i8_m128i(f.sse1, t.sse1, self… in blend() [all …]
|
H A D | i16x16_.rs | 55 sse1: add_i16_m128i(self.sse1, rhs.sse1) in add() 96 sse1: sub_i16_m128i(self.sse1, rhs.sse1) } in sub() 136 sse1: mul_i16_keep_low_m128i(self.sse1, rhs.sse1) } in mul() 230 sse1: bitand_m128i(self.sse1, rhs.sse1) } in bitand() 270 sse1: bitor_m128i(self.sse1, rhs.sse1) } in bitor() 310 sse1: bitxor_m128i(self.sse1, rhs.sse1) } in bitxor() 452 sse1: cmp_eq_mask_i16_m128i(self.sse1, rhs.sse1) } in cmp_eq() 492 sse1: cmp_gt_mask_i16_m128i(self.sse1, rhs.sse1) } in cmp_gt() 577 sse1: blend_varying_i8_m128i(f.sse1, t.sse1, self.sse1) in blend() 634 sse1: max_i16_m128i(self.sse1, rhs.sse1) } in max() [all …]
|
H A D | f32x8_.rs | 82 Self { sse0: add_m128(self.sse0, rhs.sse0), sse1: add_m128(self.sse1, rhs.sse1) } in add() 110 Self { sse0: sub_m128(self.sse0, rhs.sse0), sse1: sub_m128(self.sse1, rhs.sse1) } in sub() 138 Self { sse0: mul_m128(self.sse0, rhs.sse0), sse1: mul_m128(self.sse1, rhs.sse1) } in mul() 166 Self { sse0: div_m128(self.sse0, rhs.sse0), sse1: div_m128(self.sse1, rhs.sse1) } in div() 266 Self { sse0: bitand_m128(self.sse0, rhs.sse0), sse1: bitand_m128(self.sse1, rhs.sse1) } in bitand() 294 Self { sse0: bitor_m128(self.sse0, rhs.sse0), sse1: bitor_m128(self.sse1, rhs.sse1) } in bitor() 522 …e0: blend_varying_m128(f.sse0, t.sse0, self.sse0), sse1: blend_varying_m128(f.sse1, t.sse1, self.s… in blend() 553 Self { sse0: max_m128(self.sse0, rhs.sse0), sse1: max_m128(self.sse1, rhs.sse1) } in fast_max() 631 Self { sse0: min_m128(self.sse0, rhs.sse0), sse1: min_m128(self.sse1, rhs.sse1) } in fast_min() 908 …e0: fused_mul_add_m128(self.sse0, m.sse0, a.sse0), sse1: fused_mul_add_m128(self.sse1, m.sse1, a.s… in mul_add() [all …]
|
H A D | f64x4_.rs | 82 Self { sse0: add_m128d(self.sse0, rhs.sse0), sse1: add_m128d(self.sse1, rhs.sse1) } in add() 106 Self { sse0: sub_m128d(self.sse0, rhs.sse0), sse1: sub_m128d(self.sse1, rhs.sse1) } in sub() 130 Self { sse0: mul_m128d(self.sse0, rhs.sse0), sse1: mul_m128d(self.sse1, rhs.sse1) } in mul() 154 Self { sse0: div_m128d(self.sse0, rhs.sse0), sse1: div_m128d(self.sse1, rhs.sse1) } in div() 274 Self { sse0: bitor_m128d(self.sse0, rhs.sse0), sse1: bitor_m128d(self.sse1, rhs.sse1) } in bitor() 470 …0: blend_varying_m128d(f.sse0, t.sse0, self.sse0), sse1: blend_varying_m128d(f.sse1, t.sse1, self.… in blend() 502 Self { sse0: max_m128d(self.sse0, rhs.sse0), sse1: max_m128d(self.sse1, rhs.sse1) } in fast_max() 573 Self { sse0: min_m128d(self.sse0, rhs.sse0), sse1: min_m128d(self.sse1, rhs.sse1) } in fast_min() 716 …0: fused_mul_add_m128d(self.sse0, m.sse0, a.sse0), sse1: fused_mul_add_m128d(self.sse1, m.sse1, a.… in mul_add() 731 …0: fused_mul_sub_m128d(self.sse0, m.sse0, a.sse0), sse1: fused_mul_sub_m128d(self.sse1, m.sse1, a.… in mul_sub() [all …]
|
H A D | i8x32_.rs | 53 Self { sse0: add_i8_m128i(self.sse0, rhs.sse0), sse1: add_i8_m128i(self.sse1, rhs.sse1) } in add() 105 Self { sse0: sub_i8_m128i(self.sse0, rhs.sse0), sse1: sub_i8_m128i(self.sse1, rhs.sse1) } in sub() 193 Self { sse0: bitand_m128i(self.sse0, rhs.sse0), sse1: bitand_m128i(self.sse1, rhs.sse1) } in bitand() 245 Self { sse0: bitor_m128i(self.sse0, rhs.sse0), sse1: bitor_m128i(self.sse1, rhs.sse1) } in bitor() 297 Self { sse0: bitxor_m128i(self.sse0, rhs.sse0), sse1: bitxor_m128i(self.sse1, rhs.sse1) } in bitxor() 349 … { sse0: cmp_eq_mask_i8_m128i(self.sse0, rhs.sse0), sse1: cmp_eq_mask_i8_m128i(self.sse1, rhs.sse… in cmp_eq() 401 … { sse0: cmp_gt_mask_i8_m128i(self.sse0, rhs.sse0), sse1: cmp_gt_mask_i8_m128i(self.sse1, rhs.sse… in cmp_gt() 453 … { sse0: cmp_lt_mask_i8_m128i(self.sse0, rhs.sse0), sse1: cmp_lt_mask_i8_m128i(self.sse1, rhs.sse… in cmp_lt() 509 …nd_varying_i8_m128i(f.sse0, t.sse0, self.sse0), sse1: blend_varying_i8_m128i(f.sse1, t.sse1, self… in blend() 573 Self { sse0: max_i8_m128i(self.sse0,rhs.sse0), sse1: max_i8_m128i(self.sse1,rhs.sse1) } in max() [all …]
|
/dports/multimedia/svt-av1/SVT-AV1-3971c982311d49f9355dc8dccdcf8d21b70fa624/Source/Lib/Encoder/Codec/ |
H A D | mcomp.c | 206 *sse1 = sse; in svt_check_better_fast() 234 *sse1 = sse; in svt_check_better() 322 sse1, in svt_first_level_check() 361 sse1, in svt_second_level_check_v2() 372 sse1, in svt_second_level_check_v2() 386 sse1, in svt_second_level_check_v2() 445 sse1, distortion, &dummy, is_scaled); in first_level_check_fast() 455 sse1, distortion, &dummy, is_scaled); in first_level_check_fast() 555 sse1, distortion, is_scaled); in two_level_checks_fast() 581 mv_cost_params, sse1, distortion); in svt_av1_find_best_sub_pixel_tree_pruned() [all …]
|
/dports/multimedia/gstreamer1-plugins-svt-av1/SVT-AV1-3971c982311d49f9355dc8dccdcf8d21b70fa624/Source/Lib/Encoder/Codec/ |
H A D | mcomp.c | 206 *sse1 = sse; in svt_check_better_fast() 234 *sse1 = sse; in svt_check_better() 322 sse1, in svt_first_level_check() 361 sse1, in svt_second_level_check_v2() 372 sse1, in svt_second_level_check_v2() 386 sse1, in svt_second_level_check_v2() 445 sse1, distortion, &dummy, is_scaled); in first_level_check_fast() 455 sse1, distortion, &dummy, is_scaled); in first_level_check_fast() 555 sse1, distortion, is_scaled); in two_level_checks_fast() 581 mv_cost_params, sse1, distortion); in svt_av1_find_best_sub_pixel_tree_pruned() [all …]
|
/dports/www/firefox-legacy/firefox-52.8.0esr/media/libvpx/vpx_dsp/x86/ |
H A D | variance_mmx.c | 43 unsigned int sse0, sse1, sse2, sse3, var; in vpx_mse16x16_mmx() local 47 vpx_get8x8var_mmx(a + 8, a_stride, b + 8, b_stride, &sse1, &sum1); in vpx_mse16x16_mmx() 53 var = sse0 + sse1 + sse2 + sse3; in vpx_mse16x16_mmx() 61 unsigned int sse0, sse1, sse2, sse3, var; in vpx_variance16x16_mmx() local 65 vpx_get8x8var_mmx(a + 8, a_stride, b + 8, b_stride, &sse1, &sum1); in vpx_variance16x16_mmx() 71 var = sse0 + sse1 + sse2 + sse3; in vpx_variance16x16_mmx() 80 unsigned int sse0, sse1, var; in vpx_variance16x8_mmx() local 86 var = sse0 + sse1; in vpx_variance16x8_mmx() 95 unsigned int sse0, sse1, var; in vpx_variance8x16_mmx() local 100 b + 8 * b_stride, b_stride, &sse1, &sum1); in vpx_variance8x16_mmx() [all …]
|
/dports/graphics/dssim/dssim-3.1.2/cargo-crates/libaom-sys-0.9.1/vendor/aom_dsp/x86/ |
H A D | highbd_variance_avx2.c | 61 __m256i sse1 = _mm256_setzero_si256(); in aom_highbd_var_filter_block2d_bil_avx2() local 72 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 106 __m256i sse1 = _mm256_setzero_si256(); in aom_highbd_var_filter_block2d_bil_avx2() local 121 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 185 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 256 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 308 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 376 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 430 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 507 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() [all …]
|
/dports/www/chromium-legacy/chromium-88.0.4324.182/third_party/libaom/source/libaom/aom_dsp/x86/ |
H A D | highbd_variance_avx2.c | 61 __m256i sse1 = _mm256_setzero_si256(); in aom_highbd_var_filter_block2d_bil_avx2() local 72 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 106 __m256i sse1 = _mm256_setzero_si256(); in aom_highbd_var_filter_block2d_bil_avx2() local 121 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 185 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 256 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 308 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 376 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 430 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 507 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() [all …]
|
/dports/multimedia/aom/aom-3.2.0/aom_dsp/x86/ |
H A D | highbd_variance_avx2.c | 61 __m256i sse1 = _mm256_setzero_si256(); in aom_highbd_var_filter_block2d_bil_avx2() local 72 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 106 __m256i sse1 = _mm256_setzero_si256(); in aom_highbd_var_filter_block2d_bil_avx2() local 121 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 185 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 256 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 308 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 376 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 430 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() 507 sse1 = _mm256_add_epi32(sse1, V_R_MAD); in aom_highbd_var_filter_block2d_bil_avx2() [all …]
|
/dports/www/firefox-legacy/firefox-52.8.0esr/media/libvpx/vp9/encoder/x86/ |
H A D | vp9_variance_avx2.c | 40 unsigned int sse1; in vp9_sub_pixel_variance64x64_avx2() local 43 64, &sse1); in vp9_sub_pixel_variance64x64_avx2() 50 *sse = sse1 + sse2; in vp9_sub_pixel_variance64x64_avx2() 75 unsigned int sse1; in vp9_sub_pixel_avg_variance64x64_avx2() local 78 sec, 64, 64, &sse1); in vp9_sub_pixel_avg_variance64x64_avx2() 86 *sse = sse1 + sse2; in vp9_sub_pixel_avg_variance64x64_avx2()
|
/dports/graphics/dssim/dssim-3.1.2/cargo-crates/libaom-sys-0.9.1/vendor/av1/encoder/ |
H A D | mcomp.c | 2412 *sse1 = sse; in check_better_fast() 2439 *sse1 = sse; in check_better() 2472 sse1, distortion, &dummy, is_scaled); in first_level_check_fast() 3105 (void)sse1; in av1_return_max_sub_pixel_mv() 3132 (void)sse1; in av1_return_min_sub_pixel_mv() 3391 *sse1 = sse; in obmc_check_better_fast() 3420 *sse1 = sse; in obmc_check_better() 3434 unsigned int *sse1, int *distortion) { in obmc_first_level_check() argument 3470 sse1, distortion, &dummy); in obmc_first_level_check() 3478 sse1, distortion, &dummy); in obmc_first_level_check() [all …]
|
/dports/www/chromium-legacy/chromium-88.0.4324.182/third_party/libaom/source/libaom/av1/encoder/ |
H A D | mcomp.c | 2402 *sse1 = sse; in check_better_fast() 2429 *sse1 = sse; in check_better() 2462 sse1, distortion, &dummy, is_scaled); in first_level_check_fast() 3095 (void)sse1; in av1_return_max_sub_pixel_mv() 3122 (void)sse1; in av1_return_min_sub_pixel_mv() 3381 *sse1 = sse; in obmc_check_better_fast() 3410 *sse1 = sse; in obmc_check_better() 3424 unsigned int *sse1, int *distortion) { in obmc_first_level_check() argument 3460 sse1, distortion, &dummy); in obmc_first_level_check() 3468 sse1, distortion, &dummy); in obmc_first_level_check() [all …]
|
/dports/multimedia/aom/aom-3.2.0/av1/encoder/ |
H A D | mcomp.c | 2428 *sse1 = sse; in check_better_fast() 2455 *sse1 = sse; in check_better() 2488 sse1, distortion, &dummy, is_scaled); in first_level_check_fast() 3121 (void)sse1; in av1_return_max_sub_pixel_mv() 3148 (void)sse1; in av1_return_min_sub_pixel_mv() 3407 *sse1 = sse; in obmc_check_better_fast() 3436 *sse1 = sse; in obmc_check_better() 3450 unsigned int *sse1, int *distortion) { in obmc_first_level_check() argument 3486 sse1, distortion, &dummy); in obmc_first_level_check() 3494 sse1, distortion, &dummy); in obmc_first_level_check() [all …]
|
/dports/www/qt5-webengine/qtwebengine-everywhere-src-5.15.2/src/3rdparty/chromium/third_party/libaom/source/libaom/av1/encoder/ |
H A D | mcomp.c | 2087 *sse1 = sse; in check_better_fast() 2117 *sse1 = sse; in check_better() 2134 unsigned int *sse1, int *distortion) { in first_level_check_fast() argument 2453 unsigned int *sse1) { in av1_find_best_sub_pixel_tree_pruned_evenmore() argument 2529 unsigned int *sse1) { in av1_find_best_sub_pixel_tree_pruned_more() argument 2605 unsigned int *sse1) { in av1_find_best_sub_pixel_tree_pruned() argument 2870 (void)sse1; in av1_return_max_sub_pixel_mv() 2898 (void)sse1; in av1_return_min_sub_pixel_mv() 3131 *sse1 = sse; in obmc_check_better_fast() 3162 *sse1 = sse; in obmc_check_better() [all …]
|
/dports/www/qt5-webengine/qtwebengine-everywhere-src-5.15.2/src/3rdparty/chromium/third_party/libaom/source/libaom/aom_dsp/arm/ |
H A D | variance_neon.c | 87 uint32_t sse1, sse2; in aom_variance32x64_neon() local 91 *sse = sse1 + sse2; in aom_variance32x64_neon() 100 uint32_t sse1, sse2; in aom_variance64x32_neon() local 104 *sse = sse1 + sse2; in aom_variance64x32_neon() 113 uint32_t sse1, sse2; in aom_variance64x64_neon() local 118 sse1 += sse2; in aom_variance64x64_neon() 123 sse1 += sse2; in aom_variance64x64_neon() 128 *sse = sse1 + sse2; in aom_variance64x64_neon() 138 sum1 = sse1 = 0; in aom_variance128x128_neon() 142 sse1 += sse2; in aom_variance128x128_neon() [all …]
|
/dports/lang/fpc-source/fpc-3.2.2/tests/webtbs/ |
H A D | tw32822.pp | 14 sse1,sse2,sse3,sse4:single; 28 a: TSSE2 = ( prefix: 0; sse: (sse1: 3.4E38; sse2: 3.4E38; sse3: 3.0; sse4: 4.0)); 29 b: TSSE2 = (prefix: 0; sse: (sse1: 3.4E38; sse2: 3.4E38; sse3: 0.0; sse4: 0.0)); 30 c: TSSE2 = (prefix: 0; sse: (sse1: 0.0; sse2: 0.0; sse3: 0.0; sse4: 0.0));
|
/dports/lang/spidermonkey60/firefox-60.9.0/media/libvpx/libvpx/vpx_dsp/x86/ |
H A D | variance_avx2.c | 109 unsigned int sse1; in vpx_sub_pixel_variance64x64_avx2() local 111 src, src_stride, x_offset, y_offset, dst, dst_stride, 64, &sse1); in vpx_sub_pixel_variance64x64_avx2() 117 *sse = sse1 + sse2; in vpx_sub_pixel_variance64x64_avx2() 134 unsigned int sse1; in vpx_sub_pixel_avg_variance64x64_avx2() local 136 src, src_stride, x_offset, y_offset, dst, dst_stride, sec, 64, 64, &sse1); in vpx_sub_pixel_avg_variance64x64_avx2() 143 *sse = sse1 + sse2; in vpx_sub_pixel_avg_variance64x64_avx2()
|
/dports/www/chromium-legacy/chromium-88.0.4324.182/third_party/libvpx/source/libvpx/vpx_dsp/arm/ |
H A D | variance_neon.c | 213 uint32_t sse1, sse2; in vpx_variance32x64_neon() local 214 variance_neon_w16(src_ptr, src_stride, ref_ptr, ref_stride, 32, 32, &sse1, in vpx_variance32x64_neon() 219 *sse = sse1 + sse2; in vpx_variance32x64_neon() 228 uint32_t sse1, sse2; in vpx_variance64x32_neon() local 229 variance_neon_w16(src_ptr, src_stride, ref_ptr, ref_stride, 64, 16, &sse1, in vpx_variance64x32_neon() 234 *sse = sse1 + sse2; in vpx_variance64x32_neon() 243 uint32_t sse1, sse2; in vpx_variance64x64_neon() local 245 variance_neon_w16(src_ptr, src_stride, ref_ptr, ref_stride, 64, 16, &sse1, in vpx_variance64x64_neon() 250 sse1 += sse2; in vpx_variance64x64_neon() 256 sse1 += sse2; in vpx_variance64x64_neon() [all …]
|
/dports/www/firefox-esr/firefox-91.8.0/media/libvpx/libvpx/vpx_dsp/arm/ |
H A D | variance_neon.c | 213 uint32_t sse1, sse2; in vpx_variance32x64_neon() local 214 variance_neon_w16(src_ptr, src_stride, ref_ptr, ref_stride, 32, 32, &sse1, in vpx_variance32x64_neon() 219 *sse = sse1 + sse2; in vpx_variance32x64_neon() 228 uint32_t sse1, sse2; in vpx_variance64x32_neon() local 229 variance_neon_w16(src_ptr, src_stride, ref_ptr, ref_stride, 64, 16, &sse1, in vpx_variance64x32_neon() 234 *sse = sse1 + sse2; in vpx_variance64x32_neon() 243 uint32_t sse1, sse2; in vpx_variance64x64_neon() local 245 variance_neon_w16(src_ptr, src_stride, ref_ptr, ref_stride, 64, 16, &sse1, in vpx_variance64x64_neon() 250 sse1 += sse2; in vpx_variance64x64_neon() 256 sse1 += sse2; in vpx_variance64x64_neon() [all …]
|
/dports/net/freeswitch/freeswitch-1.10.3.-release/libs/libvpx/vpx_dsp/arm/ |
H A D | variance_neon.c | 213 uint32_t sse1, sse2; in vpx_variance32x64_neon() local 214 variance_neon_w16(src_ptr, src_stride, ref_ptr, ref_stride, 32, 32, &sse1, in vpx_variance32x64_neon() 219 *sse = sse1 + sse2; in vpx_variance32x64_neon() 228 uint32_t sse1, sse2; in vpx_variance64x32_neon() local 229 variance_neon_w16(src_ptr, src_stride, ref_ptr, ref_stride, 64, 16, &sse1, in vpx_variance64x32_neon() 234 *sse = sse1 + sse2; in vpx_variance64x32_neon() 243 uint32_t sse1, sse2; in vpx_variance64x64_neon() local 245 variance_neon_w16(src_ptr, src_stride, ref_ptr, ref_stride, 64, 16, &sse1, in vpx_variance64x64_neon() 250 sse1 += sse2; in vpx_variance64x64_neon() 256 sse1 += sse2; in vpx_variance64x64_neon() [all …]
|