/dports/www/chromium-legacy/chromium-88.0.4324.182/third_party/libaom/source/libaom/av1/encoder/arm/neon/ |
H A D | picksrt_neon.c | 28 const int32_t shift = SGRPROJ_RST_BITS + SGRPROJ_PRJ_BITS; in av1_lowbd_pixel_proj_error_neon() 48 vreinterpretq_s16_u16(vshll_n_u8(d0, SGRPROJ_RST_BITS)); in av1_lowbd_pixel_proj_error_neon() 70 const int32_t u = dat[k] << SGRPROJ_RST_BITS; in av1_lowbd_pixel_proj_error_neon() 100 xq_active << SGRPROJ_RST_BITS); in av1_lowbd_pixel_proj_error_neon() 103 xq_active << SGRPROJ_RST_BITS); in av1_lowbd_pixel_proj_error_neon() 114 const int32_t u = dat[k] << SGRPROJ_RST_BITS; in av1_lowbd_pixel_proj_error_neon()
|
/dports/multimedia/aom/aom-3.2.0/av1/encoder/arm/neon/ |
H A D | picksrt_neon.c | 27 const int32_t shift = SGRPROJ_RST_BITS + SGRPROJ_PRJ_BITS; in av1_lowbd_pixel_proj_error_neon() 47 vreinterpretq_s16_u16(vshll_n_u8(d0, SGRPROJ_RST_BITS)); in av1_lowbd_pixel_proj_error_neon() 69 const int32_t u = dat[k] << SGRPROJ_RST_BITS; in av1_lowbd_pixel_proj_error_neon() 99 xq_active << SGRPROJ_RST_BITS); in av1_lowbd_pixel_proj_error_neon() 102 xq_active << SGRPROJ_RST_BITS); in av1_lowbd_pixel_proj_error_neon() 113 const int32_t u = dat[k] << SGRPROJ_RST_BITS; in av1_lowbd_pixel_proj_error_neon()
|
/dports/graphics/dssim/dssim-3.1.2/cargo-crates/libaom-sys-0.9.1/vendor/av1/encoder/arm/neon/ |
H A D | picksrt_neon.c | 28 const int32_t shift = SGRPROJ_RST_BITS + SGRPROJ_PRJ_BITS; in av1_lowbd_pixel_proj_error_neon() 48 vreinterpretq_s16_u16(vshll_n_u8(d0, SGRPROJ_RST_BITS)); in av1_lowbd_pixel_proj_error_neon() 70 const int32_t u = dat[k] << SGRPROJ_RST_BITS; in av1_lowbd_pixel_proj_error_neon() 100 xq_active << SGRPROJ_RST_BITS); in av1_lowbd_pixel_proj_error_neon() 103 xq_active << SGRPROJ_RST_BITS); in av1_lowbd_pixel_proj_error_neon() 114 const int32_t u = dat[k] << SGRPROJ_RST_BITS; in av1_lowbd_pixel_proj_error_neon()
|
/dports/multimedia/gstreamer1-plugins-svt-av1/SVT-AV1-3971c982311d49f9355dc8dccdcf8d21b70fa624/Source/Lib/Encoder/ASM_AVX2/ |
H A D | EbRestorationPick_AVX2.c | 67 u_256 = _mm256_slli_epi16(u_256, SGRPROJ_RST_BITS); in svt_get_proj_subspace_avx2() 71 s_256 = _mm256_slli_epi16(s_256, SGRPROJ_RST_BITS); in svt_get_proj_subspace_avx2() 121 const double u = (double)(dat[i * dat_stride + j] << SGRPROJ_RST_BITS); in svt_get_proj_subspace_avx2() 122 const double s = (double)(src[i * src_stride + j] << SGRPROJ_RST_BITS) - u; in svt_get_proj_subspace_avx2() 167 u_256 = _mm256_slli_epi16(u_256, SGRPROJ_RST_BITS); in svt_get_proj_subspace_avx2() 170 s_256 = _mm256_slli_epi16(s_256, SGRPROJ_RST_BITS); in svt_get_proj_subspace_avx2() 220 const double u = (double)(dat[i * dat_stride + j] << SGRPROJ_RST_BITS); in svt_get_proj_subspace_avx2() 221 const double s = (double)(src[i * src_stride + j] << SGRPROJ_RST_BITS) - u; in svt_get_proj_subspace_avx2()
|
/dports/multimedia/svt-av1/SVT-AV1-3971c982311d49f9355dc8dccdcf8d21b70fa624/Source/Lib/Encoder/ASM_AVX2/ |
H A D | EbRestorationPick_AVX2.c | 67 u_256 = _mm256_slli_epi16(u_256, SGRPROJ_RST_BITS); in svt_get_proj_subspace_avx2() 71 s_256 = _mm256_slli_epi16(s_256, SGRPROJ_RST_BITS); in svt_get_proj_subspace_avx2() 121 const double u = (double)(dat[i * dat_stride + j] << SGRPROJ_RST_BITS); in svt_get_proj_subspace_avx2() 122 const double s = (double)(src[i * src_stride + j] << SGRPROJ_RST_BITS) - u; in svt_get_proj_subspace_avx2() 167 u_256 = _mm256_slli_epi16(u_256, SGRPROJ_RST_BITS); in svt_get_proj_subspace_avx2() 170 s_256 = _mm256_slli_epi16(s_256, SGRPROJ_RST_BITS); in svt_get_proj_subspace_avx2() 220 const double u = (double)(dat[i * dat_stride + j] << SGRPROJ_RST_BITS); in svt_get_proj_subspace_avx2() 221 const double s = (double)(src[i * src_stride + j] << SGRPROJ_RST_BITS) - u; in svt_get_proj_subspace_avx2()
|
/dports/multimedia/gstreamer1-plugins-svt-av1/SVT-AV1-3971c982311d49f9355dc8dccdcf8d21b70fa624/Source/Lib/Common/ASM_AVX2/ |
H A D | selfguided_avx2.c | 405 const __m256i rounding = round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 617 const __m256i rounding0 = round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 618 const __m256i rounding1 = round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 784 const __m256i rounding = round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in svt_apply_selfguided_restoration_avx2() 798 const __m256i u_0 = _mm256_slli_epi32(ep_0, SGRPROJ_RST_BITS); in svt_apply_selfguided_restoration_avx2() 799 const __m256i u_1 = _mm256_slli_epi32(ep_1, SGRPROJ_RST_BITS); in svt_apply_selfguided_restoration_avx2() 818 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in svt_apply_selfguided_restoration_avx2() 820 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in svt_apply_selfguided_restoration_avx2() 854 const __m256i u_0 = _mm256_slli_epi32(ep_0, SGRPROJ_RST_BITS); in svt_apply_selfguided_restoration_avx2() 855 const __m256i u_1 = _mm256_slli_epi32(ep_1, SGRPROJ_RST_BITS); in svt_apply_selfguided_restoration_avx2() [all …]
|
/dports/multimedia/svt-av1/SVT-AV1-3971c982311d49f9355dc8dccdcf8d21b70fa624/Source/Lib/Common/ASM_AVX2/ |
H A D | selfguided_avx2.c | 405 const __m256i rounding = round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 617 const __m256i rounding0 = round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 618 const __m256i rounding1 = round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 784 const __m256i rounding = round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in svt_apply_selfguided_restoration_avx2() 798 const __m256i u_0 = _mm256_slli_epi32(ep_0, SGRPROJ_RST_BITS); in svt_apply_selfguided_restoration_avx2() 799 const __m256i u_1 = _mm256_slli_epi32(ep_1, SGRPROJ_RST_BITS); in svt_apply_selfguided_restoration_avx2() 818 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in svt_apply_selfguided_restoration_avx2() 820 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in svt_apply_selfguided_restoration_avx2() 854 const __m256i u_0 = _mm256_slli_epi32(ep_0, SGRPROJ_RST_BITS); in svt_apply_selfguided_restoration_avx2() 855 const __m256i u_1 = _mm256_slli_epi32(ep_1, SGRPROJ_RST_BITS); in svt_apply_selfguided_restoration_avx2() [all …]
|
/dports/graphics/dssim/dssim-3.1.2/cargo-crates/libaom-sys-0.9.1/vendor/av1/common/x86/ |
H A D | selfguided_avx2.c | 329 round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 345 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 501 round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 503 round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 524 SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 541 SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 672 const __m256i u_0 = _mm256_slli_epi32(ep_0, SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_avx2() 673 const __m256i u_1 = _mm256_slli_epi32(ep_1, SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_avx2() 695 round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_avx2() 697 _mm256_add_epi32(v_0, rounding), SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_avx2() [all …]
|
H A D | selfguided_sse4.c | 285 round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 300 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 460 round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 462 round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 481 SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 496 SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 618 const __m128i u = _mm_slli_epi16(src, SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_sse4_1() 642 round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_sse4_1() 644 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_sse4_1() 646 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_sse4_1()
|
/dports/www/firefox-esr/firefox-91.8.0/third_party/aom/av1/common/x86/ |
H A D | selfguided_avx2.c | 329 round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 345 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 501 round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 503 round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 524 SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 541 SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 672 const __m256i u_0 = _mm256_slli_epi32(ep_0, SGRPROJ_RST_BITS); in apply_selfguided_restoration_avx2() 673 const __m256i u_1 = _mm256_slli_epi32(ep_1, SGRPROJ_RST_BITS); in apply_selfguided_restoration_avx2() 695 round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_avx2() 697 _mm256_add_epi32(v_0, rounding), SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_avx2() [all …]
|
H A D | selfguided_sse4.c | 284 round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 299 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 458 round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 460 round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 479 SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 494 SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 616 const __m128i u = _mm_slli_epi16(src, SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1() 640 round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1() 642 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1() 644 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1()
|
/dports/www/qt5-webengine/qtwebengine-everywhere-src-5.15.2/src/3rdparty/chromium/third_party/libaom/source/libaom/av1/common/x86/ |
H A D | selfguided_avx2.c | 329 round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 345 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 501 round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 503 round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 524 SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 541 SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 672 const __m256i u_0 = _mm256_slli_epi32(ep_0, SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_avx2() 673 const __m256i u_1 = _mm256_slli_epi32(ep_1, SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_avx2() 695 round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_avx2() 697 _mm256_add_epi32(v_0, rounding), SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_avx2() [all …]
|
H A D | selfguided_sse4.c | 285 round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 300 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 460 round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 462 round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 481 SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 496 SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 618 const __m128i u = _mm_slli_epi16(src, SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_sse4_1() 642 round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_sse4_1() 644 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_sse4_1() 646 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_sse4_1()
|
/dports/www/firefox/firefox-99.0/third_party/aom/av1/common/x86/ |
H A D | selfguided_avx2.c | 329 round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 345 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 501 round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 503 round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 524 SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 541 SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 672 const __m256i u_0 = _mm256_slli_epi32(ep_0, SGRPROJ_RST_BITS); in apply_selfguided_restoration_avx2() 673 const __m256i u_1 = _mm256_slli_epi32(ep_1, SGRPROJ_RST_BITS); in apply_selfguided_restoration_avx2() 695 round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_avx2() 697 _mm256_add_epi32(v_0, rounding), SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_avx2() [all …]
|
H A D | selfguided_sse4.c | 284 round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 299 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 458 round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 460 round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 479 SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 494 SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 616 const __m128i u = _mm_slli_epi16(src, SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1() 640 round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1() 642 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1() 644 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1()
|
/dports/mail/thunderbird/thunderbird-91.8.0/third_party/aom/av1/common/x86/ |
H A D | selfguided_avx2.c | 329 round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 345 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 501 round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 503 round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 524 SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 541 SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 672 const __m256i u_0 = _mm256_slli_epi32(ep_0, SGRPROJ_RST_BITS); in apply_selfguided_restoration_avx2() 673 const __m256i u_1 = _mm256_slli_epi32(ep_1, SGRPROJ_RST_BITS); in apply_selfguided_restoration_avx2() 695 round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_avx2() 697 _mm256_add_epi32(v_0, rounding), SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_avx2() [all …]
|
H A D | selfguided_sse4.c | 284 round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 299 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 458 round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 460 round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 479 SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 494 SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 616 const __m128i u = _mm_slli_epi16(src, SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1() 640 round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1() 642 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1() 644 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1()
|
/dports/www/chromium-legacy/chromium-88.0.4324.182/third_party/libaom/source/libaom/av1/common/x86/ |
H A D | selfguided_avx2.c | 329 round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 345 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 501 round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 503 round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 524 SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 541 SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 672 const __m256i u_0 = _mm256_slli_epi32(ep_0, SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_avx2() 673 const __m256i u_1 = _mm256_slli_epi32(ep_1, SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_avx2() 695 round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_avx2() 697 _mm256_add_epi32(v_0, rounding), SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_avx2() [all …]
|
H A D | selfguided_sse4.c | 285 round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 300 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 460 round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 462 round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 481 SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 496 SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 618 const __m128i u = _mm_slli_epi16(src, SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_sse4_1() 642 round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_sse4_1() 644 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_sse4_1() 646 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_sse4_1()
|
/dports/multimedia/aom/aom-3.2.0/av1/common/x86/ |
H A D | selfguided_avx2.c | 329 round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 345 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 501 round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 503 round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 524 SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 541 SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 672 const __m256i u_0 = _mm256_slli_epi32(ep_0, SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_avx2() 673 const __m256i u_1 = _mm256_slli_epi32(ep_1, SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_avx2() 695 round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_avx2() 697 _mm256_add_epi32(v_0, rounding), SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_avx2() [all …]
|
H A D | selfguided_sse4.c | 285 round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 300 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 460 round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 462 round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 481 SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 496 SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 618 const __m128i u = _mm_slli_epi16(src, SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_sse4_1() 642 round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_sse4_1() 644 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_sse4_1() 646 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in av1_apply_selfguided_restoration_sse4_1()
|
/dports/lang/spidermonkey78/firefox-78.9.0/third_party/aom/av1/common/x86/ |
H A D | selfguided_avx2.c | 329 round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 345 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 501 round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 503 round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 524 SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 541 SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 672 const __m256i u_0 = _mm256_slli_epi32(ep_0, SGRPROJ_RST_BITS); in apply_selfguided_restoration_avx2() 673 const __m256i u_1 = _mm256_slli_epi32(ep_1, SGRPROJ_RST_BITS); in apply_selfguided_restoration_avx2() 695 round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_avx2() 697 _mm256_add_epi32(v_0, rounding), SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_avx2() [all …]
|
H A D | selfguided_sse4.c | 284 round_for_shift(SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 299 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in final_filter() 458 round_for_shift(SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 460 round_for_shift(SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 479 SGRPROJ_SGR_BITS + nb0 - SGRPROJ_RST_BITS); in final_filter_fast() 494 SGRPROJ_SGR_BITS + nb1 - SGRPROJ_RST_BITS); in final_filter_fast() 616 const __m128i u = _mm_slli_epi16(src, SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1() 640 round_for_shift(SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1() 642 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1() 644 SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1()
|
/dports/lang/spidermonkey60/firefox-60.9.0/third_party/aom/av1/common/x86/ |
H A D | selfguided_sse4.c | 807 (1 << (SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS)) >> 1); in av1_selfguided_restoration_sse4_1() 810 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in av1_selfguided_restoration_sse4_1() 901 const int center = (1 << SGRPROJ_RST_BITS) - 4 * (corner + edge); in av1_highpass_filter_sse4_1() 1088 SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1() 1110 _mm_set1_epi32((1 << (SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS)) >> 1); in apply_selfguided_restoration_sse4_1() 1125 const int32_t u = ((int32_t)dat[l] << SGRPROJ_RST_BITS); in apply_selfguided_restoration_sse4_1() 1518 (1 << (SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS)) >> 1); in av1_selfguided_restoration_highbd_sse4_1() 1521 SGRPROJ_SGR_BITS + nb - SGRPROJ_RST_BITS); in av1_selfguided_restoration_highbd_sse4_1() 1612 const int center = (1 << SGRPROJ_RST_BITS) - 4 * (corner + edge); in av1_highpass_filter_highbd_sse4_1() 1792 _mm_set1_epi32((1 << (SGRPROJ_PRJ_BITS + SGRPROJ_RST_BITS)) >> 1); in apply_selfguided_restoration_highbd_sse4_1() [all …]
|
/dports/graphics/dssim/dssim-3.1.2/cargo-crates/libaom-sys-0.9.1/vendor/av1/encoder/x86/ |
H A D | pickrst_avx2.c | 505 const int32_t shift = SGRPROJ_RST_BITS + SGRPROJ_PRJ_BITS; in av1_lowbd_pixel_proj_error_avx2() 653 __m256i d = _mm256_slli_epi32(u_load, SGRPROJ_RST_BITS); in calc_proj_params_r0_r1_avx2() 654 __m256i s = _mm256_slli_epi32(s_load, SGRPROJ_RST_BITS); in calc_proj_params_r0_r1_avx2() 747 __m256i d = _mm256_slli_epi32(u_load, SGRPROJ_RST_BITS); in calc_proj_params_r0_avx2() 748 __m256i s = _mm256_slli_epi32(s_load, SGRPROJ_RST_BITS); in calc_proj_params_r0_avx2() 806 __m256i d = _mm256_slli_epi32(u_load, SGRPROJ_RST_BITS); in calc_proj_params_r1_avx2() 807 __m256i s = _mm256_slli_epi32(s_load, SGRPROJ_RST_BITS); in calc_proj_params_r1_avx2() 882 __m256i d = _mm256_slli_epi32(u_load, SGRPROJ_RST_BITS); in calc_proj_params_r0_r1_high_bd_avx2() 883 __m256i s = _mm256_slli_epi32(s_load, SGRPROJ_RST_BITS); in calc_proj_params_r0_r1_high_bd_avx2() 1092 const int32_t shift = SGRPROJ_RST_BITS + SGRPROJ_PRJ_BITS; in av1_highbd_pixel_proj_error_avx2() [all …]
|