1 /*  Copyright (C) 2013-2014  Povilas Kanapickas <povilas@radix.lt>
2 
3     Distributed under the Boost Software License, Version 1.0.
4         (See accompanying file LICENSE_1_0.txt or copy at
5             http://www.boost.org/LICENSE_1_0.txt)
6 */
7 
8 #ifndef LIBSIMDPP_SIMDPP_DETAIL_INSN_SHUFFLE_ZBYTES16_H
9 #define LIBSIMDPP_SIMDPP_DETAIL_INSN_SHUFFLE_ZBYTES16_H
10 
11 #ifndef LIBSIMDPP_SIMD_H
12     #error "This file must be included through simd.h"
13 #endif
14 
15 #include <simdpp/types.h>
16 #include <simdpp/core/bit_andnot.h>
17 #include <simdpp/core/bit_or.h>
18 #include <simdpp/core/blend.h>
19 #include <simdpp/core/cmp_lt.h>
20 #include <simdpp/core/shuffle_bytes16.h>
21 #include <simdpp/core/i_add.h>
22 #include <simdpp/detail/not_implemented.h>
23 
24 namespace simdpp {
25 namespace SIMDPP_ARCH_NAMESPACE {
26 namespace detail {
27 namespace insn {
28 
29 #if _MSC_VER
30 #pragma warning(push)
31 #pragma warning(disable: 4800)
32 #endif
33 
34 static SIMDPP_INL
i_shuffle_zbytes16(const uint8x16 & a,const uint8x16 & b,const uint8x16 & mask)35 uint8x16 i_shuffle_zbytes16(const uint8x16& a, const uint8x16& b, const uint8x16& mask)
36 {
37 #if SIMDPP_USE_NULL
38     uint8x16 r;
39 
40     for (unsigned i = 0; i < 16; i++) {
41         unsigned j = mask.el(i) & 0x0f;
42         unsigned which = mask.el(i) < 0x10;
43         bool zero = mask.el(i) & 0x80;
44         r.el(i) = zero ? 0 : (which ? a.el(j) : b.el(j));
45     }
46     return r;
47 #elif SIMDPP_USE_XOP
48     return _mm_perm_epi8(a.native(), b.native(), mask.native());
49 #elif SIMDPP_USE_SSE4_1
50     uint8<16> sel = _mm_slli_epi16(mask.native(), 3);
51 
52     uint8<16> ai = _mm_shuffle_epi8(a.native(), mask.native());
53     uint8<16> bi = _mm_shuffle_epi8(b.native(), mask.native());
54     return _mm_blendv_epi8(ai.native(), bi.native(), sel.native());
55 #elif SIMDPP_USE_SSSE3
56     mask_int8<16> select_a = cmp_lt((int8<16>) bit_and(mask, 0x1f), 0x10);
57 
58     uint8<16> ai = _mm_shuffle_epi8(a.native(), mask.native());
59     uint8<16> bi = _mm_shuffle_epi8(b.native(), mask.native());
60     return blend(ai, bi, select_a);
61 #elif SIMDPP_USE_NEON
62     return shuffle_bytes16(a, b, mask);
63 #elif SIMDPP_USE_ALTIVEC
64     int8x16 a0 = a, b0 = b;
65     int8x16 zero_mask = mask;
66     zero_mask = shift_r<7>(zero_mask); // shift in the sign bit
67     a0 = i_shuffle_bytes16(a0, b0, mask);
68     a0 = bit_andnot(a0, zero_mask);
69     return a0;
70 #elif SIMDPP_USE_MSA
71     return (v16u8) __msa_vshf_b((v16i8) mask.native(),
72                                 (v16i8) b.native(),
73                                 (v16i8) a.native());
74 #else
75     return SIMDPP_NOT_IMPLEMENTED3(a, b, mask);
76 #endif
77 }
78 
79 #if _MSC_VER
80 #pragma warning(pop)
81 #endif
82 
83 #if SIMDPP_USE_AVX2
84 static SIMDPP_INL
i_shuffle_zbytes16(const uint8x32 & a,const uint8x32 & b,const uint8x32 & mask)85 uint8x32 i_shuffle_zbytes16(const uint8x32& a, const uint8x32& b, const uint8x32& mask)
86 {
87     int8x32 sel, ai, bi, r;
88     sel = mask;
89     sel = _mm256_slli_epi16(sel.native(), 3);
90 
91     ai = _mm256_shuffle_epi8(a.native(), mask.native());
92     bi = _mm256_shuffle_epi8(b.native(), mask.native());
93     r = _mm256_blendv_epi8(ai.native(), bi.native(), sel.native());
94     return r;
95 }
96 #endif
97 
98 #if SIMDPP_USE_AVX512BW
i_shuffle_zbytes16(const uint8<64> & a,const uint8<64> & b,const uint8<64> & mask)99 SIMDPP_INL uint8<64> i_shuffle_zbytes16(const uint8<64>& a, const uint8<64>& b, const uint8<64>& mask)
100 {
101     uint8<64> sel_mask, ai, bi, r;
102     sel_mask = make_uint(0x10);
103     __mmask64 sel = _mm512_test_epi8_mask(mask.native(), sel_mask.native());
104 
105     ai = _mm512_shuffle_epi8(a.native(), mask.native());
106     bi = _mm512_shuffle_epi8(b.native(), mask.native());
107     r = _mm512_mask_blend_epi8(sel, ai.native(), bi.native());
108     return r;
109 }
110 #endif
111 
112 template<unsigned N> SIMDPP_INL
i_shuffle_zbytes16(const uint8<N> & a,const uint8<N> & b,const uint8<N> & mask)113 uint8<N> i_shuffle_zbytes16(const uint8<N>& a, const uint8<N>& b, const uint8<N>& mask)
114 {
115     SIMDPP_VEC_ARRAY_IMPL3(uint8<N>, i_shuffle_zbytes16, a, b, mask);
116 }
117 
118 template<unsigned N> SIMDPP_INL
i_shuffle_zbytes16(const uint16<N> & a,const uint16<N> & b,const uint16<N> & mask)119 uint16<N> i_shuffle_zbytes16(const uint16<N>& a, const uint16<N>& b, const uint16<N>& mask)
120 {
121     return (uint16<N>) i_shuffle_zbytes16(uint8<N*2>(a), uint8<N*2>(b), uint8<N*2>(mask));
122 }
123 template<unsigned N> SIMDPP_INL
i_shuffle_zbytes16(const uint32<N> & a,const uint32<N> & b,const uint32<N> & mask)124 uint32<N> i_shuffle_zbytes16(const uint32<N>& a, const uint32<N>& b, const uint32<N>& mask)
125 {
126     return (uint32<N>) i_shuffle_zbytes16(uint8<N*4>(a), uint8<N*4>(b), uint8<N*4>(mask));
127 }
128 template<unsigned N> SIMDPP_INL
i_shuffle_zbytes16(const uint64<N> & a,const uint64<N> & b,const uint64<N> & mask)129 uint64<N> i_shuffle_zbytes16(const uint64<N>& a, const uint64<N>& b, const uint64<N>& mask)
130 {
131     return (uint64<N>) i_shuffle_zbytes16(uint8<N*8>(a), uint8<N*8>(b), uint8<N*8>(mask));
132 }
133 template<unsigned N> SIMDPP_INL
i_shuffle_zbytes16(const float32<N> & a,const float32<N> & b,const uint32<N> & mask)134 float32<N> i_shuffle_zbytes16(const float32<N>& a, const float32<N>& b, const uint32<N>& mask)
135 {
136     return float32<N>(i_shuffle_zbytes16(uint32<N>(a), uint32<N>(b), mask));
137 }
138 template<unsigned N> SIMDPP_INL
i_shuffle_zbytes16(const float64<N> & a,const float64<N> & b,const uint64<N> & mask)139 float64<N> i_shuffle_zbytes16(const float64<N>& a, const float64<N>& b, const uint64<N>& mask)
140 {
141     return float64<N>(i_shuffle_zbytes16(uint64<N>(a), uint64<N>(b), mask));
142 }
143 
144 
145 } // namespace insn
146 } // namespace detail
147 } // namespace SIMDPP_ARCH_NAMESPACE
148 } // namespace simdpp
149 
150 #endif
151 
152