1 // REQUIRES: aarch64-registered-target
2 // RUN: %clang_cc1 -triple aarch64-none-linux-gnu -target-feature +sve -fallow-half-arguments-and-returns -S -O1 -Werror -emit-llvm -o - %s | FileCheck %s
3 // RUN: %clang_cc1 -triple aarch64-none-linux-gnu -target-feature +sve -fallow-half-arguments-and-returns -S -O1 -Werror -emit-llvm -o - -x c++ %s | FileCheck %s
4 // RUN: %clang_cc1 -DSVE_OVERLOADED_FORMS -triple aarch64-none-linux-gnu -target-feature +sve -fallow-half-arguments-and-returns -S -O1 -Werror -emit-llvm -o - %s | FileCheck %s
5 // RUN: %clang_cc1 -DSVE_OVERLOADED_FORMS -triple aarch64-none-linux-gnu -target-feature +sve -fallow-half-arguments-and-returns -S -O1 -Werror -emit-llvm -o - -x c++ %s | FileCheck %s
6 // RUN: %clang_cc1 -triple aarch64-none-linux-gnu -target-feature +sve -fallow-half-arguments-and-returns -S -O1 -Werror -o - %s >/dev/null
7 #include <arm_sve.h>
8 
9 #ifdef SVE_OVERLOADED_FORMS
10 // A simple used,unused... macro, long enough to represent any SVE builtin.
11 #define SVE_ACLE_FUNC(A1,A2_UNUSED,A3,A4_UNUSED) A1##A3
12 #else
13 #define SVE_ACLE_FUNC(A1,A2,A3,A4) A1##A2##A3##A4
14 #endif
15 
test_svld1sb_s16(svbool_t pg,const int8_t * base)16 svint16_t test_svld1sb_s16(svbool_t pg, const int8_t *base)
17 {
18   // CHECK-LABEL: test_svld1sb_s16
19   // CHECK: %[[PG:.*]] = call <vscale x 8 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv8i1(<vscale x 16 x i1> %pg)
20   // CHECK: %[[LOAD:.*]] = call <vscale x 8 x i8> @llvm.aarch64.sve.ld1.nxv8i8(<vscale x 8 x i1> %[[PG]], i8* %base)
21   // CHECK: %[[SEXT:.*]] = sext <vscale x 8 x i8> %[[LOAD]] to <vscale x 8 x i16>
22   // CHECK: ret <vscale x 8 x i16> %[[SEXT]]
23   return svld1sb_s16(pg, base);
24 }
25 
test_svld1sb_s32(svbool_t pg,const int8_t * base)26 svint32_t test_svld1sb_s32(svbool_t pg, const int8_t *base)
27 {
28   // CHECK-LABEL: test_svld1sb_s32
29   // CHECK: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
30   // CHECK: %[[LOAD:.*]] = call <vscale x 4 x i8> @llvm.aarch64.sve.ld1.nxv4i8(<vscale x 4 x i1> %[[PG]], i8* %base)
31   // CHECK: %[[SEXT:.*]] = sext <vscale x 4 x i8> %[[LOAD]] to <vscale x 4 x i32>
32   // CHECK: ret <vscale x 4 x i32> %[[SEXT]]
33   return svld1sb_s32(pg, base);
34 }
35 
test_svld1sb_s64(svbool_t pg,const int8_t * base)36 svint64_t test_svld1sb_s64(svbool_t pg, const int8_t *base)
37 {
38   // CHECK-LABEL: test_svld1sb_s64
39   // CHECK: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
40   // CHECK: %[[LOAD:.*]] = call <vscale x 2 x i8> @llvm.aarch64.sve.ld1.nxv2i8(<vscale x 2 x i1> %[[PG]], i8* %base)
41   // CHECK: %[[SEXT:.*]] = sext <vscale x 2 x i8> %[[LOAD]] to <vscale x 2 x i64>
42   // CHECK: ret <vscale x 2 x i64> %[[SEXT]]
43   return svld1sb_s64(pg, base);
44 }
45 
test_svld1sb_u16(svbool_t pg,const int8_t * base)46 svuint16_t test_svld1sb_u16(svbool_t pg, const int8_t *base)
47 {
48   // CHECK-LABEL: test_svld1sb_u16
49   // CHECK: %[[PG:.*]] = call <vscale x 8 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv8i1(<vscale x 16 x i1> %pg)
50   // CHECK: %[[LOAD:.*]] = call <vscale x 8 x i8> @llvm.aarch64.sve.ld1.nxv8i8(<vscale x 8 x i1> %[[PG]], i8* %base)
51   // CHECK: %[[SEXT:.*]] = sext <vscale x 8 x i8> %[[LOAD]] to <vscale x 8 x i16>
52   // CHECK: ret <vscale x 8 x i16> %[[SEXT]]
53   return svld1sb_u16(pg, base);
54 }
55 
test_svld1sb_u32(svbool_t pg,const int8_t * base)56 svuint32_t test_svld1sb_u32(svbool_t pg, const int8_t *base)
57 {
58   // CHECK-LABEL: test_svld1sb_u32
59   // CHECK: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
60   // CHECK: %[[LOAD:.*]] = call <vscale x 4 x i8> @llvm.aarch64.sve.ld1.nxv4i8(<vscale x 4 x i1> %[[PG]], i8* %base)
61   // CHECK: %[[SEXT:.*]] = sext <vscale x 4 x i8> %[[LOAD]] to <vscale x 4 x i32>
62   // CHECK: ret <vscale x 4 x i32> %[[SEXT]]
63   return svld1sb_u32(pg, base);
64 }
65 
test_svld1sb_u64(svbool_t pg,const int8_t * base)66 svuint64_t test_svld1sb_u64(svbool_t pg, const int8_t *base)
67 {
68   // CHECK-LABEL: test_svld1sb_u64
69   // CHECK: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
70   // CHECK: %[[LOAD:.*]] = call <vscale x 2 x i8> @llvm.aarch64.sve.ld1.nxv2i8(<vscale x 2 x i1> %[[PG]], i8* %base)
71   // CHECK: %[[SEXT:.*]] = sext <vscale x 2 x i8> %[[LOAD]] to <vscale x 2 x i64>
72   // CHECK: ret <vscale x 2 x i64> %[[SEXT]]
73   return svld1sb_u64(pg, base);
74 }
75 
test_svld1sb_vnum_s16(svbool_t pg,const int8_t * base,int64_t vnum)76 svint16_t test_svld1sb_vnum_s16(svbool_t pg, const int8_t *base, int64_t vnum)
77 {
78   // CHECK-LABEL: test_svld1sb_vnum_s16
79   // CHECK-DAG: %[[PG:.*]] = call <vscale x 8 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv8i1(<vscale x 16 x i1> %pg)
80   // CHECK-DAG: %[[BASE:.*]] = bitcast i8* %base to <vscale x 8 x i8>*
81   // CHECK-DAG: %[[GEP:.*]] = getelementptr <vscale x 8 x i8>, <vscale x 8 x i8>* %[[BASE]], i64 %vnum, i64 0
82   // CHECK: %[[LOAD:.*]] = call <vscale x 8 x i8> @llvm.aarch64.sve.ld1.nxv8i8(<vscale x 8 x i1> %[[PG]], i8* %[[GEP]])
83   // CHECK: %[[SEXT:.*]] = sext <vscale x 8 x i8> %[[LOAD]] to <vscale x 8 x i16>
84   // CHECK: ret <vscale x 8 x i16> %[[SEXT]]
85   return svld1sb_vnum_s16(pg, base, vnum);
86 }
87 
test_svld1sb_vnum_s32(svbool_t pg,const int8_t * base,int64_t vnum)88 svint32_t test_svld1sb_vnum_s32(svbool_t pg, const int8_t *base, int64_t vnum)
89 {
90   // CHECK-LABEL: test_svld1sb_vnum_s32
91   // CHECK-DAG: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
92   // CHECK-DAG: %[[BASE:.*]] = bitcast i8* %base to <vscale x 4 x i8>*
93   // CHECK-DAG: %[[GEP:.*]] = getelementptr <vscale x 4 x i8>, <vscale x 4 x i8>* %[[BASE]], i64 %vnum, i64 0
94   // CHECK: %[[LOAD:.*]] = call <vscale x 4 x i8> @llvm.aarch64.sve.ld1.nxv4i8(<vscale x 4 x i1> %[[PG]], i8* %[[GEP]])
95   // CHECK: %[[SEXT:.*]] = sext <vscale x 4 x i8> %[[LOAD]] to <vscale x 4 x i32>
96   // CHECK: ret <vscale x 4 x i32> %[[SEXT]]
97   return svld1sb_vnum_s32(pg, base, vnum);
98 }
99 
test_svld1sb_vnum_s64(svbool_t pg,const int8_t * base,int64_t vnum)100 svint64_t test_svld1sb_vnum_s64(svbool_t pg, const int8_t *base, int64_t vnum)
101 {
102   // CHECK-LABEL: test_svld1sb_vnum_s64
103   // CHECK-DAG: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
104   // CHECK-DAG: %[[BASE:.*]] = bitcast i8* %base to <vscale x 2 x i8>*
105   // CHECK-DAG: %[[GEP:.*]] = getelementptr <vscale x 2 x i8>, <vscale x 2 x i8>* %[[BASE]], i64 %vnum, i64 0
106   // CHECK: %[[LOAD:.*]] = call <vscale x 2 x i8> @llvm.aarch64.sve.ld1.nxv2i8(<vscale x 2 x i1> %[[PG]], i8* %[[GEP]])
107   // CHECK: %[[SEXT:.*]] = sext <vscale x 2 x i8> %[[LOAD]] to <vscale x 2 x i64>
108   // CHECK: ret <vscale x 2 x i64> %[[SEXT]]
109   return svld1sb_vnum_s64(pg, base, vnum);
110 }
111 
test_svld1sb_vnum_u16(svbool_t pg,const int8_t * base,int64_t vnum)112 svuint16_t test_svld1sb_vnum_u16(svbool_t pg, const int8_t *base, int64_t vnum)
113 {
114   // CHECK-LABEL: test_svld1sb_vnum_u16
115   // CHECK-DAG: %[[PG:.*]] = call <vscale x 8 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv8i1(<vscale x 16 x i1> %pg)
116   // CHECK-DAG: %[[BASE:.*]] = bitcast i8* %base to <vscale x 8 x i8>*
117   // CHECK-DAG: %[[GEP:.*]] = getelementptr <vscale x 8 x i8>, <vscale x 8 x i8>* %[[BASE]], i64 %vnum, i64 0
118   // CHECK: %[[LOAD:.*]] = call <vscale x 8 x i8> @llvm.aarch64.sve.ld1.nxv8i8(<vscale x 8 x i1> %[[PG]], i8* %[[GEP]])
119   // CHECK: %[[SEXT:.*]] = sext <vscale x 8 x i8> %[[LOAD]] to <vscale x 8 x i16>
120   // CHECK: ret <vscale x 8 x i16> %[[SEXT]]
121   return svld1sb_vnum_u16(pg, base, vnum);
122 }
123 
test_svld1sb_vnum_u32(svbool_t pg,const int8_t * base,int64_t vnum)124 svuint32_t test_svld1sb_vnum_u32(svbool_t pg, const int8_t *base, int64_t vnum)
125 {
126   // CHECK-LABEL: test_svld1sb_vnum_u32
127   // CHECK-DAG: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
128   // CHECK-DAG: %[[BASE:.*]] = bitcast i8* %base to <vscale x 4 x i8>*
129   // CHECK-DAG: %[[GEP:.*]] = getelementptr <vscale x 4 x i8>, <vscale x 4 x i8>* %[[BASE]], i64 %vnum, i64 0
130   // CHECK: %[[LOAD:.*]] = call <vscale x 4 x i8> @llvm.aarch64.sve.ld1.nxv4i8(<vscale x 4 x i1> %[[PG]], i8* %[[GEP]])
131   // CHECK: %[[SEXT:.*]] = sext <vscale x 4 x i8> %[[LOAD]] to <vscale x 4 x i32>
132   // CHECK: ret <vscale x 4 x i32> %[[SEXT]]
133   return svld1sb_vnum_u32(pg, base, vnum);
134 }
135 
test_svld1sb_vnum_u64(svbool_t pg,const int8_t * base,int64_t vnum)136 svuint64_t test_svld1sb_vnum_u64(svbool_t pg, const int8_t *base, int64_t vnum)
137 {
138   // CHECK-LABEL: test_svld1sb_vnum_u64
139   // CHECK-DAG: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
140   // CHECK-DAG: %[[BASE:.*]] = bitcast i8* %base to <vscale x 2 x i8>*
141   // CHECK-DAG: %[[GEP:.*]] = getelementptr <vscale x 2 x i8>, <vscale x 2 x i8>* %[[BASE]], i64 %vnum, i64 0
142   // CHECK: %[[LOAD:.*]] = call <vscale x 2 x i8> @llvm.aarch64.sve.ld1.nxv2i8(<vscale x 2 x i1> %[[PG]], i8* %[[GEP]])
143   // CHECK: %[[SEXT:.*]] = sext <vscale x 2 x i8> %[[LOAD]] to <vscale x 2 x i64>
144   // CHECK: ret <vscale x 2 x i64> %[[SEXT]]
145   return svld1sb_vnum_u64(pg, base, vnum);
146 }
147 
test_svld1sb_gather_u32base_s32(svbool_t pg,svuint32_t bases)148 svint32_t test_svld1sb_gather_u32base_s32(svbool_t pg, svuint32_t bases) {
149   // CHECK-LABEL: test_svld1sb_gather_u32base_s32
150   // CHECK: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
151   // CHECK: %[[LOAD:.*]] = call <vscale x 4 x i8> @llvm.aarch64.sve.ld1.gather.scalar.offset.nxv4i8.nxv4i32(<vscale x 4 x i1> %[[PG]], <vscale x 4 x i32> %bases, i64 0)
152   // CHECK: %[[SEXT:.*]] = sext <vscale x 4 x i8> %[[LOAD]] to <vscale x 4 x i32>
153   // CHECK: ret <vscale x 4 x i32> %[[SEXT]]
154   return SVE_ACLE_FUNC(svld1sb_gather, _u32base, _s32, )(pg, bases);
155 }
156 
test_svld1sb_gather_u64base_s64(svbool_t pg,svuint64_t bases)157 svint64_t test_svld1sb_gather_u64base_s64(svbool_t pg, svuint64_t bases) {
158   // CHECK-LABEL: test_svld1sb_gather_u64base_s64
159   // CHECK: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
160   // CHECK: %[[LOAD:.*]] = call <vscale x 2 x i8> @llvm.aarch64.sve.ld1.gather.scalar.offset.nxv2i8.nxv2i64(<vscale x 2 x i1> %[[PG]], <vscale x 2 x i64> %bases, i64 0)
161   // CHECK: %[[SEXT:.*]] = sext <vscale x 2 x i8> %[[LOAD]] to <vscale x 2 x i64>
162   // CHECK: ret <vscale x 2 x i64> %[[SEXT]]
163   return SVE_ACLE_FUNC(svld1sb_gather, _u64base, _s64, )(pg, bases);
164 }
165 
test_svld1sb_gather_u32base_u32(svbool_t pg,svuint32_t bases)166 svuint32_t test_svld1sb_gather_u32base_u32(svbool_t pg, svuint32_t bases) {
167   // CHECK-LABEL: test_svld1sb_gather_u32base_u32
168   // CHECK: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
169   // CHECK: %[[LOAD:.*]] = call <vscale x 4 x i8> @llvm.aarch64.sve.ld1.gather.scalar.offset.nxv4i8.nxv4i32(<vscale x 4 x i1> %[[PG]], <vscale x 4 x i32> %bases, i64 0)
170   // CHECK: %[[SEXT:.*]] = sext <vscale x 4 x i8> %[[LOAD]] to <vscale x 4 x i32>
171   // CHECK: ret <vscale x 4 x i32> %[[SEXT]]
172   return SVE_ACLE_FUNC(svld1sb_gather, _u32base, _u32, )(pg, bases);
173 }
174 
test_svld1sb_gather_u64base_u64(svbool_t pg,svuint64_t bases)175 svuint64_t test_svld1sb_gather_u64base_u64(svbool_t pg, svuint64_t bases) {
176   // CHECK-LABEL: test_svld1sb_gather_u64base_u64
177   // CHECK: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
178   // CHECK: %[[LOAD:.*]] = call <vscale x 2 x i8> @llvm.aarch64.sve.ld1.gather.scalar.offset.nxv2i8.nxv2i64(<vscale x 2 x i1> %[[PG]], <vscale x 2 x i64> %bases, i64 0)
179   // CHECK: %[[SEXT:.*]] = sext <vscale x 2 x i8> %[[LOAD]] to <vscale x 2 x i64>
180   // CHECK: ret <vscale x 2 x i64> %[[SEXT]]
181   return SVE_ACLE_FUNC(svld1sb_gather, _u64base, _u64, )(pg, bases);
182 }
183 
test_svld1sb_gather_s32offset_s32(svbool_t pg,const int8_t * base,svint32_t offsets)184 svint32_t test_svld1sb_gather_s32offset_s32(svbool_t pg, const int8_t *base, svint32_t offsets) {
185   // CHECK-LABEL: test_svld1sb_gather_s32offset_s32
186   // CHECK: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
187   // CHECK: %[[LOAD:.*]] = call <vscale x 4 x i8> @llvm.aarch64.sve.ld1.gather.sxtw.nxv4i8(<vscale x 4 x i1> %[[PG]], i8* %base, <vscale x 4 x i32> %offsets)
188   // CHECK: %[[SEXT:.*]] = sext <vscale x 4 x i8> %[[LOAD]] to <vscale x 4 x i32>
189   // CHECK: ret <vscale x 4 x i32> %[[SEXT]]
190   return SVE_ACLE_FUNC(svld1sb_gather_, s32, offset_s32, )(pg, base, offsets);
191 }
192 
test_svld1sb_gather_s64offset_s64(svbool_t pg,const int8_t * base,svint64_t offsets)193 svint64_t test_svld1sb_gather_s64offset_s64(svbool_t pg, const int8_t *base, svint64_t offsets) {
194   // CHECK-LABEL: test_svld1sb_gather_s64offset_s64
195   // CHECK: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
196   // CHECK: %[[LOAD:.*]] = call <vscale x 2 x i8> @llvm.aarch64.sve.ld1.gather.nxv2i8(<vscale x 2 x i1> %[[PG]], i8* %base, <vscale x 2 x i64> %offsets)
197   // CHECK: %[[SEXT:.*]] = sext <vscale x 2 x i8> %[[LOAD]] to <vscale x 2 x i64>
198   // CHECK: ret <vscale x 2 x i64> %[[SEXT]]
199   return SVE_ACLE_FUNC(svld1sb_gather_, s64, offset_s64, )(pg, base, offsets);
200 }
201 
test_svld1sb_gather_s32offset_u32(svbool_t pg,const int8_t * base,svint32_t offsets)202 svuint32_t test_svld1sb_gather_s32offset_u32(svbool_t pg, const int8_t *base, svint32_t offsets) {
203   // CHECK-LABEL: test_svld1sb_gather_s32offset_u32
204   // CHECK: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
205   // CHECK: %[[LOAD:.*]] = call <vscale x 4 x i8> @llvm.aarch64.sve.ld1.gather.sxtw.nxv4i8(<vscale x 4 x i1> %[[PG]], i8* %base, <vscale x 4 x i32> %offsets)
206   // CHECK: %[[SEXT:.*]] = sext <vscale x 4 x i8> %[[LOAD]] to <vscale x 4 x i32>
207   // CHECK: ret <vscale x 4 x i32> %[[SEXT]]
208   return SVE_ACLE_FUNC(svld1sb_gather_, s32, offset_u32, )(pg, base, offsets);
209 }
210 
test_svld1sb_gather_s64offset_u64(svbool_t pg,const int8_t * base,svint64_t offsets)211 svuint64_t test_svld1sb_gather_s64offset_u64(svbool_t pg, const int8_t *base, svint64_t offsets) {
212   // CHECK-LABEL: test_svld1sb_gather_s64offset_u64
213   // CHECK: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
214   // CHECK: %[[LOAD:.*]] = call <vscale x 2 x i8> @llvm.aarch64.sve.ld1.gather.nxv2i8(<vscale x 2 x i1> %[[PG]], i8* %base, <vscale x 2 x i64> %offsets)
215   // CHECK: %[[SEXT:.*]] = sext <vscale x 2 x i8> %[[LOAD]] to <vscale x 2 x i64>
216   // CHECK: ret <vscale x 2 x i64> %[[SEXT]]
217   return SVE_ACLE_FUNC(svld1sb_gather_, s64, offset_u64, )(pg, base, offsets);
218 }
219 
test_svld1sb_gather_u32offset_s32(svbool_t pg,const int8_t * base,svuint32_t offsets)220 svint32_t test_svld1sb_gather_u32offset_s32(svbool_t pg, const int8_t *base, svuint32_t offsets) {
221   // CHECK-LABEL: test_svld1sb_gather_u32offset_s32
222   // CHECK: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
223   // CHECK: %[[LOAD:.*]] = call <vscale x 4 x i8> @llvm.aarch64.sve.ld1.gather.uxtw.nxv4i8(<vscale x 4 x i1> %[[PG]], i8* %base, <vscale x 4 x i32> %offsets)
224   // CHECK: %[[SEXT:.*]] = sext <vscale x 4 x i8> %[[LOAD]] to <vscale x 4 x i32>
225   // CHECK: ret <vscale x 4 x i32> %[[SEXT]]
226   return SVE_ACLE_FUNC(svld1sb_gather_, u32, offset_s32, )(pg, base, offsets);
227 }
228 
test_svld1sb_gather_u64offset_s64(svbool_t pg,const int8_t * base,svuint64_t offsets)229 svint64_t test_svld1sb_gather_u64offset_s64(svbool_t pg, const int8_t *base, svuint64_t offsets) {
230   // CHECK-LABEL: test_svld1sb_gather_u64offset_s64
231   // CHECK: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
232   // CHECK: %[[LOAD:.*]] = call <vscale x 2 x i8> @llvm.aarch64.sve.ld1.gather.nxv2i8(<vscale x 2 x i1> %[[PG]], i8* %base, <vscale x 2 x i64> %offsets)
233   // CHECK: %[[SEXT:.*]] = sext <vscale x 2 x i8> %[[LOAD]] to <vscale x 2 x i64>
234   // CHECK: ret <vscale x 2 x i64> %[[SEXT]]
235   return SVE_ACLE_FUNC(svld1sb_gather_, u64, offset_s64, )(pg, base, offsets);
236 }
237 
test_svld1sb_gather_u32offset_u32(svbool_t pg,const int8_t * base,svuint32_t offsets)238 svuint32_t test_svld1sb_gather_u32offset_u32(svbool_t pg, const int8_t *base, svuint32_t offsets) {
239   // CHECK-LABEL: test_svld1sb_gather_u32offset_u32
240   // CHECK: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
241   // CHECK: %[[LOAD:.*]] = call <vscale x 4 x i8> @llvm.aarch64.sve.ld1.gather.uxtw.nxv4i8(<vscale x 4 x i1> %[[PG]], i8* %base, <vscale x 4 x i32> %offsets)
242   // CHECK: %[[SEXT:.*]] = sext <vscale x 4 x i8> %[[LOAD]] to <vscale x 4 x i32>
243   // CHECK: ret <vscale x 4 x i32> %[[SEXT]]
244   return SVE_ACLE_FUNC(svld1sb_gather_, u32, offset_u32, )(pg, base, offsets);
245 }
246 
test_svld1sb_gather_u64offset_u64(svbool_t pg,const int8_t * base,svuint64_t offsets)247 svuint64_t test_svld1sb_gather_u64offset_u64(svbool_t pg, const int8_t *base, svuint64_t offsets) {
248   // CHECK-LABEL: test_svld1sb_gather_u64offset_u64
249   // CHECK: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
250   // CHECK: %[[LOAD:.*]] = call <vscale x 2 x i8> @llvm.aarch64.sve.ld1.gather.nxv2i8(<vscale x 2 x i1> %[[PG]], i8* %base, <vscale x 2 x i64> %offsets)
251   // CHECK: %[[SEXT:.*]] = sext <vscale x 2 x i8> %[[LOAD]] to <vscale x 2 x i64>
252   // CHECK: ret <vscale x 2 x i64> %[[SEXT]]
253   return SVE_ACLE_FUNC(svld1sb_gather_, u64, offset_u64, )(pg, base, offsets);
254 }
255 
test_svld1sb_gather_u32base_offset_s32(svbool_t pg,svuint32_t bases,int64_t offset)256 svint32_t test_svld1sb_gather_u32base_offset_s32(svbool_t pg, svuint32_t bases, int64_t offset) {
257   // CHECK-LABEL: test_svld1sb_gather_u32base_offset_s32
258   // CHECK: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
259   // CHECK: %[[LOAD:.*]] = call <vscale x 4 x i8> @llvm.aarch64.sve.ld1.gather.scalar.offset.nxv4i8.nxv4i32(<vscale x 4 x i1> %[[PG]], <vscale x 4 x i32> %bases, i64 %offset)
260   // CHECK: %[[SEXT:.*]] = sext <vscale x 4 x i8> %[[LOAD]] to <vscale x 4 x i32>
261   // CHECK: ret <vscale x 4 x i32> %[[SEXT]]
262   return SVE_ACLE_FUNC(svld1sb_gather, _u32base, _offset_s32, )(pg, bases, offset);
263 }
264 
test_svld1sb_gather_u64base_offset_s64(svbool_t pg,svuint64_t bases,int64_t offset)265 svint64_t test_svld1sb_gather_u64base_offset_s64(svbool_t pg, svuint64_t bases, int64_t offset) {
266   // CHECK-LABEL: test_svld1sb_gather_u64base_offset_s64
267   // CHECK: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
268   // CHECK: %[[LOAD:.*]] = call <vscale x 2 x i8> @llvm.aarch64.sve.ld1.gather.scalar.offset.nxv2i8.nxv2i64(<vscale x 2 x i1> %[[PG]], <vscale x 2 x i64> %bases, i64 %offset)
269   // CHECK: %[[SEXT:.*]] = sext <vscale x 2 x i8> %[[LOAD]] to <vscale x 2 x i64>
270   // CHECK: ret <vscale x 2 x i64> %[[SEXT]]
271   return SVE_ACLE_FUNC(svld1sb_gather, _u64base, _offset_s64, )(pg, bases, offset);
272 }
273 
test_svld1sb_gather_u32base_offset_u32(svbool_t pg,svuint32_t bases,int64_t offset)274 svuint32_t test_svld1sb_gather_u32base_offset_u32(svbool_t pg, svuint32_t bases, int64_t offset) {
275   // CHECK-LABEL: test_svld1sb_gather_u32base_offset_u32
276   // CHECK: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
277   // CHECK: %[[LOAD:.*]] = call <vscale x 4 x i8> @llvm.aarch64.sve.ld1.gather.scalar.offset.nxv4i8.nxv4i32(<vscale x 4 x i1> %[[PG]], <vscale x 4 x i32> %bases, i64 %offset)
278   // CHECK: %[[SEXT:.*]] = sext <vscale x 4 x i8> %[[LOAD]] to <vscale x 4 x i32>
279   // CHECK: ret <vscale x 4 x i32> %[[SEXT]]
280   return SVE_ACLE_FUNC(svld1sb_gather, _u32base, _offset_u32, )(pg, bases, offset);
281 }
282 
test_svld1sb_gather_u64base_offset_u64(svbool_t pg,svuint64_t bases,int64_t offset)283 svuint64_t test_svld1sb_gather_u64base_offset_u64(svbool_t pg, svuint64_t bases, int64_t offset) {
284   // CHECK-LABEL: test_svld1sb_gather_u64base_offset_u64
285   // CHECK: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
286   // CHECK: %[[LOAD:.*]] = call <vscale x 2 x i8> @llvm.aarch64.sve.ld1.gather.scalar.offset.nxv2i8.nxv2i64(<vscale x 2 x i1> %[[PG]], <vscale x 2 x i64> %bases, i64 %offset)
287   // CHECK: %[[SEXT:.*]] = sext <vscale x 2 x i8> %[[LOAD]] to <vscale x 2 x i64>
288   // CHECK: ret <vscale x 2 x i64> %[[SEXT]]
289   return SVE_ACLE_FUNC(svld1sb_gather, _u64base, _offset_u64, )(pg, bases, offset);
290 }
291