1 // REQUIRES: aarch64-registered-target
2 // RUN: %clang_cc1 -triple aarch64-none-linux-gnu -target-feature +sve -fallow-half-arguments-and-returns -S -O1 -Werror -Wall -emit-llvm -o - %s | FileCheck %s
3 // RUN: %clang_cc1 -DSVE_OVERLOADED_FORMS -triple aarch64-none-linux-gnu -target-feature +sve -fallow-half-arguments-and-returns -S -O1 -Werror -Wall -emit-llvm -o - %s | FileCheck %s
4 // RUN: %clang_cc1 -triple aarch64-none-linux-gnu -target-feature +sve -fallow-half-arguments-and-returns -S -O1 -Werror -Wall -o - %s >/dev/null 2>%t
5 // RUN: FileCheck --check-prefix=ASM --allow-empty %s <%t
6 
7 // If this check fails please read test/CodeGen/aarch64-sve-intrinsics/README for instructions on how to resolve it.
8 // ASM-NOT: warning
9 #include <arm_sve.h>
10 
11 #ifdef SVE_OVERLOADED_FORMS
12 // A simple used,unused... macro, long enough to represent any SVE builtin.
13 #define SVE_ACLE_FUNC(A1,A2_UNUSED,A3,A4_UNUSED) A1##A3
14 #else
15 #define SVE_ACLE_FUNC(A1,A2,A3,A4) A1##A2##A3##A4
16 #endif
17 
test_svnot_s8_z(svbool_t pg,svint8_t op)18 svint8_t test_svnot_s8_z(svbool_t pg, svint8_t op)
19 {
20   // CHECK-LABEL: test_svnot_s8_z
21   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 16 x i8> @llvm.aarch64.sve.not.nxv16i8(<vscale x 16 x i8> zeroinitializer, <vscale x 16 x i1> %pg, <vscale x 16 x i8> %op)
22   // CHECK: ret <vscale x 16 x i8> %[[INTRINSIC]]
23   return SVE_ACLE_FUNC(svnot,_s8,_z,)(pg, op);
24 }
25 
test_svnot_s16_z(svbool_t pg,svint16_t op)26 svint16_t test_svnot_s16_z(svbool_t pg, svint16_t op)
27 {
28   // CHECK-LABEL: test_svnot_s16_z
29   // CHECK: %[[PG:.*]] = call <vscale x 8 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv8i1(<vscale x 16 x i1> %pg)
30   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 8 x i16> @llvm.aarch64.sve.not.nxv8i16(<vscale x 8 x i16> zeroinitializer, <vscale x 8 x i1> %[[PG]], <vscale x 8 x i16> %op)
31   // CHECK: ret <vscale x 8 x i16> %[[INTRINSIC]]
32   return SVE_ACLE_FUNC(svnot,_s16,_z,)(pg, op);
33 }
34 
test_svnot_s32_z(svbool_t pg,svint32_t op)35 svint32_t test_svnot_s32_z(svbool_t pg, svint32_t op)
36 {
37   // CHECK-LABEL: test_svnot_s32_z
38   // CHECK: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
39   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 4 x i32> @llvm.aarch64.sve.not.nxv4i32(<vscale x 4 x i32> zeroinitializer, <vscale x 4 x i1> %[[PG]], <vscale x 4 x i32> %op)
40   // CHECK: ret <vscale x 4 x i32> %[[INTRINSIC]]
41   return SVE_ACLE_FUNC(svnot,_s32,_z,)(pg, op);
42 }
43 
test_svnot_s64_z(svbool_t pg,svint64_t op)44 svint64_t test_svnot_s64_z(svbool_t pg, svint64_t op)
45 {
46   // CHECK-LABEL: test_svnot_s64_z
47   // CHECK: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
48   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 2 x i64> @llvm.aarch64.sve.not.nxv2i64(<vscale x 2 x i64> zeroinitializer, <vscale x 2 x i1> %[[PG]], <vscale x 2 x i64> %op)
49   // CHECK: ret <vscale x 2 x i64> %[[INTRINSIC]]
50   return SVE_ACLE_FUNC(svnot,_s64,_z,)(pg, op);
51 }
52 
test_svnot_u8_z(svbool_t pg,svuint8_t op)53 svuint8_t test_svnot_u8_z(svbool_t pg, svuint8_t op)
54 {
55   // CHECK-LABEL: test_svnot_u8_z
56   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 16 x i8> @llvm.aarch64.sve.not.nxv16i8(<vscale x 16 x i8> zeroinitializer, <vscale x 16 x i1> %pg, <vscale x 16 x i8> %op)
57   // CHECK: ret <vscale x 16 x i8> %[[INTRINSIC]]
58   return SVE_ACLE_FUNC(svnot,_u8,_z,)(pg, op);
59 }
60 
test_svnot_u16_z(svbool_t pg,svuint16_t op)61 svuint16_t test_svnot_u16_z(svbool_t pg, svuint16_t op)
62 {
63   // CHECK-LABEL: test_svnot_u16_z
64   // CHECK: %[[PG:.*]] = call <vscale x 8 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv8i1(<vscale x 16 x i1> %pg)
65   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 8 x i16> @llvm.aarch64.sve.not.nxv8i16(<vscale x 8 x i16> zeroinitializer, <vscale x 8 x i1> %[[PG]], <vscale x 8 x i16> %op)
66   // CHECK: ret <vscale x 8 x i16> %[[INTRINSIC]]
67   return SVE_ACLE_FUNC(svnot,_u16,_z,)(pg, op);
68 }
69 
test_svnot_u32_z(svbool_t pg,svuint32_t op)70 svuint32_t test_svnot_u32_z(svbool_t pg, svuint32_t op)
71 {
72   // CHECK-LABEL: test_svnot_u32_z
73   // CHECK: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
74   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 4 x i32> @llvm.aarch64.sve.not.nxv4i32(<vscale x 4 x i32> zeroinitializer, <vscale x 4 x i1> %[[PG]], <vscale x 4 x i32> %op)
75   // CHECK: ret <vscale x 4 x i32> %[[INTRINSIC]]
76   return SVE_ACLE_FUNC(svnot,_u32,_z,)(pg, op);
77 }
78 
test_svnot_u64_z(svbool_t pg,svuint64_t op)79 svuint64_t test_svnot_u64_z(svbool_t pg, svuint64_t op)
80 {
81   // CHECK-LABEL: test_svnot_u64_z
82   // CHECK: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
83   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 2 x i64> @llvm.aarch64.sve.not.nxv2i64(<vscale x 2 x i64> zeroinitializer, <vscale x 2 x i1> %[[PG]], <vscale x 2 x i64> %op)
84   // CHECK: ret <vscale x 2 x i64> %[[INTRINSIC]]
85   return SVE_ACLE_FUNC(svnot,_u64,_z,)(pg, op);
86 }
87 
test_svnot_s8_m(svint8_t inactive,svbool_t pg,svint8_t op)88 svint8_t test_svnot_s8_m(svint8_t inactive, svbool_t pg, svint8_t op)
89 {
90   // CHECK-LABEL: test_svnot_s8_m
91   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 16 x i8> @llvm.aarch64.sve.not.nxv16i8(<vscale x 16 x i8> %inactive, <vscale x 16 x i1> %pg, <vscale x 16 x i8> %op)
92   // CHECK: ret <vscale x 16 x i8> %[[INTRINSIC]]
93   return SVE_ACLE_FUNC(svnot,_s8,_m,)(inactive, pg, op);
94 }
95 
test_svnot_s16_m(svint16_t inactive,svbool_t pg,svint16_t op)96 svint16_t test_svnot_s16_m(svint16_t inactive, svbool_t pg, svint16_t op)
97 {
98   // CHECK-LABEL: test_svnot_s16_m
99   // CHECK: %[[PG:.*]] = call <vscale x 8 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv8i1(<vscale x 16 x i1> %pg)
100   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 8 x i16> @llvm.aarch64.sve.not.nxv8i16(<vscale x 8 x i16> %inactive, <vscale x 8 x i1> %[[PG]], <vscale x 8 x i16> %op)
101   // CHECK: ret <vscale x 8 x i16> %[[INTRINSIC]]
102   return SVE_ACLE_FUNC(svnot,_s16,_m,)(inactive, pg, op);
103 }
104 
test_svnot_s32_m(svint32_t inactive,svbool_t pg,svint32_t op)105 svint32_t test_svnot_s32_m(svint32_t inactive, svbool_t pg, svint32_t op)
106 {
107   // CHECK-LABEL: test_svnot_s32_m
108   // CHECK: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
109   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 4 x i32> @llvm.aarch64.sve.not.nxv4i32(<vscale x 4 x i32> %inactive, <vscale x 4 x i1> %[[PG]], <vscale x 4 x i32> %op)
110   // CHECK: ret <vscale x 4 x i32> %[[INTRINSIC]]
111   return SVE_ACLE_FUNC(svnot,_s32,_m,)(inactive, pg, op);
112 }
113 
test_svnot_s64_m(svint64_t inactive,svbool_t pg,svint64_t op)114 svint64_t test_svnot_s64_m(svint64_t inactive, svbool_t pg, svint64_t op)
115 {
116   // CHECK-LABEL: test_svnot_s64_m
117   // CHECK: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
118   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 2 x i64> @llvm.aarch64.sve.not.nxv2i64(<vscale x 2 x i64> %inactive, <vscale x 2 x i1> %[[PG]], <vscale x 2 x i64> %op)
119   // CHECK: ret <vscale x 2 x i64> %[[INTRINSIC]]
120   return SVE_ACLE_FUNC(svnot,_s64,_m,)(inactive, pg, op);
121 }
122 
test_svnot_u8_m(svuint8_t inactive,svbool_t pg,svuint8_t op)123 svuint8_t test_svnot_u8_m(svuint8_t inactive, svbool_t pg, svuint8_t op)
124 {
125   // CHECK-LABEL: test_svnot_u8_m
126   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 16 x i8> @llvm.aarch64.sve.not.nxv16i8(<vscale x 16 x i8> %inactive, <vscale x 16 x i1> %pg, <vscale x 16 x i8> %op)
127   // CHECK: ret <vscale x 16 x i8> %[[INTRINSIC]]
128   return SVE_ACLE_FUNC(svnot,_u8,_m,)(inactive, pg, op);
129 }
130 
test_svnot_u16_m(svuint16_t inactive,svbool_t pg,svuint16_t op)131 svuint16_t test_svnot_u16_m(svuint16_t inactive, svbool_t pg, svuint16_t op)
132 {
133   // CHECK-LABEL: test_svnot_u16_m
134   // CHECK: %[[PG:.*]] = call <vscale x 8 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv8i1(<vscale x 16 x i1> %pg)
135   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 8 x i16> @llvm.aarch64.sve.not.nxv8i16(<vscale x 8 x i16> %inactive, <vscale x 8 x i1> %[[PG]], <vscale x 8 x i16> %op)
136   // CHECK: ret <vscale x 8 x i16> %[[INTRINSIC]]
137   return SVE_ACLE_FUNC(svnot,_u16,_m,)(inactive, pg, op);
138 }
139 
test_svnot_u32_m(svuint32_t inactive,svbool_t pg,svuint32_t op)140 svuint32_t test_svnot_u32_m(svuint32_t inactive, svbool_t pg, svuint32_t op)
141 {
142   // CHECK-LABEL: test_svnot_u32_m
143   // CHECK: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
144   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 4 x i32> @llvm.aarch64.sve.not.nxv4i32(<vscale x 4 x i32> %inactive, <vscale x 4 x i1> %[[PG]], <vscale x 4 x i32> %op)
145   // CHECK: ret <vscale x 4 x i32> %[[INTRINSIC]]
146   return SVE_ACLE_FUNC(svnot,_u32,_m,)(inactive, pg, op);
147 }
148 
test_svnot_u64_m(svuint64_t inactive,svbool_t pg,svuint64_t op)149 svuint64_t test_svnot_u64_m(svuint64_t inactive, svbool_t pg, svuint64_t op)
150 {
151   // CHECK-LABEL: test_svnot_u64_m
152   // CHECK: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
153   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 2 x i64> @llvm.aarch64.sve.not.nxv2i64(<vscale x 2 x i64> %inactive, <vscale x 2 x i1> %[[PG]], <vscale x 2 x i64> %op)
154   // CHECK: ret <vscale x 2 x i64> %[[INTRINSIC]]
155   return SVE_ACLE_FUNC(svnot,_u64,_m,)(inactive, pg, op);
156 }
157 
test_svnot_s8_x(svbool_t pg,svint8_t op)158 svint8_t test_svnot_s8_x(svbool_t pg, svint8_t op)
159 {
160   // CHECK-LABEL: test_svnot_s8_x
161   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 16 x i8> @llvm.aarch64.sve.not.nxv16i8(<vscale x 16 x i8> undef, <vscale x 16 x i1> %pg, <vscale x 16 x i8> %op)
162   // CHECK: ret <vscale x 16 x i8> %[[INTRINSIC]]
163   return SVE_ACLE_FUNC(svnot,_s8,_x,)(pg, op);
164 }
165 
test_svnot_s16_x(svbool_t pg,svint16_t op)166 svint16_t test_svnot_s16_x(svbool_t pg, svint16_t op)
167 {
168   // CHECK-LABEL: test_svnot_s16_x
169   // CHECK: %[[PG:.*]] = call <vscale x 8 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv8i1(<vscale x 16 x i1> %pg)
170   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 8 x i16> @llvm.aarch64.sve.not.nxv8i16(<vscale x 8 x i16> undef, <vscale x 8 x i1> %[[PG]], <vscale x 8 x i16> %op)
171   // CHECK: ret <vscale x 8 x i16> %[[INTRINSIC]]
172   return SVE_ACLE_FUNC(svnot,_s16,_x,)(pg, op);
173 }
174 
test_svnot_s32_x(svbool_t pg,svint32_t op)175 svint32_t test_svnot_s32_x(svbool_t pg, svint32_t op)
176 {
177   // CHECK-LABEL: test_svnot_s32_x
178   // CHECK: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
179   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 4 x i32> @llvm.aarch64.sve.not.nxv4i32(<vscale x 4 x i32> undef, <vscale x 4 x i1> %[[PG]], <vscale x 4 x i32> %op)
180   // CHECK: ret <vscale x 4 x i32> %[[INTRINSIC]]
181   return SVE_ACLE_FUNC(svnot,_s32,_x,)(pg, op);
182 }
183 
test_svnot_s64_x(svbool_t pg,svint64_t op)184 svint64_t test_svnot_s64_x(svbool_t pg, svint64_t op)
185 {
186   // CHECK-LABEL: test_svnot_s64_x
187   // CHECK: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
188   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 2 x i64> @llvm.aarch64.sve.not.nxv2i64(<vscale x 2 x i64> undef, <vscale x 2 x i1> %[[PG]], <vscale x 2 x i64> %op)
189   // CHECK: ret <vscale x 2 x i64> %[[INTRINSIC]]
190   return SVE_ACLE_FUNC(svnot,_s64,_x,)(pg, op);
191 }
192 
test_svnot_u8_x(svbool_t pg,svuint8_t op)193 svuint8_t test_svnot_u8_x(svbool_t pg, svuint8_t op)
194 {
195   // CHECK-LABEL: test_svnot_u8_x
196   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 16 x i8> @llvm.aarch64.sve.not.nxv16i8(<vscale x 16 x i8> undef, <vscale x 16 x i1> %pg, <vscale x 16 x i8> %op)
197   // CHECK: ret <vscale x 16 x i8> %[[INTRINSIC]]
198   return SVE_ACLE_FUNC(svnot,_u8,_x,)(pg, op);
199 }
200 
test_svnot_u16_x(svbool_t pg,svuint16_t op)201 svuint16_t test_svnot_u16_x(svbool_t pg, svuint16_t op)
202 {
203   // CHECK-LABEL: test_svnot_u16_x
204   // CHECK: %[[PG:.*]] = call <vscale x 8 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv8i1(<vscale x 16 x i1> %pg)
205   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 8 x i16> @llvm.aarch64.sve.not.nxv8i16(<vscale x 8 x i16> undef, <vscale x 8 x i1> %[[PG]], <vscale x 8 x i16> %op)
206   // CHECK: ret <vscale x 8 x i16> %[[INTRINSIC]]
207   return SVE_ACLE_FUNC(svnot,_u16,_x,)(pg, op);
208 }
209 
test_svnot_u32_x(svbool_t pg,svuint32_t op)210 svuint32_t test_svnot_u32_x(svbool_t pg, svuint32_t op)
211 {
212   // CHECK-LABEL: test_svnot_u32_x
213   // CHECK: %[[PG:.*]] = call <vscale x 4 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv4i1(<vscale x 16 x i1> %pg)
214   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 4 x i32> @llvm.aarch64.sve.not.nxv4i32(<vscale x 4 x i32> undef, <vscale x 4 x i1> %[[PG]], <vscale x 4 x i32> %op)
215   // CHECK: ret <vscale x 4 x i32> %[[INTRINSIC]]
216   return SVE_ACLE_FUNC(svnot,_u32,_x,)(pg, op);
217 }
218 
test_svnot_u64_x(svbool_t pg,svuint64_t op)219 svuint64_t test_svnot_u64_x(svbool_t pg, svuint64_t op)
220 {
221   // CHECK-LABEL: test_svnot_u64_x
222   // CHECK: %[[PG:.*]] = call <vscale x 2 x i1> @llvm.aarch64.sve.convert.from.svbool.nxv2i1(<vscale x 16 x i1> %pg)
223   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 2 x i64> @llvm.aarch64.sve.not.nxv2i64(<vscale x 2 x i64> undef, <vscale x 2 x i1> %[[PG]], <vscale x 2 x i64> %op)
224   // CHECK: ret <vscale x 2 x i64> %[[INTRINSIC]]
225   return SVE_ACLE_FUNC(svnot,_u64,_x,)(pg, op);
226 }
227 
test_svnot_b_z(svbool_t pg,svbool_t op)228 svbool_t test_svnot_b_z(svbool_t pg, svbool_t op)
229 {
230   // CHECK-LABEL: test_svnot_b_z
231   // CHECK: %[[INTRINSIC:.*]] = call <vscale x 16 x i1> @llvm.aarch64.sve.eor.z.nxv16i1(<vscale x 16 x i1> %pg, <vscale x 16 x i1> %op, <vscale x 16 x i1> %pg)
232   // CHECK: ret <vscale x 16 x i1> %[[INTRINSIC]]
233   return SVE_ACLE_FUNC(svnot,_b,_z,)(pg, op);
234 }
235