1 // REQUIRES: arm-registered-target
2 // RUN: %clang_cc1 -triple arm64-apple-ios9 -emit-llvm -o - -fallow-half-arguments-and-returns %s | FileCheck %s --check-prefix=CHECK
3 // RUN: %clang_cc1 -triple armv7-apple-ios9 -emit-llvm -o - -fallow-half-arguments-and-returns %s | FileCheck %s --check-prefix=CHECK
4 // RUN: %clang_cc1 -triple x86_64-apple-macos10.13 -emit-llvm -o - %s | FileCheck %s --check-prefix=CHECK
5 
6 typedef __fp16 half4 __attribute__ ((vector_size (8)));
7 typedef short short4 __attribute__ ((vector_size (8)));
8 
9 half4 hv0, hv1;
10 short4 sv0;
11 
12 // CHECK-LABEL: testFP16Vec0
13 // CHECK: %[[V0:.*]] = load <4 x half>, <4 x half>* @hv0, align 8
14 // CHECK: %[[CONV:.*]] = fpext <4 x half> %[[V0]] to <4 x float>
15 // CHECK: %[[V1:.*]] = load <4 x half>, <4 x half>* @hv1, align 8
16 // CHECK: %[[CONV1:.*]] = fpext <4 x half> %[[V1]] to <4 x float>
17 // CHECK: %[[ADD:.*]] = fadd <4 x float> %[[CONV]], %[[CONV1]]
18 // CHECK: %[[CONV2:.*]] = fptrunc <4 x float> %[[ADD]] to <4 x half>
19 // CHECK: store <4 x half> %[[CONV2]], <4 x half>* @hv0, align 8
20 // CHECK: %[[V2:.*]] = load <4 x half>, <4 x half>* @hv0, align 8
21 // CHECK: %[[CONV3:.*]] = fpext <4 x half> %[[V2]] to <4 x float>
22 // CHECK: %[[V3:.*]] = load <4 x half>, <4 x half>* @hv1, align 8
23 // CHECK: %[[CONV4:.*]] = fpext <4 x half> %[[V3]] to <4 x float>
24 // CHECK: %[[SUB:.*]] = fsub <4 x float> %[[CONV3]], %[[CONV4]]
25 // CHECK: %[[CONV5:.*]] = fptrunc <4 x float> %[[SUB]] to <4 x half>
26 // CHECK: store <4 x half> %[[CONV5]], <4 x half>* @hv0, align 8
27 // CHECK: %[[V4:.*]] = load <4 x half>, <4 x half>* @hv0, align 8
28 // CHECK: %[[CONV6:.*]] = fpext <4 x half> %[[V4]] to <4 x float>
29 // CHECK: %[[V5:.*]] = load <4 x half>, <4 x half>* @hv1, align 8
30 // CHECK: %[[CONV7:.*]] = fpext <4 x half> %[[V5]] to <4 x float>
31 // CHECK: %[[MUL:.*]] = fmul <4 x float> %[[CONV6]], %[[CONV7]]
32 // CHECK: %[[CONV8:.*]] = fptrunc <4 x float> %[[MUL]] to <4 x half>
33 // CHECK: store <4 x half> %[[CONV8]], <4 x half>* @hv0, align 8
34 // CHECK: %[[V6:.*]] = load <4 x half>, <4 x half>* @hv0, align 8
35 // CHECK: %[[CONV9:.*]] = fpext <4 x half> %[[V6]] to <4 x float>
36 // CHECK: %[[V7:.*]] = load <4 x half>, <4 x half>* @hv1, align 8
37 // CHECK: %[[CONV10:.*]] = fpext <4 x half> %[[V7]] to <4 x float>
38 // CHECK: %[[DIV:.*]] = fdiv <4 x float> %[[CONV9]], %[[CONV10]]
39 // CHECK: %[[CONV11:.*]] = fptrunc <4 x float> %[[DIV]] to <4 x half>
40 // CHECK: store <4 x half> %[[CONV11]], <4 x half>* @hv0, align 8
41 
testFP16Vec0()42 void testFP16Vec0() {
43   hv0 = hv0 + hv1;
44   hv0 = hv0 - hv1;
45   hv0 = hv0 * hv1;
46   hv0 = hv0 / hv1;
47 }
48 
49 // CHECK-LABEL: testFP16Vec1
50 // CHECK: %[[V0:.*]] = load <4 x half>, <4 x half>* @hv1, align 8
51 // CHECK: %[[CONV:.*]] = fpext <4 x half> %[[V0]] to <4 x float>
52 // CHECK: %[[V1:.*]] = load <4 x half>, <4 x half>* @hv0, align 8
53 // CHECK: %[[CONV1:.*]] = fpext <4 x half> %[[V1]] to <4 x float>
54 // CHECK: %[[ADD:.*]] = fadd <4 x float> %[[CONV1]], %[[CONV]]
55 // CHECK: %[[CONV2:.*]] = fptrunc <4 x float> %[[ADD]] to <4 x half>
56 // CHECK: store <4 x half> %[[CONV2]], <4 x half>* @hv0, align 8
57 // CHECK: %[[V2:.*]] = load <4 x half>, <4 x half>* @hv1, align 8
58 // CHECK: %[[CONV3:.*]] = fpext <4 x half> %[[V2]] to <4 x float>
59 // CHECK: %[[V3:.*]] = load <4 x half>, <4 x half>* @hv0, align 8
60 // CHECK: %[[CONV4:.*]] = fpext <4 x half> %[[V3]] to <4 x float>
61 // CHECK: %[[SUB:.*]] = fsub <4 x float> %[[CONV4]], %[[CONV3]]
62 // CHECK: %[[CONV5:.*]] = fptrunc <4 x float> %[[SUB]] to <4 x half>
63 // CHECK: store <4 x half> %[[CONV5]], <4 x half>* @hv0, align 8
64 // CHECK: %[[V4:.*]] = load <4 x half>, <4 x half>* @hv1, align 8
65 // CHECK: %[[CONV6:.*]] = fpext <4 x half> %[[V4]] to <4 x float>
66 // CHECK: %[[V5:.*]] = load <4 x half>, <4 x half>* @hv0, align 8
67 // CHECK: %[[CONV7:.*]] = fpext <4 x half> %[[V5]] to <4 x float>
68 // CHECK: %[[MUL:.*]] = fmul <4 x float> %[[CONV7]], %[[CONV6]]
69 // CHECK: %[[CONV8:.*]] = fptrunc <4 x float> %[[MUL]] to <4 x half>
70 // CHECK: store <4 x half> %[[CONV8]], <4 x half>* @hv0, align 8
71 // CHECK: %[[V6:.*]] = load <4 x half>, <4 x half>* @hv1, align 8
72 // CHECK: %[[CONV9:.*]] = fpext <4 x half> %[[V6]] to <4 x float>
73 // CHECK: %[[V7:.*]] = load <4 x half>, <4 x half>* @hv0, align 8
74 // CHECK: %[[CONV10:.*]] = fpext <4 x half> %[[V7]] to <4 x float>
75 // CHECK: %[[DIV:.*]] = fdiv <4 x float> %[[CONV10]], %[[CONV9]]
76 // CHECK: %[[CONV11:.*]] = fptrunc <4 x float> %[[DIV]] to <4 x half>
77 // CHECK: store <4 x half> %[[CONV11]], <4 x half>* @hv0, align 8
78 
testFP16Vec1()79 void testFP16Vec1() {
80   hv0 += hv1;
81   hv0 -= hv1;
82   hv0 *= hv1;
83   hv0 /= hv1;
84 }
85 
86 // CHECK-LABEL: testFP16Vec2
87 // CHECK: %[[CADDR:.*]] = alloca i32, align 4
88 // CHECK: store i32 %[[C:.*]], i32* %[[CADDR]], align 4
89 // CHECK: %[[V0:.*]] = load i32, i32* %[[CADDR]], align 4
90 // CHECK: %[[TOBOOL:.*]] = icmp ne i32 %[[V0]], 0
91 // CHECK: br i1 %[[TOBOOL]], label %{{.*}}, label %{{.*}}
92 //
93 // CHECK: %[[V1:.*]] = load <4 x half>, <4 x half>* @hv0, align 8
94 // CHECK: br label %{{.*}}
95 //
96 // CHECK: %[[V2:.*]] = load <4 x half>, <4 x half>* @hv1, align 8
97 // CHECK: br label %{{.*}}
98 //
99 // CHECK: %[[COND:.*]] = phi <4 x half> [ %[[V1]], %{{.*}} ], [ %[[V2]], %{{.*}} ]
100 // CHECK: store <4 x half> %[[COND]], <4 x half>* @hv0, align 8
101 
testFP16Vec2(int c)102 void testFP16Vec2(int c) {
103   hv0 = c ? hv0 : hv1;
104 }
105 
106 // CHECK-LABEL: testFP16Vec3
107 // CHECK: %[[V0:.*]] = load <4 x half>, <4 x half>* @hv0, align 8
108 // CHECK: %[[CONV:.*]] = fpext <4 x half> %[[V0]] to <4 x float>
109 // CHECK: %[[V1:.*]] = load <4 x half>, <4 x half>* @hv1, align 8
110 // CHECK: %[[CONV1:.*]] = fpext <4 x half> %[[V1]] to <4 x float>
111 // CHECK: %[[CMP:.*]] = fcmp oeq <4 x float> %[[CONV]], %[[CONV1]]
112 // CHECK: %[[SEXT:.*]] = sext <4 x i1> %[[CMP]] to <4 x i32>
113 // CHECK: %[[CONV2:.*]] = trunc <4 x i32> %[[SEXT]] to <4 x i16>
114 // CHECK: store <4 x i16> %[[CONV2]], <4 x i16>* @sv0, align 8
115 // CHECK: %[[V2:.*]] = load <4 x half>, <4 x half>* @hv0, align 8
116 // CHECK: %[[CONV3:.*]] = fpext <4 x half> %[[V2]] to <4 x float>
117 // CHECK: %[[V3:.*]] = load <4 x half>, <4 x half>* @hv1, align 8
118 // CHECK: %[[CONV4:.*]] = fpext <4 x half> %[[V3]] to <4 x float>
119 // CHECK: %[[CMP5:.*]] = fcmp une <4 x float> %[[CONV3]], %[[CONV4]]
120 // CHECK: %[[SEXT6:.*]] = sext <4 x i1> %[[CMP5]] to <4 x i32>
121 // CHECK: %[[CONV7:.*]] = trunc <4 x i32> %[[SEXT6]] to <4 x i16>
122 // CHECK: store <4 x i16> %[[CONV7]], <4 x i16>* @sv0, align 8
123 // CHECK: %[[V4:.*]] = load <4 x half>, <4 x half>* @hv0, align 8
124 // CHECK: %[[CONV8:.*]] = fpext <4 x half> %[[V4]] to <4 x float>
125 // CHECK: %[[V5:.*]] = load <4 x half>, <4 x half>* @hv1, align 8
126 // CHECK: %[[CONV9:.*]] = fpext <4 x half> %[[V5]] to <4 x float>
127 // CHECK: %[[CMP10:.*]] = fcmp olt <4 x float> %[[CONV8]], %[[CONV9]]
128 // CHECK: %[[SEXT11:.*]] = sext <4 x i1> %[[CMP10]] to <4 x i32>
129 // CHECK: %[[CONV12:.*]] = trunc <4 x i32> %[[SEXT11]] to <4 x i16>
130 // CHECK: store <4 x i16> %[[CONV12]], <4 x i16>* @sv0, align 8
131 // CHECK: %[[V6:.*]] = load <4 x half>, <4 x half>* @hv0, align 8
132 // CHECK: %[[CONV13:.*]] = fpext <4 x half> %[[V6]] to <4 x float>
133 // CHECK: %[[V7:.*]] = load <4 x half>, <4 x half>* @hv1, align 8
134 // CHECK: %[[CONV14:.*]] = fpext <4 x half> %[[V7]] to <4 x float>
135 // CHECK: %[[CMP15:.*]] = fcmp ogt <4 x float> %[[CONV13]], %[[CONV14]]
136 // CHECK: %[[SEXT16:.*]] = sext <4 x i1> %[[CMP15]] to <4 x i32>
137 // CHECK: %[[CONV17:.*]] = trunc <4 x i32> %[[SEXT16]] to <4 x i16>
138 // CHECK: store <4 x i16> %[[CONV17]], <4 x i16>* @sv0, align 8
139 // CHECK: %[[V8:.*]] = load <4 x half>, <4 x half>* @hv0, align 8
140 // CHECK: %[[CONV18:.*]] = fpext <4 x half> %[[V8]] to <4 x float>
141 // CHECK: %[[V9:.*]] = load <4 x half>, <4 x half>* @hv1, align 8
142 // CHECK: %[[CONV19:.*]] = fpext <4 x half> %[[V9]] to <4 x float>
143 // CHECK: %[[CMP20:.*]] = fcmp ole <4 x float> %[[CONV18]], %[[CONV19]]
144 // CHECK: %[[SEXT21:.*]] = sext <4 x i1> %[[CMP20]] to <4 x i32>
145 // CHECK: %[[CONV22:.*]] = trunc <4 x i32> %[[SEXT21]] to <4 x i16>
146 // CHECK: store <4 x i16> %[[CONV22]], <4 x i16>* @sv0, align 8
147 // CHECK: %[[V10:.*]] = load <4 x half>, <4 x half>* @hv0, align 8
148 // CHECK: %[[CONV23:.*]] = fpext <4 x half> %[[V10]] to <4 x float>
149 // CHECK: %[[V11:.*]] = load <4 x half>, <4 x half>* @hv1, align 8
150 // CHECK: %[[CONV24:.*]] = fpext <4 x half> %[[V11]] to <4 x float>
151 // CHECK: %[[CMP25:.*]] = fcmp oge <4 x float> %[[CONV23]], %[[CONV24]]
152 // CHECK: %[[SEXT26:.*]] = sext <4 x i1> %[[CMP25]] to <4 x i32>
153 // CHECK: %[[CONV27:.*]] = trunc <4 x i32> %[[SEXT26]] to <4 x i16>
154 // CHECK: store <4 x i16> %[[CONV27]], <4 x i16>* @sv0, align 8
155 
testFP16Vec3()156 void testFP16Vec3() {
157   sv0 = hv0 == hv1;
158   sv0 = hv0 != hv1;
159   sv0 = hv0 < hv1;
160   sv0 = hv0 > hv1;
161   sv0 = hv0 <= hv1;
162   sv0 = hv0 >= hv1;
163 }
164