1 // RUN: %clang_cc1 -verify -fopenmp -x c++ -triple x86_64-unknown-unknown -emit-llvm %s -o - | FileCheck %s
2 // RUN: %clang_cc1 -fopenmp -x c++ -std=c++11 -triple x86_64-unknown-unknown -emit-pch -o %t %s
3 // RUN: %clang_cc1 -fopenmp -x c++ -triple x86_64-unknown-unknown -std=c++11 -include-pch %t -verify %s -emit-llvm -o - | FileCheck %s
4 
5 // RUN: %clang_cc1 -verify -fopenmp-simd -x c++ -triple x86_64-unknown-unknown -emit-llvm %s -o - | FileCheck --check-prefix SIMD-ONLY0 %s
6 // RUN: %clang_cc1 -fopenmp-simd -x c++ -std=c++11 -triple x86_64-unknown-unknown -emit-pch -o %t %s
7 // RUN: %clang_cc1 -fopenmp-simd -x c++ -triple x86_64-unknown-unknown -std=c++11 -include-pch %t -verify %s -emit-llvm -o - | FileCheck --check-prefix SIMD-ONLY0 %s
8 // SIMD-ONLY0-NOT: {{__kmpc|__tgt}}
9 // expected-no-diagnostics
10 #ifndef HEADER
11 #define HEADER
12 
13 void foo();
14 void bar();
15 
16 // CHECK: define{{.*}} void @{{.*}}baz{{.*}}(i32 %n)
baz(int n)17 void baz(int n) {
18   static float a[10];
19   static double b;
20   // CHECK: call i8* @llvm.stacksave()
21   // CHECK: [[A_BUF_SIZE:%.+]] = mul nuw i64 10, [[NUM_ELEMS:%[^,]+]]
22 
23   // float a_buffer[10][n];
24   // CHECK: [[A_BUF:%.+]] = alloca float, i64 [[A_BUF_SIZE]],
25 
26   // double b_buffer[10];
27   // CHECK: [[B_BUF:%.+]] = alloca double, i64 10,
28 #pragma omp for simd reduction(inscan, +:a[:n], b)
29   for (int i = 0; i < 10; ++i) {
30     // CHECK: call void @__kmpc_for_static_init_4(
31     // CHECK: call i8* @llvm.stacksave()
32     // CHECK: store float 0.000000e+00, float* %
33     // CHECK: store double 0.000000e+00, double* [[B_PRIV_ADDR:%.+]],
34     // CHECK: br label %[[DISPATCH:[^,]+]]
35     // CHECK: [[INPUT_PHASE:.+]]:
36     // CHECK: call void @{{.+}}foo{{.+}}()
37 
38     // a_buffer[i][0..n] = a_priv[[0..n];
39     // CHECK: [[BASE_IDX_I:%.+]] = load i32, i32* [[IV_ADDR:%.+]],
40     // CHECK: [[BASE_IDX:%.+]] = zext i32 [[BASE_IDX_I]] to i64
41     // CHECK: [[IDX:%.+]] = mul nsw i64 [[BASE_IDX]], [[NUM_ELEMS]]
42     // CHECK: [[A_BUF_IDX:%.+]] = getelementptr inbounds float, float* [[A_BUF]], i64 [[IDX]]
43     // CHECK: [[A_PRIV:%.+]] = getelementptr inbounds [10 x float], [10 x float]* [[A_PRIV_ADDR:%.+]], i64 0, i64 0
44     // CHECK: [[BYTES:%.+]] = mul nuw i64 [[NUM_ELEMS:%.+]], 4
45     // CHECK: [[DEST:%.+]] = bitcast float* [[A_BUF_IDX]] to i8*
46     // CHECK: [[SRC:%.+]] = bitcast float* [[A_PRIV]] to i8*
47     // CHECK: call void @llvm.memcpy.p0i8.p0i8.i64(i8* {{.*}}[[DEST]], i8* {{.*}}[[SRC]], i64 [[BYTES]], i1 false)
48 
49     // b_buffer[i] = b_priv;
50     // CHECK: [[B_BUF_IDX:%.+]] = getelementptr inbounds double, double* [[B_BUF]], i64 [[BASE_IDX]]
51     // CHECK: [[B_PRIV:%.+]] = load double, double* [[B_PRIV_ADDR]],
52     // CHECK: store double [[B_PRIV]], double* [[B_BUF_IDX]],
53     // CHECK: br label %[[LOOP_CONTINUE:.+]]
54 
55     // CHECK: [[DISPATCH]]:
56     // CHECK: br label %[[INPUT_PHASE]]
57     // CHECK: [[LOOP_CONTINUE]]:
58     // CHECK: call void @llvm.stackrestore(i8* %
59     // CHECK: call void @__kmpc_for_static_fini(
60     // CHECK: call void @__kmpc_barrier(
61     foo();
62 #pragma omp scan inclusive(a[:n], b)
63     // CHECK: [[LOG2_10:%.+]] = call double @llvm.log2.f64(double 1.000000e+01)
64     // CHECK: [[CEIL_LOG2_10:%.+]] = call double @llvm.ceil.f64(double [[LOG2_10]])
65     // CHECK: [[CEIL_LOG2_10_INT:%.+]] = fptoui double [[CEIL_LOG2_10]] to i32
66     // CHECK: br label %[[OUTER_BODY:[^,]+]]
67     // CHECK: [[OUTER_BODY]]:
68     // CHECK: [[K:%.+]] = phi i32 [ 0, %{{.+}} ], [ [[K_NEXT:%.+]], %{{.+}} ]
69     // CHECK: [[K2POW:%.+]] = phi i64 [ 1, %{{.+}} ], [ [[K2POW_NEXT:%.+]], %{{.+}} ]
70     // CHECK: [[CMP:%.+]] = icmp uge i64 9, [[K2POW]]
71     // CHECK: br i1 [[CMP]], label %[[INNER_BODY:[^,]+]], label %[[INNER_EXIT:[^,]+]]
72     // CHECK: [[INNER_BODY]]:
73     // CHECK: [[I:%.+]] = phi i64 [ 9, %[[OUTER_BODY]] ], [ [[I_PREV:%.+]], %{{.+}} ]
74 
75     // a_buffer[i] += a_buffer[i-pow(2, k)];
76     // CHECK: [[IDX:%.+]] = mul nsw i64 [[I]], [[NUM_ELEMS]]
77     // CHECK: [[A_BUF_IDX:%.+]] = getelementptr inbounds float, float* [[A_BUF]], i64 [[IDX]]
78     // CHECK: [[IDX_SUB_K2POW:%.+]] = sub nuw i64 [[I]], [[K2POW]]
79     // CHECK: [[IDX:%.+]] = mul nsw i64 [[IDX_SUB_K2POW]], [[NUM_ELEMS]]
80     // CHECK: [[A_BUF_IDX_SUB_K2POW:%.+]] = getelementptr inbounds float, float* [[A_BUF]], i64 [[IDX]]
81     // CHECK: [[B_BUF_IDX:%.+]] = getelementptr inbounds double, double* [[B_BUF]], i64 [[I]]
82     // CHECK: [[IDX_SUB_K2POW:%.+]] = sub nuw i64 [[I]], [[K2POW]]
83     // CHECK: [[B_BUF_IDX_SUB_K2POW:%.+]] = getelementptr inbounds double, double* [[B_BUF]], i64 [[IDX_SUB_K2POW]]
84     // CHECK: [[A_BUF_END:%.+]] = getelementptr float, float* [[A_BUF_IDX]], i64 [[NUM_ELEMS]]
85     // CHECK: [[ISEMPTY:%.+]] = icmp eq float* [[A_BUF_IDX]], [[A_BUF_END]]
86     // CHECK: br i1 [[ISEMPTY]], label %[[RED_DONE:[^,]+]], label %[[RED_BODY:[^,]+]]
87     // CHECK: [[RED_BODY]]:
88     // CHECK: [[A_BUF_IDX_SUB_K2POW_ELEM:%.+]] = phi float* [ [[A_BUF_IDX_SUB_K2POW]], %[[INNER_BODY]] ], [ [[A_BUF_IDX_SUB_K2POW_NEXT:%.+]], %[[RED_BODY]] ]
89     // CHECK: [[A_BUF_IDX_ELEM:%.+]] = phi float* [ [[A_BUF_IDX]], %[[INNER_BODY]] ], [ [[A_BUF_IDX_NEXT:%.+]], %[[RED_BODY]] ]
90     // CHECK: [[A_BUF_IDX_VAL:%.+]] = load float, float* [[A_BUF_IDX_ELEM]],
91     // CHECK: [[A_BUF_IDX_SUB_K2POW_VAL:%.+]] = load float, float* [[A_BUF_IDX_SUB_K2POW_ELEM]],
92     // CHECK: [[RED:%.+]] = fadd float [[A_BUF_IDX_VAL]], [[A_BUF_IDX_SUB_K2POW_VAL]]
93     // CHECK: store float [[RED]], float* [[A_BUF_IDX_ELEM]],
94     // CHECK: [[A_BUF_IDX_NEXT]] = getelementptr float, float* [[A_BUF_IDX_ELEM]], i32 1
95     // CHECK: [[A_BUF_IDX_SUB_K2POW_NEXT]] = getelementptr float, float* [[A_BUF_IDX_SUB_K2POW_ELEM]], i32 1
96     // CHECK: [[DONE:%.+]] = icmp eq float* [[A_BUF_IDX_NEXT]], [[A_BUF_END]]
97     // CHECK: br i1 [[DONE]], label %[[RED_DONE]], label %[[RED_BODY]]
98     // CHECK: [[RED_DONE]]:
99 
100     // b_buffer[i] += b_buffer[i-pow(2, k)];
101     // CHECK: [[B_BUF_IDX_VAL:%.+]] = load double, double* [[B_BUF_IDX]],
102     // CHECK: [[B_BUF_IDX_SUB_K2POW_VAL:%.+]] = load double, double* [[B_BUF_IDX_SUB_K2POW]],
103     // CHECK: [[RED:%.+]] = fadd double [[B_BUF_IDX_VAL]], [[B_BUF_IDX_SUB_K2POW_VAL]]
104     // CHECK: store double [[RED]], double* [[B_BUF_IDX]],
105 
106     // --i;
107     // CHECK: [[I_PREV:%.+]] = sub nuw i64 [[I]], 1
108     // CHECK: [[CMP:%.+]] = icmp uge i64 [[I_PREV]], [[K2POW]]
109     // CHECK: br i1 [[CMP]], label %[[INNER_BODY]], label %[[INNER_EXIT]]
110     // CHECK: [[INNER_EXIT]]:
111 
112     // ++k;
113     // CHECK: [[K_NEXT]] = add nuw i32 [[K]], 1
114     // k2pow <<= 1;
115     // CHECK: [[K2POW_NEXT]] = shl nuw i64 [[K2POW]], 1
116     // CHECK: [[CMP:%.+]] = icmp ne i32 [[K_NEXT]], [[CEIL_LOG2_10_INT]]
117     // CHECK: br i1 [[CMP]], label %[[OUTER_BODY]], label %[[OUTER_EXIT:[^,]+]]
118     // CHECK: [[OUTER_EXIT]]:
119     bar();
120     // CHECK: call void @__kmpc_for_static_init_4(
121     // CHECK: call i8* @llvm.stacksave()
122     // CHECK: store float 0.000000e+00, float* %
123     // CHECK: store double 0.000000e+00, double* [[B_PRIV_ADDR:%.+]],
124     // CHECK: br label %[[DISPATCH:[^,]+]]
125 
126     // Skip the before scan body.
127     // CHECK: call void @{{.+}}foo{{.+}}()
128 
129     // CHECK: [[EXIT_INSCAN:[^,]+]]:
130     // CHECK: br label %[[LOOP_CONTINUE:[^,]+]]
131 
132     // CHECK: [[DISPATCH]]:
133     // a_priv[[0..n] = a_buffer[i][0..n];
134     // CHECK: [[BASE_IDX_I:%.+]] = load i32, i32* [[IV_ADDR:%.+]],
135     // CHECK: [[BASE_IDX:%.+]] = zext i32 [[BASE_IDX_I]] to i64
136     // CHECK: [[IDX:%.+]] = mul nsw i64 [[BASE_IDX]], [[NUM_ELEMS]]
137     // CHECK: [[A_BUF_IDX:%.+]] = getelementptr inbounds float, float* [[A_BUF]], i64 [[IDX]]
138     // CHECK: [[A_PRIV:%.+]] = getelementptr inbounds [10 x float], [10 x float]* [[A_PRIV_ADDR:%.+]], i64 0, i64 0
139     // CHECK: [[BYTES:%.+]] = mul nuw i64 [[NUM_ELEMS:%.+]], 4
140     // CHECK: [[DEST:%.+]] = bitcast float* [[A_PRIV]] to i8*
141     // CHECK: [[SRC:%.+]] = bitcast float* [[A_BUF_IDX]] to i8*
142     // CHECK: call void @llvm.memcpy.p0i8.p0i8.i64(i8* {{.*}}[[DEST]], i8* {{.*}}[[SRC]], i64 [[BYTES]], i1 false)
143 
144     // b_priv = b_buffer[i];
145     // CHECK: [[B_BUF_IDX:%.+]] = getelementptr inbounds double, double* [[B_BUF]], i64 [[BASE_IDX]]
146     // CHECK: [[B_BUF_IDX_VAL:%.+]] = load double, double* [[B_BUF_IDX]],
147     // CHECK: store double [[B_BUF_IDX_VAL]], double* [[B_PRIV_ADDR]],
148     // CHECK: br label %[[SCAN_PHASE:[^,]+]]
149 
150     // CHECK: [[SCAN_PHASE]]:
151     // CHECK: call void @{{.+}}bar{{.+}}()
152     // CHECK: br label %[[EXIT_INSCAN]]
153 
154     // CHECK: [[LOOP_CONTINUE]]:
155     // CHECK: call void @llvm.stackrestore(i8* %
156     // CHECK: call void @__kmpc_for_static_fini(
157     // CHECK: call void @llvm.stackrestore(i8*
158     // CHECK: call void @__kmpc_barrier(
159   }
160 
161   // CHECK: call i8* @llvm.stacksave()
162   // CHECK: [[A_BUF_SIZE:%.+]] = mul nuw i64 10, [[NUM_ELEMS:%[^,]+]]
163 
164   // float a_buffer[10][n];
165   // CHECK: [[A_BUF:%.+]] = alloca float, i64 [[A_BUF_SIZE]],
166 
167   // double b_buffer[10];
168   // CHECK: [[B_BUF:%.+]] = alloca double, i64 10,
169 #pragma omp for simd reduction(inscan, +:a[:n], b)
170   for (int i = 0; i < 10; ++i) {
171     // CHECK: call void @__kmpc_for_static_init_4(
172     // CHECK: call i8* @llvm.stacksave()
173     // CHECK: store float 0.000000e+00, float* %
174     // CHECK: store double 0.000000e+00, double* [[B_PRIV_ADDR:%.+]],
175     // CHECK: br label %[[DISPATCH:[^,]+]]
176 
177     // Skip the before scan body.
178     // CHECK: call void @{{.+}}foo{{.+}}()
179 
180     // CHECK: [[EXIT_INSCAN:[^,]+]]:
181 
182     // a_buffer[i][0..n] = a_priv[[0..n];
183     // CHECK: [[BASE_IDX_I:%.+]] = load i32, i32* [[IV_ADDR:%.+]],
184     // CHECK: [[BASE_IDX:%.+]] = zext i32 [[BASE_IDX_I]] to i64
185     // CHECK: [[IDX:%.+]] = mul nsw i64 [[BASE_IDX]], [[NUM_ELEMS]]
186     // CHECK: [[A_BUF_IDX:%.+]] = getelementptr inbounds float, float* [[A_BUF]], i64 [[IDX]]
187     // CHECK: [[A_PRIV:%.+]] = getelementptr inbounds [10 x float], [10 x float]* [[A_PRIV_ADDR:%.+]], i64 0, i64 0
188     // CHECK: [[BYTES:%.+]] = mul nuw i64 [[NUM_ELEMS:%.+]], 4
189     // CHECK: [[DEST:%.+]] = bitcast float* [[A_BUF_IDX]] to i8*
190     // CHECK: [[SRC:%.+]] = bitcast float* [[A_PRIV]] to i8*
191     // CHECK: call void @llvm.memcpy.p0i8.p0i8.i64(i8* {{.*}}[[DEST]], i8* {{.*}}[[SRC]], i64 [[BYTES]], i1 false)
192 
193     // b_buffer[i] = b_priv;
194     // CHECK: [[B_BUF_IDX:%.+]] = getelementptr inbounds double, double* [[B_BUF]], i64 [[BASE_IDX]]
195     // CHECK: [[B_PRIV:%.+]] = load double, double* [[B_PRIV_ADDR]],
196     // CHECK: store double [[B_PRIV]], double* [[B_BUF_IDX]],
197     // CHECK: br label %[[LOOP_CONTINUE:[^,]+]]
198 
199     // CHECK: [[DISPATCH]]:
200     // CHECK: br label %[[INPUT_PHASE:[^,]+]]
201 
202     // CHECK: [[INPUT_PHASE]]:
203     // CHECK: call void @{{.+}}bar{{.+}}()
204     // CHECK: br label %[[EXIT_INSCAN]]
205 
206     // CHECK: [[LOOP_CONTINUE]]:
207     // CHECK: call void @llvm.stackrestore(i8* %
208     // CHECK: call void @__kmpc_for_static_fini(
209     // CHECK: call void @__kmpc_barrier(
210     foo();
211 #pragma omp scan exclusive(a[:n], b)
212     // CHECK: [[LOG2_10:%.+]] = call double @llvm.log2.f64(double 1.000000e+01)
213     // CHECK: [[CEIL_LOG2_10:%.+]] = call double @llvm.ceil.f64(double [[LOG2_10]])
214     // CHECK: [[CEIL_LOG2_10_INT:%.+]] = fptoui double [[CEIL_LOG2_10]] to i32
215     // CHECK: br label %[[OUTER_BODY:[^,]+]]
216     // CHECK: [[OUTER_BODY]]:
217     // CHECK: [[K:%.+]] = phi i32 [ 0, %{{.+}} ], [ [[K_NEXT:%.+]], %{{.+}} ]
218     // CHECK: [[K2POW:%.+]] = phi i64 [ 1, %{{.+}} ], [ [[K2POW_NEXT:%.+]], %{{.+}} ]
219     // CHECK: [[CMP:%.+]] = icmp uge i64 9, [[K2POW]]
220     // CHECK: br i1 [[CMP]], label %[[INNER_BODY:[^,]+]], label %[[INNER_EXIT:[^,]+]]
221     // CHECK: [[INNER_BODY]]:
222     // CHECK: [[I:%.+]] = phi i64 [ 9, %[[OUTER_BODY]] ], [ [[I_PREV:%.+]], %{{.+}} ]
223 
224     // a_buffer[i] += a_buffer[i-pow(2, k)];
225     // CHECK: [[IDX:%.+]] = mul nsw i64 [[I]], [[NUM_ELEMS]]
226     // CHECK: [[A_BUF_IDX:%.+]] = getelementptr inbounds float, float* [[A_BUF]], i64 [[IDX]]
227     // CHECK: [[IDX_SUB_K2POW:%.+]] = sub nuw i64 [[I]], [[K2POW]]
228     // CHECK: [[IDX:%.+]] = mul nsw i64 [[IDX_SUB_K2POW]], [[NUM_ELEMS]]
229     // CHECK: [[A_BUF_IDX_SUB_K2POW:%.+]] = getelementptr inbounds float, float* [[A_BUF]], i64 [[IDX]]
230     // CHECK: [[B_BUF_IDX:%.+]] = getelementptr inbounds double, double* [[B_BUF]], i64 [[I]]
231     // CHECK: [[IDX_SUB_K2POW:%.+]] = sub nuw i64 [[I]], [[K2POW]]
232     // CHECK: [[B_BUF_IDX_SUB_K2POW:%.+]] = getelementptr inbounds double, double* [[B_BUF]], i64 [[IDX_SUB_K2POW]]
233     // CHECK: [[A_BUF_END:%.+]] = getelementptr float, float* [[A_BUF_IDX]], i64 [[NUM_ELEMS]]
234     // CHECK: [[ISEMPTY:%.+]] = icmp eq float* [[A_BUF_IDX]], [[A_BUF_END]]
235     // CHECK: br i1 [[ISEMPTY]], label %[[RED_DONE:[^,]+]], label %[[RED_BODY:[^,]+]]
236     // CHECK: [[RED_BODY]]:
237     // CHECK: [[A_BUF_IDX_SUB_K2POW_ELEM:%.+]] = phi float* [ [[A_BUF_IDX_SUB_K2POW]], %[[INNER_BODY]] ], [ [[A_BUF_IDX_SUB_K2POW_NEXT:%.+]], %[[RED_BODY]] ]
238     // CHECK: [[A_BUF_IDX_ELEM:%.+]] = phi float* [ [[A_BUF_IDX]], %[[INNER_BODY]] ], [ [[A_BUF_IDX_NEXT:%.+]], %[[RED_BODY]] ]
239     // CHECK: [[A_BUF_IDX_VAL:%.+]] = load float, float* [[A_BUF_IDX_ELEM]],
240     // CHECK: [[A_BUF_IDX_SUB_K2POW_VAL:%.+]] = load float, float* [[A_BUF_IDX_SUB_K2POW_ELEM]],
241     // CHECK: [[RED:%.+]] = fadd float [[A_BUF_IDX_VAL]], [[A_BUF_IDX_SUB_K2POW_VAL]]
242     // CHECK: store float [[RED]], float* [[A_BUF_IDX_ELEM]],
243     // CHECK: [[A_BUF_IDX_NEXT]] = getelementptr float, float* [[A_BUF_IDX_ELEM]], i32 1
244     // CHECK: [[A_BUF_IDX_SUB_K2POW_NEXT]] = getelementptr float, float* [[A_BUF_IDX_SUB_K2POW_ELEM]], i32 1
245     // CHECK: [[DONE:%.+]] = icmp eq float* [[A_BUF_IDX_NEXT]], [[A_BUF_END]]
246     // CHECK: br i1 [[DONE]], label %[[RED_DONE]], label %[[RED_BODY]]
247     // CHECK: [[RED_DONE]]:
248 
249     // b_buffer[i] += b_buffer[i-pow(2, k)];
250     // CHECK: [[B_BUF_IDX_VAL:%.+]] = load double, double* [[B_BUF_IDX]],
251     // CHECK: [[B_BUF_IDX_SUB_K2POW_VAL:%.+]] = load double, double* [[B_BUF_IDX_SUB_K2POW]],
252     // CHECK: [[RED:%.+]] = fadd double [[B_BUF_IDX_VAL]], [[B_BUF_IDX_SUB_K2POW_VAL]]
253     // CHECK: store double [[RED]], double* [[B_BUF_IDX]],
254 
255     // --i;
256     // CHECK: [[I_PREV:%.+]] = sub nuw i64 [[I]], 1
257     // CHECK: [[CMP:%.+]] = icmp uge i64 [[I_PREV]], [[K2POW]]
258     // CHECK: br i1 [[CMP]], label %[[INNER_BODY]], label %[[INNER_EXIT]]
259     // CHECK: [[INNER_EXIT]]:
260 
261     // ++k;
262     // CHECK: [[K_NEXT]] = add nuw i32 [[K]], 1
263     // k2pow <<= 1;
264     // CHECK: [[K2POW_NEXT]] = shl nuw i64 [[K2POW]], 1
265     // CHECK: [[CMP:%.+]] = icmp ne i32 [[K_NEXT]], [[CEIL_LOG2_10_INT]]
266     // CHECK: br i1 [[CMP]], label %[[OUTER_BODY]], label %[[OUTER_EXIT:[^,]+]]
267     // CHECK: [[OUTER_EXIT]]:
268     bar();
269     // CHECK: call void @__kmpc_for_static_init_4(
270     // CHECK: call i8* @llvm.stacksave()
271     // CHECK: store float 0.000000e+00, float* %
272     // CHECK: store double 0.000000e+00, double* [[B_PRIV_ADDR:%.+]],
273     // CHECK: br label %[[DISPATCH:[^,]+]]
274 
275     // CHECK: [[SCAN_PHASE:.+]]:
276     // CHECK: call void @{{.+}}foo{{.+}}()
277     // CHECK: br label %[[LOOP_CONTINUE:.+]]
278 
279     // CHECK: [[DISPATCH]]:
280     // if (i >0)
281     //   a_priv[[0..n] = a_buffer[i-1][0..n];
282     // CHECK: [[BASE_IDX_I:%.+]] = load i32, i32* [[IV_ADDR:%.+]],
283     // CHECK: [[BASE_IDX:%.+]] = zext i32 [[BASE_IDX_I]] to i64
284     // CHECK: [[CMP:%.+]] = icmp eq i64 [[BASE_IDX]], 0
285     // CHECK: br i1 [[CMP]], label %[[IF_DONE:[^,]+]], label %[[IF_THEN:[^,]+]]
286     // CHECK: [[IF_THEN]]:
287     // CHECK: [[BASE_IDX_SUB_1:%.+]] = sub nuw i64 [[BASE_IDX]], 1
288     // CHECK: [[IDX:%.+]] = mul nsw i64 [[BASE_IDX_SUB_1]], [[NUM_ELEMS]]
289     // CHECK: [[A_BUF_IDX:%.+]] = getelementptr inbounds float, float* [[A_BUF]], i64 [[IDX]]
290     // CHECK: [[A_PRIV:%.+]] = getelementptr inbounds [10 x float], [10 x float]* [[A_PRIV_ADDR:%.+]], i64 0, i64 0
291     // CHECK: [[BYTES:%.+]] = mul nuw i64 [[NUM_ELEMS:%.+]], 4
292     // CHECK: [[DEST:%.+]] = bitcast float* [[A_PRIV]] to i8*
293     // CHECK: [[SRC:%.+]] = bitcast float* [[A_BUF_IDX]] to i8*
294     // CHECK: call void @llvm.memcpy.p0i8.p0i8.i64(i8* {{.*}}[[DEST]], i8* {{.*}}[[SRC]], i64 [[BYTES]], i1 false)
295 
296     // b_priv = b_buffer[i];
297     // CHECK: [[B_BUF_IDX:%.+]] = getelementptr inbounds double, double* [[B_BUF]], i64 [[BASE_IDX_SUB_1]]
298     // CHECK: [[B_BUF_IDX_VAL:%.+]] = load double, double* [[B_BUF_IDX]],
299     // CHECK: store double [[B_BUF_IDX_VAL]], double* [[B_PRIV_ADDR]],
300     // CHECK: br label %[[SCAN_PHASE]]
301 
302     // CHECK: [[LOOP_CONTINUE]]:
303     // CHECK: call void @llvm.stackrestore(i8* %
304     // CHECK: call void @__kmpc_for_static_fini(
305     // CHECK: call void @llvm.stackrestore(i8*
306     // CHECK: call void @__kmpc_barrier(
307   }
308 }
309 
310 #endif
311 // CHECK: !{!"llvm.loop.vectorize.enable", i1 true}
312 
313