1; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -verify-machineinstrs < %s | FileCheck -check-prefixes=GCN,SI,FUNC %s
2; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=tonga -verify-machineinstrs < %s | FileCheck -check-prefixes=GCN,VI,FUNC %s
3; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=gfx900 -verify-machineinstrs < %s | FileCheck -check-prefixes=GCN,GFX9,FUNC %s
4; RUN: llc -amdgpu-scalarize-global-loads=false -march=r600 -mcpu=cypress -verify-machineinstrs < %s | FileCheck -check-prefixes=EG,FUNC %s
5
6; FUNC-LABEL: {{^}}s_uaddo_i64_zext:
7; GCN: s_add_u32
8; GCN: s_addc_u32
9; GCN: v_cmp_lt_u64_e32 vcc
10
11; EG: ADDC_UINT
12; EG: ADDC_UINT
13define amdgpu_kernel void @s_uaddo_i64_zext(i64 addrspace(1)* %out, i64 %a, i64 %b) #0 {
14  %uadd = call { i64, i1 } @llvm.uadd.with.overflow.i64(i64 %a, i64 %b)
15  %val = extractvalue { i64, i1 } %uadd, 0
16  %carry = extractvalue { i64, i1 } %uadd, 1
17  %ext = zext i1 %carry to i64
18  %add2 = add i64 %val, %ext
19  store i64 %add2, i64 addrspace(1)* %out, align 8
20  ret void
21}
22
23; FIXME: Could do scalar
24
25; FUNC-LABEL: {{^}}s_uaddo_i32:
26; SI: v_add_i32_e32 v{{[0-9]+}}, vcc, s{{[0-9]+}}, v{{[0-9]+}}
27; VI: v_add_u32_e32 v{{[0-9]+}}, vcc, s{{[0-9]+}}, v{{[0-9]+}}
28; GFX9: v_add_co_u32_e32 v{{[0-9]+}}, vcc, s{{[0-9]+}}, v{{[0-9]+}}
29
30; GCN: v_cndmask_b32_e64 v{{[0-9]+}}, 0, 1, vcc
31
32; EG: ADDC_UINT
33; EG: ADD_INT
34define amdgpu_kernel void @s_uaddo_i32(i32 addrspace(1)* %out, i1 addrspace(1)* %carryout, i32 %a, i32 %b) #0 {
35  %uadd = call { i32, i1 } @llvm.uadd.with.overflow.i32(i32 %a, i32 %b)
36  %val = extractvalue { i32, i1 } %uadd, 0
37  %carry = extractvalue { i32, i1 } %uadd, 1
38  store i32 %val, i32 addrspace(1)* %out, align 4
39  store i1 %carry, i1 addrspace(1)* %carryout
40  ret void
41}
42
43; FUNC-LABEL: {{^}}v_uaddo_i32:
44; SI: v_add_i32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
45; VI: v_add_u32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
46; GFX9: v_add_co_u32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
47
48; GCN: v_cndmask_b32_e64 v{{[0-9]+}}, 0, 1, vcc
49
50; EG: ADDC_UINT
51; EG: ADD_INT
52define amdgpu_kernel void @v_uaddo_i32(i32 addrspace(1)* %out, i1 addrspace(1)* %carryout, i32 addrspace(1)* %a.ptr, i32 addrspace(1)* %b.ptr) #0 {
53  %tid = call i32 @llvm.amdgcn.workitem.id.x()
54  %tid.ext = sext i32 %tid to i64
55  %a.gep = getelementptr inbounds i32, i32 addrspace(1)* %a.ptr
56  %b.gep = getelementptr inbounds i32, i32 addrspace(1)* %b.ptr
57  %a = load i32, i32 addrspace(1)* %a.gep, align 4
58  %b = load i32, i32 addrspace(1)* %b.gep, align 4
59  %uadd = call { i32, i1 } @llvm.uadd.with.overflow.i32(i32 %a, i32 %b)
60  %val = extractvalue { i32, i1 } %uadd, 0
61  %carry = extractvalue { i32, i1 } %uadd, 1
62  store i32 %val, i32 addrspace(1)* %out, align 4
63  store i1 %carry, i1 addrspace(1)* %carryout
64  ret void
65}
66
67; FUNC-LABEL: {{^}}v_uaddo_i32_novcc:
68; SI: v_add_i32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
69; VI: v_add_u32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
70; GFX9: v_add_co_u32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
71
72; GCN: v_cndmask_b32_e64 v{{[0-9]+}}, 0, 1, vcc
73
74; EG: ADDC_UINT
75; EG: ADD_INT
76define amdgpu_kernel void @v_uaddo_i32_novcc(i32 addrspace(1)* %out, i1 addrspace(1)* %carryout, i32 addrspace(1)* %a.ptr, i32 addrspace(1)* %b.ptr) #0 {
77  %tid = call i32 @llvm.amdgcn.workitem.id.x()
78  %tid.ext = sext i32 %tid to i64
79  %a.gep = getelementptr inbounds i32, i32 addrspace(1)* %a.ptr
80  %b.gep = getelementptr inbounds i32, i32 addrspace(1)* %b.ptr
81  %a = load i32, i32 addrspace(1)* %a.gep, align 4
82  %b = load i32, i32 addrspace(1)* %b.gep, align 4
83  %uadd = call { i32, i1 } @llvm.uadd.with.overflow.i32(i32 %a, i32 %b)
84  %val = extractvalue { i32, i1 } %uadd, 0
85  %carry = extractvalue { i32, i1 } %uadd, 1
86  store volatile i32 %val, i32 addrspace(1)* %out, align 4
87  call void asm sideeffect "", "~{VCC}"() #0
88  store volatile i1 %carry, i1 addrspace(1)* %carryout
89  ret void
90}
91
92; FUNC-LABEL: {{^}}s_uaddo_i64:
93; GCN: s_add_u32
94; GCN: s_addc_u32
95
96; EG: ADDC_UINT
97; EG: ADD_INT
98define amdgpu_kernel void @s_uaddo_i64(i64 addrspace(1)* %out, i1 addrspace(1)* %carryout, i64 %a, i64 %b) #0 {
99  %uadd = call { i64, i1 } @llvm.uadd.with.overflow.i64(i64 %a, i64 %b)
100  %val = extractvalue { i64, i1 } %uadd, 0
101  %carry = extractvalue { i64, i1 } %uadd, 1
102  store i64 %val, i64 addrspace(1)* %out, align 8
103  store i1 %carry, i1 addrspace(1)* %carryout
104  ret void
105}
106
107; FUNC-LABEL: {{^}}v_uaddo_i64:
108; SI: v_add_i32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
109; SI: v_addc_u32_e32 v{{[0-9]+}}, vcc,
110
111; VI: v_add_u32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
112; VI: v_addc_u32_e32 v{{[0-9]+}}, vcc,
113
114; GFX9: v_add_co_u32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
115; GFX9: v_addc_co_u32_e32 v{{[0-9]+}}, vcc,
116
117; EG: ADDC_UINT
118; EG: ADD_INT
119define amdgpu_kernel void @v_uaddo_i64(i64 addrspace(1)* %out, i1 addrspace(1)* %carryout, i64 addrspace(1)* %a.ptr, i64 addrspace(1)* %b.ptr) #0 {
120  %tid = call i32 @llvm.amdgcn.workitem.id.x()
121  %tid.ext = sext i32 %tid to i64
122  %a.gep = getelementptr inbounds i64, i64 addrspace(1)* %a.ptr
123  %b.gep = getelementptr inbounds i64, i64 addrspace(1)* %b.ptr
124  %a = load i64, i64 addrspace(1)* %a.gep
125  %b = load i64, i64 addrspace(1)* %b.gep
126  %uadd = call { i64, i1 } @llvm.uadd.with.overflow.i64(i64 %a, i64 %b)
127  %val = extractvalue { i64, i1 } %uadd, 0
128  %carry = extractvalue { i64, i1 } %uadd, 1
129  store i64 %val, i64 addrspace(1)* %out
130  store i1 %carry, i1 addrspace(1)* %carryout
131  ret void
132}
133
134; FUNC-LABEL: {{^}}v_uaddo_i16:
135; VI: v_add_u16_e32
136; VI: v_cmp_lt_u16_e32
137
138; GFX9: v_add_u16_e32
139; GFX9: v_cmp_lt_u16_e32
140define amdgpu_kernel void @v_uaddo_i16(i16 addrspace(1)* %out, i1 addrspace(1)* %carryout, i16 addrspace(1)* %a.ptr, i16 addrspace(1)* %b.ptr) #0 {
141  %tid = call i32 @llvm.amdgcn.workitem.id.x()
142  %tid.ext = sext i32 %tid to i64
143  %a.gep = getelementptr inbounds i16, i16 addrspace(1)* %a.ptr
144  %b.gep = getelementptr inbounds i16, i16 addrspace(1)* %b.ptr
145  %a = load i16, i16 addrspace(1)* %a.gep
146  %b = load i16, i16 addrspace(1)* %b.gep
147  %uadd = call { i16, i1 } @llvm.uadd.with.overflow.i16(i16 %a, i16 %b)
148  %val = extractvalue { i16, i1 } %uadd, 0
149  %carry = extractvalue { i16, i1 } %uadd, 1
150  store i16 %val, i16 addrspace(1)* %out
151  store i1 %carry, i1 addrspace(1)* %carryout
152  ret void
153}
154
155declare i32 @llvm.amdgcn.workitem.id.x() #1
156declare { i16, i1 } @llvm.uadd.with.overflow.i16(i16, i16) #1
157declare { i32, i1 } @llvm.uadd.with.overflow.i32(i32, i32) #1
158declare { i64, i1 } @llvm.uadd.with.overflow.i64(i64, i64) #1
159
160attributes #0 = { nounwind }
161attributes #1 = { nounwind readnone }
162