150
|
1 ; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=tahiti -verify-machineinstrs < %s | FileCheck -check-prefix=GCN -check-prefix=GCN-NOHSA -check-prefix=SI %s
|
|
2 ; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=bonaire -verify-machineinstrs < %s | FileCheck -check-prefix=GCN -check-prefix=GCN-NOHSA -check-prefix=CI -check-prefix=CI-NOHSA %s
|
|
3 ; RUN: llc -amdgpu-scalarize-global-loads=false -mtriple=amdgcn--amdhsa -mcpu=bonaire -verify-machineinstrs < %s | FileCheck -check-prefix=GCN -check-prefix=CI --check-prefix=GCN-HSA %s
|
|
4
|
|
5 declare i32 @llvm.amdgcn.workitem.id.x() #0
|
|
6 declare i32 @llvm.amdgcn.workitem.id.y() #0
|
|
7
|
|
8 ; In this test both the pointer and the offset operands to the
|
|
9 ; BUFFER_LOAD instructions end up being stored in vgprs. This
|
|
10 ; requires us to add the pointer and offset together, store the
|
|
11 ; result in the offset operand (vaddr), and then store 0 in an
|
|
12 ; sgpr register pair and use that for the pointer operand
|
|
13 ; (low 64-bits of srsrc).
|
|
14
|
|
15 ; GCN-LABEL: {{^}}mubuf:
|
|
16
|
|
17 ; Make sure we aren't using VGPRs for the source operand of s_mov_b64
|
|
18 ; GCN-NOT: s_mov_b64 s[{{[0-9]+:[0-9]+}}], v
|
|
19
|
|
20 ; Make sure we aren't using VGPR's for the srsrc operand of BUFFER_LOAD_*
|
|
21 ; instructions
|
|
22 ; GCN-NOHSA: buffer_load_ubyte v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}], s[{{[0-9]+:[0-9]+}}], 0 addr64
|
|
23 ; GCN-NOHSA: buffer_load_ubyte v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}], s[{{[0-9]+:[0-9]+}}], 0 addr64
|
|
24 ; GCN-HSA: flat_load_ubyte v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}
|
|
25 ; GCN-HSA: flat_load_ubyte v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}
|
|
26
|
|
27 define amdgpu_kernel void @mubuf(i32 addrspace(1)* %out, i8 addrspace(1)* %in) #1 {
|
|
28 entry:
|
|
29 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
|
|
30 %tmp1 = call i32 @llvm.amdgcn.workitem.id.y()
|
|
31 %tmp2 = sext i32 %tmp to i64
|
|
32 %tmp3 = sext i32 %tmp1 to i64
|
|
33 br label %loop
|
|
34
|
|
35 loop: ; preds = %loop, %entry
|
|
36 %tmp4 = phi i64 [ 0, %entry ], [ %tmp5, %loop ]
|
|
37 %tmp5 = add i64 %tmp2, %tmp4
|
|
38 %tmp6 = getelementptr i8, i8 addrspace(1)* %in, i64 %tmp5
|
|
39 %tmp7 = load i8, i8 addrspace(1)* %tmp6, align 1
|
|
40 %tmp8 = or i64 %tmp5, 1
|
|
41 %tmp9 = getelementptr i8, i8 addrspace(1)* %in, i64 %tmp8
|
|
42 %tmp10 = load i8, i8 addrspace(1)* %tmp9, align 1
|
|
43 %tmp11 = add i8 %tmp7, %tmp10
|
|
44 %tmp12 = sext i8 %tmp11 to i32
|
|
45 store i32 %tmp12, i32 addrspace(1)* %out
|
|
46 %tmp13 = icmp slt i64 %tmp5, 10
|
|
47 br i1 %tmp13, label %loop, label %done
|
|
48
|
|
49 done: ; preds = %loop
|
|
50 ret void
|
|
51 }
|
|
52
|
|
53 ; Test moving an SMRD instruction to the VALU
|
|
54 ; FIXME: movs can be moved before nop to reduce count
|
|
55
|
|
56 ; GCN-LABEL: {{^}}smrd_valu:
|
|
57 ; SI: s_movk_i32 [[OFFSET:s[0-9]+]], 0x2ee0
|
|
58 ; GCN: v_readfirstlane_b32 s[[PTR_LO:[0-9]+]], v{{[0-9]+}}
|
|
59 ; GCN: v_readfirstlane_b32 s[[PTR_HI:[0-9]+]], v{{[0-9]+}}
|
173
|
60 ; SI: s_mov_b32
|
|
61 ; SI: s_nop 1
|
150
|
62 ; SI: s_load_dword [[OUT:s[0-9]+]], s{{\[}}[[PTR_LO]]:[[PTR_HI]]{{\]}}, [[OFFSET]]
|
|
63
|
|
64 ; CI: s_load_dword [[OUT:s[0-9]+]], s{{\[}}[[PTR_LO]]:[[PTR_HI]]{{\]}}, 0xbb8
|
|
65 ; GCN: v_mov_b32_e32 [[V_OUT:v[0-9]+]], [[OUT]]
|
|
66 ; GCN-NOHSA: buffer_store_dword [[V_OUT]]
|
|
67 ; GCN-HSA: flat_store_dword {{.*}}, [[V_OUT]]
|
|
68 define amdgpu_kernel void @smrd_valu(i32 addrspace(4)* addrspace(1)* %in, i32 %a, i32 %b, i32 addrspace(1)* %out) #1 {
|
|
69 entry:
|
|
70 %tmp = icmp ne i32 %a, 0
|
|
71 br i1 %tmp, label %if, label %else
|
|
72
|
|
73 if: ; preds = %entry
|
|
74 %tmp1 = load i32 addrspace(4)*, i32 addrspace(4)* addrspace(1)* %in
|
|
75 br label %endif
|
|
76
|
|
77 else: ; preds = %entry
|
|
78 %tmp2 = getelementptr i32 addrspace(4)*, i32 addrspace(4)* addrspace(1)* %in
|
|
79 %tmp3 = load i32 addrspace(4)*, i32 addrspace(4)* addrspace(1)* %tmp2
|
|
80 br label %endif
|
|
81
|
|
82 endif: ; preds = %else, %if
|
|
83 %tmp4 = phi i32 addrspace(4)* [ %tmp1, %if ], [ %tmp3, %else ]
|
|
84 %tmp5 = getelementptr i32, i32 addrspace(4)* %tmp4, i32 3000
|
|
85 %tmp6 = load i32, i32 addrspace(4)* %tmp5
|
|
86 store i32 %tmp6, i32 addrspace(1)* %out
|
|
87 ret void
|
|
88 }
|
|
89
|
|
90 ; Test moving an SMRD with an immediate offset to the VALU
|
|
91
|
|
92 ; GCN-LABEL: {{^}}smrd_valu2:
|
|
93 ; GCN-NOHSA-NOT: v_add
|
|
94 ; GCN-NOHSA: buffer_load_dword v{{[0-9]+}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], 0 addr64 offset:16{{$}}
|
|
95 ; GCN-HSA: flat_load_dword v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}]
|
|
96 define amdgpu_kernel void @smrd_valu2(i32 addrspace(1)* %out, [8 x i32] addrspace(4)* %in) #1 {
|
|
97 entry:
|
|
98 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
|
|
99 %tmp1 = add i32 %tmp, 4
|
|
100 %tmp2 = getelementptr [8 x i32], [8 x i32] addrspace(4)* %in, i32 %tmp, i32 4
|
|
101 %tmp3 = load i32, i32 addrspace(4)* %tmp2
|
|
102 store i32 %tmp3, i32 addrspace(1)* %out
|
|
103 ret void
|
|
104 }
|
|
105
|
|
106 ; Use a big offset that will use the SMRD literal offset on CI
|
|
107 ; GCN-LABEL: {{^}}smrd_valu_ci_offset:
|
|
108 ; GCN-NOHSA-NOT: v_add
|
|
109 ; GCN-NOHSA: s_movk_i32 [[OFFSET:s[0-9]+]], 0x4e20{{$}}
|
|
110 ; GCN-NOHSA-NOT: v_add
|
|
111 ; GCN-NOHSA: buffer_load_dword v{{[0-9]+}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET]] addr64{{$}}
|
|
112 ; GCN-NOHSA: v_add_i32_e32
|
|
113 ; GCN-NOHSA: buffer_store_dword
|
|
114 ; GCN-HSA: flat_load_dword v{{[0-9]+}}, v[{{[0-9]+:[0-9]+}}]
|
|
115 ; GCN-HSA: flat_store_dword v[{{[0-9]+:[0-9]+}}], v{{[0-9]+}}
|
|
116 define amdgpu_kernel void @smrd_valu_ci_offset(i32 addrspace(1)* %out, i32 addrspace(4)* %in, i32 %c) #1 {
|
|
117 entry:
|
|
118 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
|
|
119 %tmp2 = getelementptr i32, i32 addrspace(4)* %in, i32 %tmp
|
|
120 %tmp3 = getelementptr i32, i32 addrspace(4)* %tmp2, i32 5000
|
|
121 %tmp4 = load i32, i32 addrspace(4)* %tmp3
|
|
122 %tmp5 = add i32 %tmp4, %c
|
|
123 store i32 %tmp5, i32 addrspace(1)* %out
|
|
124 ret void
|
|
125 }
|
|
126
|
|
127 ; GCN-LABEL: {{^}}smrd_valu_ci_offset_x2:
|
|
128 ; GCN-NOHSA-NOT: v_add
|
|
129 ; GCN-NOHSA: s_mov_b32 [[OFFSET:s[0-9]+]], 0x9c40{{$}}
|
|
130 ; GCN-NOHSA-NOT: v_add
|
|
131 ; GCN-NOHSA: buffer_load_dwordx2 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET]] addr64{{$}}
|
|
132 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
133 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
134 ; GCN-NOHSA: buffer_store_dwordx2
|
|
135 ; GCN-HSA: flat_load_dwordx2 v[{{[0-9]+:[0-9]+}}], v[{{[0-9]+:[0-9]+}}]
|
|
136 define amdgpu_kernel void @smrd_valu_ci_offset_x2(i64 addrspace(1)* %out, i64 addrspace(4)* %in, i64 %c) #1 {
|
|
137 entry:
|
|
138 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
|
|
139 %tmp2 = getelementptr i64, i64 addrspace(4)* %in, i32 %tmp
|
|
140 %tmp3 = getelementptr i64, i64 addrspace(4)* %tmp2, i32 5000
|
|
141 %tmp4 = load i64, i64 addrspace(4)* %tmp3
|
|
142 %tmp5 = or i64 %tmp4, %c
|
|
143 store i64 %tmp5, i64 addrspace(1)* %out
|
|
144 ret void
|
|
145 }
|
|
146
|
|
147 ; GCN-LABEL: {{^}}smrd_valu_ci_offset_x4:
|
|
148 ; GCN-NOHSA-NOT: v_add
|
|
149 ; GCN-NOHSA: s_movk_i32 [[OFFSET:s[0-9]+]], 0x4d20{{$}}
|
|
150 ; GCN-NOHSA-NOT: v_add
|
|
151 ; GCN-NOHSA: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET]] addr64{{$}}
|
|
152 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
153 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
154 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
155 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
156 ; GCN-NOHSA: buffer_store_dwordx4
|
|
157 ; GCN-HSA: flat_load_dwordx4 v[{{[0-9]+:[0-9]+}}], v[{{[0-9]+:[0-9]+}}]
|
|
158 define amdgpu_kernel void @smrd_valu_ci_offset_x4(<4 x i32> addrspace(1)* %out, <4 x i32> addrspace(4)* %in, <4 x i32> %c) #1 {
|
|
159 entry:
|
|
160 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
|
|
161 %tmp2 = getelementptr <4 x i32>, <4 x i32> addrspace(4)* %in, i32 %tmp
|
|
162 %tmp3 = getelementptr <4 x i32>, <4 x i32> addrspace(4)* %tmp2, i32 1234
|
|
163 %tmp4 = load <4 x i32>, <4 x i32> addrspace(4)* %tmp3
|
|
164 %tmp5 = or <4 x i32> %tmp4, %c
|
|
165 store <4 x i32> %tmp5, <4 x i32> addrspace(1)* %out
|
|
166 ret void
|
|
167 }
|
|
168
|
|
169 ; Original scalar load uses SGPR offset on SI and 32-bit literal on
|
|
170 ; CI.
|
|
171
|
|
172 ; GCN-LABEL: {{^}}smrd_valu_ci_offset_x8:
|
|
173 ; GCN-NOHSA-DAG: s_mov_b32 [[OFFSET0:s[0-9]+]], 0x9a40{{$}}
|
|
174 ; CI-NOHSA-DAG: s_mov_b32 [[OFFSET1:s[0-9]+]], 0x9a50{{$}}
|
|
175 ; CI-NOHSA-NOT: v_add
|
|
176 ; SI: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], 0 addr64 offset:16
|
|
177 ; CI-NOHSA: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET1]] addr64{{$}}
|
|
178 ; GCN-NOHSA: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET0]] addr64{{$}}
|
|
179
|
|
180 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
181 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
182 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
183 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
184 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
185 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
186 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
187 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
188 ; GCN-NOHSA: buffer_store_dwordx4
|
|
189 ; GCN-NOHSA: buffer_store_dwordx4
|
|
190 ; GCN-HSA: flat_load_dwordx4
|
|
191 ; GCN-HSA: flat_load_dwordx4
|
|
192 define amdgpu_kernel void @smrd_valu_ci_offset_x8(<8 x i32> addrspace(1)* %out, <8 x i32> addrspace(4)* %in, <8 x i32> %c) #1 {
|
|
193 entry:
|
|
194 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
|
|
195 %tmp2 = getelementptr <8 x i32>, <8 x i32> addrspace(4)* %in, i32 %tmp
|
|
196 %tmp3 = getelementptr <8 x i32>, <8 x i32> addrspace(4)* %tmp2, i32 1234
|
|
197 %tmp4 = load <8 x i32>, <8 x i32> addrspace(4)* %tmp3
|
|
198 %tmp5 = or <8 x i32> %tmp4, %c
|
|
199 store <8 x i32> %tmp5, <8 x i32> addrspace(1)* %out
|
|
200 ret void
|
|
201 }
|
|
202
|
|
203 ; GCN-LABEL: {{^}}smrd_valu_ci_offset_x16:
|
|
204
|
|
205 ; SI: s_mov_b32 {{s[0-9]+}}, 0x13480
|
|
206 ; SI: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], 0 addr64 offset:32
|
|
207 ; SI: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], 0 addr64 offset:48
|
221
|
208 ; SI: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], 0 addr64 offset:16
|
150
|
209 ; SI: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], {{s[0-9]+}} addr64
|
|
210 ; CI-NOHSA-DAG: s_mov_b32 [[OFFSET0:s[0-9]+]], 0x13480{{$}}
|
|
211 ; CI-NOHSA-DAG: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET0]] addr64{{$}}
|
|
212 ; CI-NOHSA-DAG: s_mov_b32 [[OFFSET1:s[0-9]+]], 0x13490{{$}}
|
|
213 ; CI-NOHSA-DAG: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET1]] addr64{{$}}
|
|
214 ; CI-NOHSA-DAG: s_mov_b32 [[OFFSET2:s[0-9]+]], 0x134a0{{$}}
|
|
215 ; CI-NOHSA-DAG: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET2]] addr64{{$}}
|
|
216 ; CI-NOHSA-DAG: s_mov_b32 [[OFFSET3:s[0-9]+]], 0x134b0{{$}}
|
|
217 ; CI-NOHSA-DAG: buffer_load_dwordx4 v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, s[{{[0-9]+:[0-9]+}}], [[OFFSET3]] addr64{{$}}
|
|
218
|
|
219 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
220 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
221 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
222 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
223 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
224 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
225 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
226 ; GCN-NOHSA: v_or_b32_e32 {{v[0-9]+}}, {{s[0-9]+}}, {{v[0-9]+}}
|
|
227 ; GCN-NOHSA: buffer_store_dwordx4
|
|
228 ; GCN-NOHSA: buffer_store_dwordx4
|
|
229 ; GCN-NOHSA: buffer_store_dwordx4
|
|
230 ; GCN-NOHSA: buffer_store_dwordx4
|
|
231
|
|
232 ; GCN-HSA: flat_load_dwordx4
|
|
233 ; GCN-HSA: flat_load_dwordx4
|
|
234 ; GCN-HSA: flat_load_dwordx4
|
|
235 ; GCN-HSA: flat_load_dwordx4
|
|
236
|
|
237 ; GCN: s_endpgm
|
|
238 define amdgpu_kernel void @smrd_valu_ci_offset_x16(<16 x i32> addrspace(1)* %out, <16 x i32> addrspace(4)* %in, <16 x i32> %c) #1 {
|
|
239 entry:
|
|
240 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
|
|
241 %tmp2 = getelementptr <16 x i32>, <16 x i32> addrspace(4)* %in, i32 %tmp
|
|
242 %tmp3 = getelementptr <16 x i32>, <16 x i32> addrspace(4)* %tmp2, i32 1234
|
|
243 %tmp4 = load <16 x i32>, <16 x i32> addrspace(4)* %tmp3
|
|
244 %tmp5 = or <16 x i32> %tmp4, %c
|
|
245 store <16 x i32> %tmp5, <16 x i32> addrspace(1)* %out
|
|
246 ret void
|
|
247 }
|
|
248
|
|
249 ; GCN-LABEL: {{^}}smrd_valu2_salu_user:
|
|
250 ; GCN-NOHSA: buffer_load_dword [[MOVED:v[0-9]+]], v{{\[[0-9]+:[0-9]+\]}}, s{{\[[0-9]+:[0-9]+\]}}, 0 addr64 offset:16{{$}}
|
|
251 ; GCN-HSA: flat_load_dword [[MOVED:v[0-9]+]], v[{{[0-9+:[0-9]+}}]
|
|
252 ; GCN: v_add_i32_e32 [[ADD:v[0-9]+]], vcc, s{{[0-9]+}}, [[MOVED]]
|
|
253 ; GCN-NOHSA: buffer_store_dword [[ADD]]
|
|
254 ; GCN-HSA: flat_store_dword {{.*}}, [[ADD]]
|
|
255 define amdgpu_kernel void @smrd_valu2_salu_user(i32 addrspace(1)* %out, [8 x i32] addrspace(4)* %in, i32 %a) #1 {
|
|
256 entry:
|
|
257 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
|
|
258 %tmp1 = add i32 %tmp, 4
|
|
259 %tmp2 = getelementptr [8 x i32], [8 x i32] addrspace(4)* %in, i32 %tmp, i32 4
|
|
260 %tmp3 = load i32, i32 addrspace(4)* %tmp2
|
|
261 %tmp4 = add i32 %tmp3, %a
|
|
262 store i32 %tmp4, i32 addrspace(1)* %out
|
|
263 ret void
|
|
264 }
|
|
265
|
|
266 ; GCN-LABEL: {{^}}smrd_valu2_max_smrd_offset:
|
|
267 ; GCN-NOHSA: buffer_load_dword v{{[0-9]+}}, v{{\[[0-9]+:[0-9]+\]}}, s{{\[[0-9]+:[0-9]+\]}}, 0 addr64 offset:1020{{$}}
|
173
|
268 ; GCN-HSA: flat_load_dword v{{[0-9]}}, v[{{[0-9]+:[0-9]+}}]
|
150
|
269 define amdgpu_kernel void @smrd_valu2_max_smrd_offset(i32 addrspace(1)* %out, [1024 x i32] addrspace(4)* %in) #1 {
|
|
270 entry:
|
|
271 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
|
|
272 %tmp1 = add i32 %tmp, 4
|
|
273 %tmp2 = getelementptr [1024 x i32], [1024 x i32] addrspace(4)* %in, i32 %tmp, i32 255
|
|
274 %tmp3 = load i32, i32 addrspace(4)* %tmp2
|
|
275 store i32 %tmp3, i32 addrspace(1)* %out
|
|
276 ret void
|
|
277 }
|
|
278
|
|
279 ; GCN-LABEL: {{^}}smrd_valu2_mubuf_offset:
|
|
280 ; GCN-NOHSA-NOT: v_add
|
|
281 ; GCN-NOHSA: buffer_load_dword v{{[0-9]+}}, v{{\[[0-9]+:[0-9]+\]}}, s{{\[[0-9]+:[0-9]+\]}}, 0 addr64 offset:1024{{$}}
|
|
282 ; GCN-HSA: flat_load_dword v{{[0-9]}}, v[{{[0-9]+:[0-9]+}}]
|
|
283 define amdgpu_kernel void @smrd_valu2_mubuf_offset(i32 addrspace(1)* %out, [1024 x i32] addrspace(4)* %in) #1 {
|
|
284 entry:
|
|
285 %tmp = call i32 @llvm.amdgcn.workitem.id.x()
|
|
286 %tmp1 = add i32 %tmp, 4
|
|
287 %tmp2 = getelementptr [1024 x i32], [1024 x i32] addrspace(4)* %in, i32 %tmp, i32 256
|
|
288 %tmp3 = load i32, i32 addrspace(4)* %tmp2
|
|
289 store i32 %tmp3, i32 addrspace(1)* %out
|
|
290 ret void
|
|
291 }
|
|
292
|
|
293 ; GCN-LABEL: {{^}}s_load_imm_v8i32:
|
|
294 ; GCN-NOHSA: buffer_load_dwordx4
|
|
295 ; GCN-NOHSA: buffer_load_dwordx4
|
|
296 ; GCN-HSA: flat_load_dwordx4
|
|
297 ; GCN-HSA: flat_load_dwordx4
|
|
298 define amdgpu_kernel void @s_load_imm_v8i32(<8 x i32> addrspace(1)* %out, i32 addrspace(4)* nocapture readonly %in) #1 {
|
|
299 entry:
|
|
300 %tmp0 = tail call i32 @llvm.amdgcn.workitem.id.x()
|
|
301 %tmp1 = getelementptr inbounds i32, i32 addrspace(4)* %in, i32 %tmp0
|
|
302 %tmp2 = bitcast i32 addrspace(4)* %tmp1 to <8 x i32> addrspace(4)*
|
|
303 %tmp3 = load <8 x i32>, <8 x i32> addrspace(4)* %tmp2, align 4
|
|
304 store <8 x i32> %tmp3, <8 x i32> addrspace(1)* %out, align 32
|
|
305 ret void
|
|
306 }
|
|
307
|
|
308 ; GCN-LABEL: {{^}}s_load_imm_v8i32_salu_user:
|
|
309 ; GCN-NOHSA: buffer_load_dwordx4
|
|
310 ; GCN-NOHSA: buffer_load_dwordx4
|
|
311 ; GCN-NOHSA: v_add_i32_e32
|
|
312 ; GCN-NOHSA: v_add_i32_e32
|
|
313 ; GCN-NOHSA: v_add_i32_e32
|
|
314 ; GCN-NOHSA: v_add_i32_e32
|
|
315 ; GCN-NOHSA: v_add_i32_e32
|
|
316 ; GCN-NOHSA: v_add_i32_e32
|
|
317 ; GCN-NOHSA: v_add_i32_e32
|
|
318 ; GCN-NOHSA: buffer_store_dword
|
|
319 ; GCN-HSA: flat_load_dwordx4
|
|
320 ; GCN-HSA: flat_load_dwordx4
|
|
321 define amdgpu_kernel void @s_load_imm_v8i32_salu_user(i32 addrspace(1)* %out, i32 addrspace(4)* nocapture readonly %in) #1 {
|
|
322 entry:
|
|
323 %tmp0 = tail call i32 @llvm.amdgcn.workitem.id.x()
|
|
324 %tmp1 = getelementptr inbounds i32, i32 addrspace(4)* %in, i32 %tmp0
|
|
325 %tmp2 = bitcast i32 addrspace(4)* %tmp1 to <8 x i32> addrspace(4)*
|
|
326 %tmp3 = load <8 x i32>, <8 x i32> addrspace(4)* %tmp2, align 4
|
|
327
|
|
328 %elt0 = extractelement <8 x i32> %tmp3, i32 0
|
|
329 %elt1 = extractelement <8 x i32> %tmp3, i32 1
|
|
330 %elt2 = extractelement <8 x i32> %tmp3, i32 2
|
|
331 %elt3 = extractelement <8 x i32> %tmp3, i32 3
|
|
332 %elt4 = extractelement <8 x i32> %tmp3, i32 4
|
|
333 %elt5 = extractelement <8 x i32> %tmp3, i32 5
|
|
334 %elt6 = extractelement <8 x i32> %tmp3, i32 6
|
|
335 %elt7 = extractelement <8 x i32> %tmp3, i32 7
|
|
336
|
|
337 %add0 = add i32 %elt0, %elt1
|
|
338 %add1 = add i32 %add0, %elt2
|
|
339 %add2 = add i32 %add1, %elt3
|
|
340 %add3 = add i32 %add2, %elt4
|
|
341 %add4 = add i32 %add3, %elt5
|
|
342 %add5 = add i32 %add4, %elt6
|
|
343 %add6 = add i32 %add5, %elt7
|
|
344
|
|
345 store i32 %add6, i32 addrspace(1)* %out
|
|
346 ret void
|
|
347 }
|
|
348
|
|
349 ; GCN-LABEL: {{^}}s_load_imm_v16i32:
|
|
350 ; GCN-NOHSA: buffer_load_dwordx4
|
|
351 ; GCN-NOHSA: buffer_load_dwordx4
|
|
352 ; GCN-NOHSA: buffer_load_dwordx4
|
|
353 ; GCN-NOHSA: buffer_load_dwordx4
|
|
354 ; GCN-HSA: flat_load_dwordx4
|
|
355 ; GCN-HSA: flat_load_dwordx4
|
|
356 ; GCN-HSA: flat_load_dwordx4
|
|
357 ; GCN-HSA: flat_load_dwordx4
|
|
358 define amdgpu_kernel void @s_load_imm_v16i32(<16 x i32> addrspace(1)* %out, i32 addrspace(4)* nocapture readonly %in) #1 {
|
|
359 entry:
|
|
360 %tmp0 = tail call i32 @llvm.amdgcn.workitem.id.x()
|
|
361 %tmp1 = getelementptr inbounds i32, i32 addrspace(4)* %in, i32 %tmp0
|
|
362 %tmp2 = bitcast i32 addrspace(4)* %tmp1 to <16 x i32> addrspace(4)*
|
|
363 %tmp3 = load <16 x i32>, <16 x i32> addrspace(4)* %tmp2, align 4
|
|
364 store <16 x i32> %tmp3, <16 x i32> addrspace(1)* %out, align 32
|
|
365 ret void
|
|
366 }
|
|
367
|
|
368 ; GCN-LABEL: {{^}}s_load_imm_v16i32_salu_user:
|
|
369 ; GCN-NOHSA: buffer_load_dwordx4
|
|
370 ; GCN-NOHSA: buffer_load_dwordx4
|
|
371 ; GCN-NOHSA: buffer_load_dwordx4
|
|
372 ; GCN-NOHSA: buffer_load_dwordx4
|
|
373 ; GCN-NOHSA: v_add_i32_e32
|
|
374 ; GCN-NOHSA: v_add_i32_e32
|
|
375 ; GCN-NOHSA: v_add_i32_e32
|
|
376 ; GCN-NOHSA: v_add_i32_e32
|
|
377 ; GCN-NOHSA: v_add_i32_e32
|
|
378 ; GCN-NOHSA: v_add_i32_e32
|
|
379 ; GCN-NOHSA: v_add_i32_e32
|
|
380 ; GCN-NOHSA: v_add_i32_e32
|
|
381 ; GCN-NOHSA: v_add_i32_e32
|
|
382 ; GCN-NOHSA: v_add_i32_e32
|
|
383 ; GCN-NOHSA: v_add_i32_e32
|
|
384 ; GCN-NOHSA: v_add_i32_e32
|
|
385 ; GCN-NOHSA: v_add_i32_e32
|
|
386 ; GCN-NOHSA: v_add_i32_e32
|
|
387 ; GCN-NOHSA: v_add_i32_e32
|
|
388 ; GCN-NOHSA: buffer_store_dword
|
|
389 ; GCN-HSA: flat_load_dwordx4
|
|
390 ; GCN-HSA: flat_load_dwordx4
|
|
391 ; GCN-HSA: flat_load_dwordx4
|
|
392 ; GCN-HSA: flat_load_dwordx4
|
|
393 define amdgpu_kernel void @s_load_imm_v16i32_salu_user(i32 addrspace(1)* %out, i32 addrspace(4)* nocapture readonly %in) #1 {
|
|
394 entry:
|
|
395 %tmp0 = tail call i32 @llvm.amdgcn.workitem.id.x()
|
|
396 %tmp1 = getelementptr inbounds i32, i32 addrspace(4)* %in, i32 %tmp0
|
|
397 %tmp2 = bitcast i32 addrspace(4)* %tmp1 to <16 x i32> addrspace(4)*
|
|
398 %tmp3 = load <16 x i32>, <16 x i32> addrspace(4)* %tmp2, align 4
|
|
399
|
|
400 %elt0 = extractelement <16 x i32> %tmp3, i32 0
|
|
401 %elt1 = extractelement <16 x i32> %tmp3, i32 1
|
|
402 %elt2 = extractelement <16 x i32> %tmp3, i32 2
|
|
403 %elt3 = extractelement <16 x i32> %tmp3, i32 3
|
|
404 %elt4 = extractelement <16 x i32> %tmp3, i32 4
|
|
405 %elt5 = extractelement <16 x i32> %tmp3, i32 5
|
|
406 %elt6 = extractelement <16 x i32> %tmp3, i32 6
|
|
407 %elt7 = extractelement <16 x i32> %tmp3, i32 7
|
|
408 %elt8 = extractelement <16 x i32> %tmp3, i32 8
|
|
409 %elt9 = extractelement <16 x i32> %tmp3, i32 9
|
|
410 %elt10 = extractelement <16 x i32> %tmp3, i32 10
|
|
411 %elt11 = extractelement <16 x i32> %tmp3, i32 11
|
|
412 %elt12 = extractelement <16 x i32> %tmp3, i32 12
|
|
413 %elt13 = extractelement <16 x i32> %tmp3, i32 13
|
|
414 %elt14 = extractelement <16 x i32> %tmp3, i32 14
|
|
415 %elt15 = extractelement <16 x i32> %tmp3, i32 15
|
|
416
|
|
417 %add0 = add i32 %elt0, %elt1
|
|
418 %add1 = add i32 %add0, %elt2
|
|
419 %add2 = add i32 %add1, %elt3
|
|
420 %add3 = add i32 %add2, %elt4
|
|
421 %add4 = add i32 %add3, %elt5
|
|
422 %add5 = add i32 %add4, %elt6
|
|
423 %add6 = add i32 %add5, %elt7
|
|
424 %add7 = add i32 %add6, %elt8
|
|
425 %add8 = add i32 %add7, %elt9
|
|
426 %add9 = add i32 %add8, %elt10
|
|
427 %add10 = add i32 %add9, %elt11
|
|
428 %add11 = add i32 %add10, %elt12
|
|
429 %add12 = add i32 %add11, %elt13
|
|
430 %add13 = add i32 %add12, %elt14
|
|
431 %add14 = add i32 %add13, %elt15
|
|
432
|
|
433 store i32 %add14, i32 addrspace(1)* %out
|
|
434 ret void
|
|
435 }
|
|
436
|
|
437 ; Make sure we legalize vopc operands after moving an sopc to the value.
|
|
438
|
|
439 ; {{^}}sopc_vopc_legalize_bug:
|
|
440 ; GCN: s_load_dword [[SGPR:s[0-9]+]]
|
|
441 ; GCN: v_cmp_le_u32_e32 vcc, [[SGPR]], v{{[0-9]+}}
|
|
442 ; GCN: s_and_b64 vcc, exec, vcc
|
|
443 ; GCN: s_cbranch_vccnz [[EXIT:[A-Z0-9_]+]]
|
|
444 ; GCN: v_mov_b32_e32 [[ONE:v[0-9]+]], 1
|
|
445 ; GCN-NOHSA: buffer_store_dword [[ONE]]
|
|
446 ; GCN-HSA: flat_store_dword v[{{[0-9]+:[0-9]+}}], [[ONE]]
|
|
447 ; GCN: {{^}}[[EXIT]]:
|
|
448 ; GCN: s_endpgm
|
|
449 define amdgpu_kernel void @sopc_vopc_legalize_bug(i32 %cond, i32 addrspace(1)* %out, i32 addrspace(1)* %in) {
|
|
450 bb3: ; preds = %bb2
|
|
451 %tmp0 = bitcast i32 %cond to float
|
|
452 %tmp1 = fadd float %tmp0, 2.500000e-01
|
|
453 %tmp2 = bitcast float %tmp1 to i32
|
|
454 %tmp3 = icmp ult i32 %tmp2, %cond
|
|
455 br i1 %tmp3, label %bb6, label %bb7
|
|
456
|
|
457 bb6:
|
|
458 store i32 1, i32 addrspace(1)* %out
|
|
459 br label %bb7
|
|
460
|
|
461 bb7: ; preds = %bb3
|
|
462 ret void
|
|
463 }
|
|
464
|
|
465 ; GCN-LABEL: {{^}}phi_visit_order:
|
|
466 ; GCN: v_add_i32_e64 v{{[0-9]+}}, s{{\[[0-9]+:[0-9]+\]}}, 1, v{{[0-9]+}}
|
|
467 define amdgpu_kernel void @phi_visit_order() {
|
|
468 bb:
|
|
469 br label %bb1
|
|
470
|
|
471 bb1:
|
|
472 %tmp = phi i32 [ 0, %bb ], [ %tmp5, %bb4 ]
|
|
473 %tid = call i32 @llvm.amdgcn.workitem.id.x()
|
|
474 %cnd = icmp eq i32 %tid, 0
|
|
475 br i1 %cnd, label %bb4, label %bb2
|
|
476
|
|
477 bb2:
|
|
478 %tmp3 = add nsw i32 %tmp, 1
|
|
479 br label %bb4
|
|
480
|
|
481 bb4:
|
|
482 %tmp5 = phi i32 [ %tmp3, %bb2 ], [ %tmp, %bb1 ]
|
|
483 store volatile i32 %tmp5, i32 addrspace(1)* undef
|
|
484 br label %bb1
|
|
485 }
|
|
486
|
|
487 ; GCN-LABEL: {{^}}phi_imm_in_sgprs
|
|
488 ; GCN: s_movk_i32 [[A:s[0-9]+]], 0x400
|
|
489 ; GCN: s_movk_i32 [[B:s[0-9]+]], 0x400
|
|
490 ; GCN: [[LOOP_LABEL:[0-9a-zA-Z_]+]]:
|
|
491 ; GCN: s_xor_b32 [[B]], [[B]], [[A]]
|
|
492 ; GCN: s_cbranch_scc{{[01]}} [[LOOP_LABEL]]
|
|
493 define amdgpu_kernel void @phi_imm_in_sgprs(i32 addrspace(3)* %out, i32 %cond) {
|
|
494 entry:
|
|
495 br label %loop
|
|
496
|
|
497 loop:
|
|
498 %i = phi i32 [0, %entry], [%i.add, %loop]
|
|
499 %offset = phi i32 [1024, %entry], [%offset.xor, %loop]
|
|
500 %offset.xor = xor i32 %offset, 1024
|
|
501 %offset.i = add i32 %offset.xor, %i
|
|
502 %ptr = getelementptr i32, i32 addrspace(3)* %out, i32 %offset.i
|
|
503 store i32 0, i32 addrspace(3)* %ptr
|
|
504 %i.add = add i32 %i, 1
|
|
505 %cmp = icmp ult i32 %i.add, %cond
|
|
506 br i1 %cmp, label %loop, label %exit
|
|
507
|
|
508 exit:
|
|
509 ret void
|
|
510 }
|
|
511
|
|
512 attributes #0 = { nounwind readnone }
|
|
513 attributes #1 = { nounwind }
|