Mercurial > hg > CbC > CbC_llvm
comparison test/CodeGen/AMDGPU/operand-folding.ll @ 95:afa8332a0e37 LLVM3.8
LLVM 3.8
author | Kaito Tokumori <e105711@ie.u-ryukyu.ac.jp> |
---|---|
date | Tue, 13 Oct 2015 17:48:58 +0900 |
parents | |
children | 1172e4bd9c6f |
comparison
equal
deleted
inserted
replaced
84:f3e34b893a5f | 95:afa8332a0e37 |
---|---|
1 ; RUN: llc < %s -march=amdgcn -mcpu=SI -verify-machineinstrs | FileCheck %s | |
2 | |
3 ; CHECK-LABEL: {{^}}fold_sgpr: | |
4 ; CHECK: v_add_i32_e32 v{{[0-9]+}}, vcc, s | |
5 define void @fold_sgpr(i32 addrspace(1)* %out, i32 %fold) { | |
6 entry: | |
7 %tmp0 = icmp ne i32 %fold, 0 | |
8 br i1 %tmp0, label %if, label %endif | |
9 | |
10 if: | |
11 %id = call i32 @llvm.r600.read.tidig.x() | |
12 %offset = add i32 %fold, %id | |
13 %tmp1 = getelementptr i32, i32 addrspace(1)* %out, i32 %offset | |
14 store i32 0, i32 addrspace(1)* %tmp1 | |
15 br label %endif | |
16 | |
17 endif: | |
18 ret void | |
19 } | |
20 | |
21 ; CHECK-LABEL: {{^}}fold_imm: | |
22 ; CHECK: v_or_b32_e32 v{{[0-9]+}}, 5 | |
23 define void @fold_imm(i32 addrspace(1)* %out, i32 %cmp) { | |
24 entry: | |
25 %fold = add i32 3, 2 | |
26 %tmp0 = icmp ne i32 %cmp, 0 | |
27 br i1 %tmp0, label %if, label %endif | |
28 | |
29 if: | |
30 %id = call i32 @llvm.r600.read.tidig.x() | |
31 %val = or i32 %id, %fold | |
32 store i32 %val, i32 addrspace(1)* %out | |
33 br label %endif | |
34 | |
35 endif: | |
36 ret void | |
37 } | |
38 | |
39 ; CHECK-LABEL: {{^}}fold_64bit_constant_add: | |
40 ; CHECK-NOT: s_mov_b64 | |
41 ; FIXME: It would be better if we could use v_add here and drop the extra | |
42 ; v_mov_b32 instructions. | |
43 ; CHECK-DAG: s_add_u32 [[LO:s[0-9]+]], s{{[0-9]+}}, 1 | |
44 ; CHECK-DAG: s_addc_u32 [[HI:s[0-9]+]], s{{[0-9]+}}, 0 | |
45 ; CHECK-DAG: v_mov_b32_e32 v[[VLO:[0-9]+]], [[LO]] | |
46 ; CHECK-DAG: v_mov_b32_e32 v[[VHI:[0-9]+]], [[HI]] | |
47 ; CHECK: buffer_store_dwordx2 v{{\[}}[[VLO]]:[[VHI]]{{\]}}, | |
48 | |
49 define void @fold_64bit_constant_add(i64 addrspace(1)* %out, i32 %cmp, i64 %val) { | |
50 entry: | |
51 %tmp0 = add i64 %val, 1 | |
52 store i64 %tmp0, i64 addrspace(1)* %out | |
53 ret void | |
54 } | |
55 | |
56 ; Inline constants should always be folded. | |
57 | |
58 ; CHECK-LABEL: {{^}}vector_inline: | |
59 ; CHECK: v_xor_b32_e32 v{{[0-9]+}}, 5, v{{[0-9]+}} | |
60 ; CHECK: v_xor_b32_e32 v{{[0-9]+}}, 5, v{{[0-9]+}} | |
61 ; CHECK: v_xor_b32_e32 v{{[0-9]+}}, 5, v{{[0-9]+}} | |
62 ; CHECK: v_xor_b32_e32 v{{[0-9]+}}, 5, v{{[0-9]+}} | |
63 | |
64 define void @vector_inline(<4 x i32> addrspace(1)* %out) { | |
65 entry: | |
66 %tmp0 = call i32 @llvm.r600.read.tidig.x() | |
67 %tmp1 = add i32 %tmp0, 1 | |
68 %tmp2 = add i32 %tmp0, 2 | |
69 %tmp3 = add i32 %tmp0, 3 | |
70 %vec0 = insertelement <4 x i32> undef, i32 %tmp0, i32 0 | |
71 %vec1 = insertelement <4 x i32> %vec0, i32 %tmp1, i32 1 | |
72 %vec2 = insertelement <4 x i32> %vec1, i32 %tmp2, i32 2 | |
73 %vec3 = insertelement <4 x i32> %vec2, i32 %tmp3, i32 3 | |
74 %tmp4 = xor <4 x i32> <i32 5, i32 5, i32 5, i32 5>, %vec3 | |
75 store <4 x i32> %tmp4, <4 x i32> addrspace(1)* %out | |
76 ret void | |
77 } | |
78 | |
79 ; Immediates with one use should be folded | |
80 ; CHECK-LABEL: {{^}}imm_one_use: | |
81 ; CHECK: v_xor_b32_e32 v{{[0-9]+}}, 0x64, v{{[0-9]+}} | |
82 | |
83 define void @imm_one_use(i32 addrspace(1)* %out) { | |
84 entry: | |
85 %tmp0 = call i32 @llvm.r600.read.tidig.x() | |
86 %tmp1 = xor i32 %tmp0, 100 | |
87 store i32 %tmp1, i32 addrspace(1)* %out | |
88 ret void | |
89 } | |
90 ; CHECK-LABEL: {{^}}vector_imm: | |
91 ; CHECK: s_movk_i32 [[IMM:s[0-9]+]], 0x64 | |
92 ; CHECK: v_xor_b32_e32 v{{[0-9]}}, [[IMM]], v{{[0-9]}} | |
93 ; CHECK: v_xor_b32_e32 v{{[0-9]}}, [[IMM]], v{{[0-9]}} | |
94 ; CHECK: v_xor_b32_e32 v{{[0-9]}}, [[IMM]], v{{[0-9]}} | |
95 ; CHECK: v_xor_b32_e32 v{{[0-9]}}, [[IMM]], v{{[0-9]}} | |
96 | |
97 define void @vector_imm(<4 x i32> addrspace(1)* %out) { | |
98 entry: | |
99 %tmp0 = call i32 @llvm.r600.read.tidig.x() | |
100 %tmp1 = add i32 %tmp0, 1 | |
101 %tmp2 = add i32 %tmp0, 2 | |
102 %tmp3 = add i32 %tmp0, 3 | |
103 %vec0 = insertelement <4 x i32> undef, i32 %tmp0, i32 0 | |
104 %vec1 = insertelement <4 x i32> %vec0, i32 %tmp1, i32 1 | |
105 %vec2 = insertelement <4 x i32> %vec1, i32 %tmp2, i32 2 | |
106 %vec3 = insertelement <4 x i32> %vec2, i32 %tmp3, i32 3 | |
107 %tmp4 = xor <4 x i32> <i32 100, i32 100, i32 100, i32 100>, %vec3 | |
108 store <4 x i32> %tmp4, <4 x i32> addrspace(1)* %out | |
109 ret void | |
110 } | |
111 | |
112 declare i32 @llvm.r600.read.tidig.x() #0 | |
113 attributes #0 = { readnone } |