Mercurial > hg > CbC > CbC_llvm
diff llvm/test/CodeGen/AMDGPU/xnor.ll @ 150:1d019706d866
LLVM10
author | anatofuz |
---|---|
date | Thu, 13 Feb 2020 15:10:13 +0900 |
parents | |
children | 2e18cbf3894f |
line wrap: on
line diff
--- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/llvm/test/CodeGen/AMDGPU/xnor.ll Thu Feb 13 15:10:13 2020 +0900 @@ -0,0 +1,198 @@ +; RUN: llc -march=amdgcn -mcpu=gfx600 -verify-machineinstrs < %s | FileCheck --check-prefix=GCN --check-prefix=GFX600 %s +; RUN: llc -march=amdgcn -mcpu=gfx700 -verify-machineinstrs < %s | FileCheck --check-prefix=GCN --check-prefix=GFX700 %s +; RUN: llc -march=amdgcn -mcpu=gfx801 -verify-machineinstrs < %s | FileCheck --check-prefix=GCN --check-prefix=GFX801 %s +; RUN: llc -march=amdgcn -mcpu=gfx900 -verify-machineinstrs < %s | FileCheck --check-prefix=GCN --check-prefix=GFX900 %s +; RUN: llc -march=amdgcn -mcpu=gfx906 -verify-machineinstrs < %s | FileCheck --check-prefix=GCN-DL --check-prefix=GFX906 %s + +; GCN-LABEL: {{^}}scalar_xnor_i32_one_use +; GCN: s_xnor_b32 +define amdgpu_kernel void @scalar_xnor_i32_one_use( + i32 addrspace(1)* %r0, i32 %a, i32 %b) { +entry: + %xor = xor i32 %a, %b + %r0.val = xor i32 %xor, -1 + store i32 %r0.val, i32 addrspace(1)* %r0 + ret void +} + +; GCN-LABEL: {{^}}scalar_xnor_i32_mul_use +; GCN-NOT: s_xnor_b32 +; GCN: s_xor_b32 +; GCN: s_not_b32 +; GCN: s_add_i32 +define amdgpu_kernel void @scalar_xnor_i32_mul_use( + i32 addrspace(1)* %r0, i32 addrspace(1)* %r1, i32 %a, i32 %b) { +entry: + %xor = xor i32 %a, %b + %r0.val = xor i32 %xor, -1 + %r1.val = add i32 %xor, %a + store i32 %r0.val, i32 addrspace(1)* %r0 + store i32 %r1.val, i32 addrspace(1)* %r1 + ret void +} + +; GCN-LABEL: {{^}}scalar_xnor_i64_one_use +; GCN: s_xnor_b64 +define amdgpu_kernel void @scalar_xnor_i64_one_use( + i64 addrspace(1)* %r0, i64 %a, i64 %b) { +entry: + %xor = xor i64 %a, %b + %r0.val = xor i64 %xor, -1 + store i64 %r0.val, i64 addrspace(1)* %r0 + ret void +} + +; GCN-LABEL: {{^}}scalar_xnor_i64_mul_use +; GCN-NOT: s_xnor_b64 +; GCN: s_xor_b64 +; GCN: s_not_b64 +; GCN: s_add_u32 +; GCN: s_addc_u32 +define amdgpu_kernel void @scalar_xnor_i64_mul_use( + i64 addrspace(1)* %r0, i64 addrspace(1)* %r1, i64 %a, i64 %b) { +entry: + %xor = xor i64 %a, %b + %r0.val = xor i64 %xor, -1 + %r1.val = add i64 %xor, %a + store i64 %r0.val, i64 addrspace(1)* %r0 + store i64 %r1.val, i64 addrspace(1)* %r1 + ret void +} + +; GCN-LABEL: {{^}}vector_xnor_i32_one_use +; GCN-NOT: s_xnor_b32 +; GCN: v_not_b32 +; GCN: v_xor_b32 +; GCN-DL: v_xnor_b32 +define i32 @vector_xnor_i32_one_use(i32 %a, i32 %b) { +entry: + %xor = xor i32 %a, %b + %r = xor i32 %xor, -1 + ret i32 %r +} + +; GCN-LABEL: {{^}}vector_xnor_i64_one_use +; GCN-NOT: s_xnor_b64 +; GCN: v_not_b32 +; GCN: v_not_b32 +; GCN: v_xor_b32 +; GCN: v_xor_b32 +; GCN-DL: v_xnor_b32 +; GCN-DL: v_xnor_b32 +define i64 @vector_xnor_i64_one_use(i64 %a, i64 %b) { +entry: + %xor = xor i64 %a, %b + %r = xor i64 %xor, -1 + ret i64 %r +} + +; GCN-LABEL: {{^}}xnor_s_v_i32_one_use +; GCN-NOT: s_xnor_b32 +; GCN: s_not_b32 +; GCN: v_xor_b32 +define amdgpu_kernel void @xnor_s_v_i32_one_use(i32 addrspace(1)* %out, i32 %s) { + %v = call i32 @llvm.amdgcn.workitem.id.x() #1 + %xor = xor i32 %s, %v + %d = xor i32 %xor, -1 + store i32 %d, i32 addrspace(1)* %out + ret void +} + +; GCN-LABEL: {{^}}xnor_v_s_i32_one_use +; GCN-NOT: s_xnor_b32 +; GCN: s_not_b32 +; GCN: v_xor_b32 +define amdgpu_kernel void @xnor_v_s_i32_one_use(i32 addrspace(1)* %out, i32 %s) { + %v = call i32 @llvm.amdgcn.workitem.id.x() #1 + %xor = xor i32 %v, %s + %d = xor i32 %xor, -1 + store i32 %d, i32 addrspace(1)* %out + ret void +} + +; GCN-LABEL: {{^}}xnor_i64_s_v_one_use +; GCN-NOT: s_xnor_b64 +; GCN: s_not_b64 +; GCN: v_xor_b32 +; GCN: v_xor_b32 +; GCN-DL: v_xnor_b32 +; GCN-DL: v_xnor_b32 +define amdgpu_kernel void @xnor_i64_s_v_one_use( + i64 addrspace(1)* %r0, i64 %a) { +entry: + %b32 = call i32 @llvm.amdgcn.workitem.id.x() #1 + %b64 = zext i32 %b32 to i64 + %b = shl i64 %b64, 29 + %xor = xor i64 %a, %b + %r0.val = xor i64 %xor, -1 + store i64 %r0.val, i64 addrspace(1)* %r0 + ret void +} + +; GCN-LABEL: {{^}}xnor_i64_v_s_one_use +; GCN-NOT: s_xnor_b64 +; GCN: s_not_b64 +; GCN: v_xor_b32 +; GCN: v_xor_b32 +; GCN-DL: v_xnor_b32 +; GCN-DL: v_xnor_b32 +define amdgpu_kernel void @xnor_i64_v_s_one_use( + i64 addrspace(1)* %r0, i64 %a) { +entry: + %b32 = call i32 @llvm.amdgcn.workitem.id.x() #1 + %b64 = zext i32 %b32 to i64 + %b = shl i64 %b64, 29 + %xor = xor i64 %b, %a + %r0.val = xor i64 %xor, -1 + store i64 %r0.val, i64 addrspace(1)* %r0 + ret void +} + +; GCN-LABEL: {{^}}vector_xor_na_b_i32_one_use +; GCN-NOT: s_xnor_b32 +; GCN: v_not_b32 +; GCN: v_xor_b32 +; GCN-DL: v_xnor_b32 +define i32 @vector_xor_na_b_i32_one_use(i32 %a, i32 %b) { +entry: + %na = xor i32 %a, -1 + %r = xor i32 %na, %b + ret i32 %r +} + +; GCN-LABEL: {{^}}vector_xor_a_nb_i32_one_use +; GCN-NOT: s_xnor_b32 +; GCN: v_not_b32 +; GCN: v_xor_b32 +; GCN-DL: v_xnor_b32 +define i32 @vector_xor_a_nb_i32_one_use(i32 %a, i32 %b) { +entry: + %nb = xor i32 %b, -1 + %r = xor i32 %a, %nb + ret i32 %r +} + +; GCN-LABEL: {{^}}scalar_xor_a_nb_i64_one_use +; GCN: s_xnor_b64 +define amdgpu_kernel void @scalar_xor_a_nb_i64_one_use( + i64 addrspace(1)* %r0, i64 %a, i64 %b) { +entry: + %nb = xor i64 %b, -1 + %r0.val = xor i64 %a, %nb + store i64 %r0.val, i64 addrspace(1)* %r0 + ret void +} + +; GCN-LABEL: {{^}}scalar_xor_na_b_i64_one_use +; GCN: s_xnor_b64 +define amdgpu_kernel void @scalar_xor_na_b_i64_one_use( + i64 addrspace(1)* %r0, i64 %a, i64 %b) { +entry: + %na = xor i64 %a, -1 + %r0.val = xor i64 %na, %b + store i64 %r0.val, i64 addrspace(1)* %r0 + ret void +} + +; Function Attrs: nounwind readnone +declare i32 @llvm.amdgcn.workitem.id.x() #0