Mercurial > hg > CbC > CbC_llvm
diff test/Transforms/InstCombine/align-addr.ll @ 147:c2174574ed3a
LLVM 10
author | Shinji KONO <kono@ie.u-ryukyu.ac.jp> |
---|---|
date | Wed, 14 Aug 2019 16:55:33 +0900 |
parents | 3a76565eade5 |
children |
line wrap: on
line diff
--- a/test/Transforms/InstCombine/align-addr.ll Sat Feb 17 09:57:20 2018 +0900 +++ b/test/Transforms/InstCombine/align-addr.ll Wed Aug 14 16:55:33 2019 +0900 @@ -1,13 +1,33 @@ +; NOTE: Assertions have been autogenerated by utils/update_test_checks.py ; RUN: opt < %s -instcombine -S | FileCheck %s target datalayout = "E-p:64:64:64-p1:32:32:32-a0:0:8-f32:32:32-f64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-v64:64:64-v128:128:128" ; Instcombine should be able to prove vector alignment in the ; presence of a few mild address computation tricks. +define void @test0(i8* %b, i64 %n, i64 %u, i64 %y) nounwind { ; CHECK-LABEL: @test0( -; CHECK: align 16 - -define void @test0(i8* %b, i64 %n, i64 %u, i64 %y) nounwind { +; CHECK-NEXT: entry: +; CHECK-NEXT: [[C:%.*]] = ptrtoint i8* [[B:%.*]] to i64 +; CHECK-NEXT: [[D:%.*]] = and i64 [[C]], -16 +; CHECK-NEXT: [[E:%.*]] = inttoptr i64 [[D]] to double* +; CHECK-NEXT: [[V:%.*]] = shl i64 [[U:%.*]], 1 +; CHECK-NEXT: [[Z:%.*]] = and i64 [[Y:%.*]], -2 +; CHECK-NEXT: [[T1421:%.*]] = icmp eq i64 [[N:%.*]], 0 +; CHECK-NEXT: br i1 [[T1421]], label [[RETURN:%.*]], label [[BB:%.*]] +; CHECK: bb: +; CHECK-NEXT: [[I:%.*]] = phi i64 [ [[INDVAR_NEXT:%.*]], [[BB]] ], [ 20, [[ENTRY:%.*]] ] +; CHECK-NEXT: [[J:%.*]] = mul i64 [[I]], [[V]] +; CHECK-NEXT: [[H:%.*]] = add i64 [[J]], [[Z]] +; CHECK-NEXT: [[T8:%.*]] = getelementptr double, double* [[E]], i64 [[H]] +; CHECK-NEXT: [[P:%.*]] = bitcast double* [[T8]] to <2 x double>* +; CHECK-NEXT: store <2 x double> zeroinitializer, <2 x double>* [[P]], align 16 +; CHECK-NEXT: [[INDVAR_NEXT]] = add i64 [[I]], 1 +; CHECK-NEXT: [[EXITCOND:%.*]] = icmp eq i64 [[INDVAR_NEXT]], [[N]] +; CHECK-NEXT: br i1 [[EXITCOND]], label [[RETURN]], label [[BB]] +; CHECK: return: +; CHECK-NEXT: ret void +; entry: %c = ptrtoint i8* %b to i64 %d = and i64 %c, -16 @@ -35,24 +55,26 @@ ; When we see a unaligned load from an insufficiently aligned global or ; alloca, increase the alignment of the load, turning it into an aligned load. -; CHECK-LABEL: @test1( -; CHECK: tmp = load -; CHECK: GLOBAL{{.*}}align 16 - @GLOBAL = internal global [4 x i32] zeroinitializer define <16 x i8> @test1(<2 x i64> %x) { +; CHECK-LABEL: @test1( +; CHECK-NEXT: entry: +; CHECK-NEXT: [[TMP:%.*]] = load <16 x i8>, <16 x i8>* bitcast ([4 x i32]* @GLOBAL to <16 x i8>*), align 16 +; CHECK-NEXT: ret <16 x i8> [[TMP]] +; entry: - %tmp = load <16 x i8>, <16 x i8>* bitcast ([4 x i32]* @GLOBAL to <16 x i8>*), align 1 - ret <16 x i8> %tmp + %tmp = load <16 x i8>, <16 x i8>* bitcast ([4 x i32]* @GLOBAL to <16 x i8>*), align 1 + ret <16 x i8> %tmp } @GLOBAL_as1 = internal addrspace(1) global [4 x i32] zeroinitializer define <16 x i8> @test1_as1(<2 x i64> %x) { ; CHECK-LABEL: @test1_as1( -; CHECK: tmp = load -; CHECK: GLOBAL_as1{{.*}}align 16 +; CHECK-NEXT: [[TMP:%.*]] = load <16 x i8>, <16 x i8> addrspace(1)* bitcast ([4 x i32] addrspace(1)* @GLOBAL_as1 to <16 x i8> addrspace(1)*), align 16 +; CHECK-NEXT: ret <16 x i8> [[TMP]] +; %tmp = load <16 x i8>, <16 x i8> addrspace(1)* bitcast ([4 x i32] addrspace(1)* @GLOBAL_as1 to <16 x i8> addrspace(1)*), align 1 ret <16 x i8> %tmp } @@ -61,8 +83,9 @@ define <16 x i8> @test1_as1_gep(<2 x i64> %x) { ; CHECK-LABEL: @test1_as1_gep( -; CHECK: tmp = load -; CHECK: GLOBAL_as1_gep{{.*}}align 16 +; CHECK-NEXT: [[TMP:%.*]] = load <16 x i8>, <16 x i8> addrspace(1)* bitcast (i32 addrspace(1)* getelementptr inbounds ([8 x i32], [8 x i32] addrspace(1)* @GLOBAL_as1_gep, i32 0, i32 4) to <16 x i8> addrspace(1)*), align 16 +; CHECK-NEXT: ret <16 x i8> [[TMP]] +; %tmp = load <16 x i8>, <16 x i8> addrspace(1)* bitcast (i32 addrspace(1)* getelementptr ([8 x i32], [8 x i32] addrspace(1)* @GLOBAL_as1_gep, i16 0, i16 4) to <16 x i8> addrspace(1)*), align 1 ret <16 x i8> %tmp } @@ -70,11 +93,12 @@ ; When a load or store lacks an explicit alignment, add one. +define double @test2(double* %p, double %n) nounwind { ; CHECK-LABEL: @test2( -; CHECK: load double, double* %p, align 8 -; CHECK: store double %n, double* %p, align 8 - -define double @test2(double* %p, double %n) nounwind { +; CHECK-NEXT: [[T:%.*]] = load double, double* [[P:%.*]], align 8 +; CHECK-NEXT: store double [[N:%.*]], double* [[P]], align 8 +; CHECK-NEXT: ret double [[T]] +; %t = load double, double* %p store double %n, double* %p ret double %t @@ -89,9 +113,13 @@ define void @test3(%struct.s* sret %a4) { ; Check that the alignment is bumped up the alignment of the sret type. ; CHECK-LABEL: @test3( +; CHECK-NEXT: [[A4_CAST:%.*]] = bitcast %struct.s* [[A4:%.*]] to i8* +; CHECK-NEXT: call void @llvm.memset.p0i8.i64(i8* align 4 dereferenceable(16) [[A4_CAST]], i8 0, i64 16, i1 false) +; CHECK-NEXT: call void @use(i8* [[A4_CAST]]) +; CHECK-NEXT: ret void +; %a4.cast = bitcast %struct.s* %a4 to i8* call void @llvm.memset.p0i8.i64(i8* %a4.cast, i8 0, i64 16, i1 false) -; CHECK: call void @llvm.memset.p0i8.i64(i8* align 4 %a4.cast, i8 0, i64 16, i1 false) call void @use(i8* %a4.cast) ret void }