diff options
Diffstat (limited to 'llvm/test/Analysis/ScalarEvolution')
31 files changed, 73 insertions, 73 deletions
diff --git a/llvm/test/Analysis/ScalarEvolution/2008-07-12-UnneededSelect1.ll b/llvm/test/Analysis/ScalarEvolution/2008-07-12-UnneededSelect1.ll index 6896e7a4728..7e42530798f 100644 --- a/llvm/test/Analysis/ScalarEvolution/2008-07-12-UnneededSelect1.ll +++ b/llvm/test/Analysis/ScalarEvolution/2008-07-12-UnneededSelect1.ll @@ -16,11 +16,11 @@ bb.nph: ; preds = %entry bb: ; preds = %bb1, %bb.nph %j.01 = phi i32 [ %8, %bb1 ], [ 0, %bb.nph ] ; <i32> [#uses=1] - load i32* %srcptr, align 4 ; <i32>:1 [#uses=2] + load i32, i32* %srcptr, align 4 ; <i32>:1 [#uses=2] and i32 %1, 255 ; <i32>:2 [#uses=1] and i32 %1, -256 ; <i32>:3 [#uses=1] getelementptr [256 x i8], [256 x i8]* @lut, i32 0, i32 %2 ; <i8*>:4 [#uses=1] - load i8* %4, align 1 ; <i8>:5 [#uses=1] + load i8, i8* %4, align 1 ; <i8>:5 [#uses=1] zext i8 %5 to i32 ; <i32>:6 [#uses=1] or i32 %6, %3 ; <i32>:7 [#uses=1] store i32 %7, i32* %dstptr, align 4 diff --git a/llvm/test/Analysis/ScalarEvolution/2008-12-08-FiniteSGE.ll b/llvm/test/Analysis/ScalarEvolution/2008-12-08-FiniteSGE.ll index 1d4a27ccc86..0c24ee4eaff 100644 --- a/llvm/test/Analysis/ScalarEvolution/2008-12-08-FiniteSGE.ll +++ b/llvm/test/Analysis/ScalarEvolution/2008-12-08-FiniteSGE.ll @@ -10,7 +10,7 @@ bb1: ; preds = %bb1, %bb1.thread %indvar = phi i32 [ 0, %bb1.thread ], [ %indvar.next, %bb1 ] ; <i32> [#uses=4] %i.0.reg2mem.0 = sub i32 255, %indvar ; <i32> [#uses=2] %0 = getelementptr i32, i32* %alp, i32 %i.0.reg2mem.0 ; <i32*> [#uses=1] - %1 = load i32* %0, align 4 ; <i32> [#uses=1] + %1 = load i32, i32* %0, align 4 ; <i32> [#uses=1] %2 = getelementptr i32, i32* %lam, i32 %i.0.reg2mem.0 ; <i32*> [#uses=1] store i32 %1, i32* %2, align 4 %3 = sub i32 254, %indvar ; <i32> [#uses=1] diff --git a/llvm/test/Analysis/ScalarEvolution/2009-01-02-SignedNegativeStride.ll b/llvm/test/Analysis/ScalarEvolution/2009-01-02-SignedNegativeStride.ll index f19d18c72e6..ebcecbf7429 100644 --- a/llvm/test/Analysis/ScalarEvolution/2009-01-02-SignedNegativeStride.ll +++ b/llvm/test/Analysis/ScalarEvolution/2009-01-02-SignedNegativeStride.ll @@ -9,12 +9,12 @@ define void @func_15() nounwind { entry: - %0 = load i16* @g_16, align 2 ; <i16> [#uses=1] + %0 = load i16, i16* @g_16, align 2 ; <i16> [#uses=1] %1 = icmp sgt i16 %0, 0 ; <i1> [#uses=1] br i1 %1, label %bb2, label %bb.nph bb.nph: ; preds = %entry - %g_16.promoted = load i16* @g_16 ; <i16> [#uses=1] + %g_16.promoted = load i16, i16* @g_16 ; <i16> [#uses=1] br label %bb bb: ; preds = %bb1, %bb.nph diff --git a/llvm/test/Analysis/ScalarEvolution/2009-07-04-GroupConstantsWidthMismatch.ll b/llvm/test/Analysis/ScalarEvolution/2009-07-04-GroupConstantsWidthMismatch.ll index a4358aa6321..d18bdaf7cba 100644 --- a/llvm/test/Analysis/ScalarEvolution/2009-07-04-GroupConstantsWidthMismatch.ll +++ b/llvm/test/Analysis/ScalarEvolution/2009-07-04-GroupConstantsWidthMismatch.ll @@ -3,11 +3,11 @@ define void @test() { entry: - %0 = load i16* undef, align 1 + %0 = load i16, i16* undef, align 1 %1 = lshr i16 %0, 8 %2 = and i16 %1, 3 %3 = zext i16 %2 to i32 - %4 = load i8* undef, align 1 + %4 = load i8, i8* undef, align 1 %5 = lshr i8 %4, 4 %6 = and i8 %5, 1 %7 = zext i8 %6 to i32 diff --git a/llvm/test/Analysis/ScalarEvolution/2012-03-26-LoadConstant.ll b/llvm/test/Analysis/ScalarEvolution/2012-03-26-LoadConstant.ll index 8c6c9b6d1eb..3ca552a7dfd 100644 --- a/llvm/test/Analysis/ScalarEvolution/2012-03-26-LoadConstant.ll +++ b/llvm/test/Analysis/ScalarEvolution/2012-03-26-LoadConstant.ll @@ -19,20 +19,20 @@ lbl_818: ; preds = %for.end, %entry br label %for.cond for.cond: ; preds = %for.body, %lbl_818 - %0 = load i32* @g_814, align 4 + %0 = load i32, i32* @g_814, align 4 %cmp = icmp sle i32 %0, 0 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond %idxprom = sext i32 %0 to i64 %arrayidx = getelementptr inbounds [0 x i32], [0 x i32]* getelementptr inbounds ([1 x [0 x i32]]* @g_244, i32 0, i64 0), i32 0, i64 %idxprom - %1 = load i32* %arrayidx, align 1 + %1 = load i32, i32* %arrayidx, align 1 store i32 %1, i32* @func_21_l_773, align 4 store i32 1, i32* @g_814, align 4 br label %for.cond for.end: ; preds = %for.cond - %2 = load i32* @func_21_l_773, align 4 + %2 = load i32, i32* @func_21_l_773, align 4 %tobool = icmp ne i32 %2, 0 br i1 %tobool, label %lbl_818, label %if.end diff --git a/llvm/test/Analysis/ScalarEvolution/avoid-infinite-recursion-0.ll b/llvm/test/Analysis/ScalarEvolution/avoid-infinite-recursion-0.ll index 7eeb3084958..0976ef92985 100644 --- a/llvm/test/Analysis/ScalarEvolution/avoid-infinite-recursion-0.ll +++ b/llvm/test/Analysis/ScalarEvolution/avoid-infinite-recursion-0.ll @@ -7,7 +7,7 @@ target triple = "x86_64-unknown-linux-gnu" define i32 @test() { entry: - %0 = load i32** undef, align 8 ; <i32*> [#uses=1] + %0 = load i32*, i32** undef, align 8 ; <i32*> [#uses=1] %1 = ptrtoint i32* %0 to i64 ; <i64> [#uses=1] %2 = sub i64 undef, %1 ; <i64> [#uses=1] %3 = lshr i64 %2, 3 ; <i64> [#uses=1] diff --git a/llvm/test/Analysis/ScalarEvolution/avoid-smax-0.ll b/llvm/test/Analysis/ScalarEvolution/avoid-smax-0.ll index e921544f9b4..a282ee6993f 100644 --- a/llvm/test/Analysis/ScalarEvolution/avoid-smax-0.ll +++ b/llvm/test/Analysis/ScalarEvolution/avoid-smax-0.ll @@ -11,7 +11,7 @@ entry: br i1 %0, label %bb, label %return bb: - load i32* %q, align 4 + load i32, i32* %q, align 4 icmp eq i32 %1, 0 br i1 %2, label %return, label %bb3.preheader @@ -21,7 +21,7 @@ bb3.preheader: bb3: %i.0 = phi i32 [ %7, %bb3 ], [ 0, %bb3.preheader ] getelementptr i32, i32* %p, i32 %i.0 - load i32* %3, align 4 + load i32, i32* %3, align 4 add i32 %4, 1 getelementptr i32, i32* %p, i32 %i.0 store i32 %5, i32* %6, align 4 diff --git a/llvm/test/Analysis/ScalarEvolution/avoid-smax-1.ll b/llvm/test/Analysis/ScalarEvolution/avoid-smax-1.ll index 685a106c296..e6c62ee6b47 100644 --- a/llvm/test/Analysis/ScalarEvolution/avoid-smax-1.ll +++ b/llvm/test/Analysis/ScalarEvolution/avoid-smax-1.ll @@ -36,7 +36,7 @@ bb6: ; preds = %bb7, %bb.nph7 %8 = shl i32 %x.06, 1 ; <i32> [#uses=1] %9 = add i32 %6, %8 ; <i32> [#uses=1] %10 = getelementptr i8, i8* %r, i32 %9 ; <i8*> [#uses=1] - %11 = load i8* %10, align 1 ; <i8> [#uses=1] + %11 = load i8, i8* %10, align 1 ; <i8> [#uses=1] %12 = getelementptr i8, i8* %j, i32 %7 ; <i8*> [#uses=1] store i8 %11, i8* %12, align 1 %13 = add i32 %x.06, 1 ; <i32> [#uses=2] @@ -103,7 +103,7 @@ bb14: ; preds = %bb15, %bb.nph3 %29 = shl i32 %x.12, 2 ; <i32> [#uses=1] %30 = add i32 %29, %25 ; <i32> [#uses=1] %31 = getelementptr i8, i8* %r, i32 %30 ; <i8*> [#uses=1] - %32 = load i8* %31, align 1 ; <i8> [#uses=1] + %32 = load i8, i8* %31, align 1 ; <i8> [#uses=1] %.sum = add i32 %26, %x.12 ; <i32> [#uses=1] %33 = getelementptr i8, i8* %j, i32 %.sum ; <i8*> [#uses=1] store i8 %32, i8* %33, align 1 @@ -111,7 +111,7 @@ bb14: ; preds = %bb15, %bb.nph3 %35 = or i32 %34, 2 ; <i32> [#uses=1] %36 = add i32 %35, %25 ; <i32> [#uses=1] %37 = getelementptr i8, i8* %r, i32 %36 ; <i8*> [#uses=1] - %38 = load i8* %37, align 1 ; <i8> [#uses=1] + %38 = load i8, i8* %37, align 1 ; <i8> [#uses=1] %.sum6 = add i32 %27, %x.12 ; <i32> [#uses=1] %39 = getelementptr i8, i8* %j, i32 %.sum6 ; <i8*> [#uses=1] store i8 %38, i8* %39, align 1 diff --git a/llvm/test/Analysis/ScalarEvolution/infer-prestart-no-wrap.ll b/llvm/test/Analysis/ScalarEvolution/infer-prestart-no-wrap.ll index c9689f7fe14..078ca03ff14 100644 --- a/llvm/test/Analysis/ScalarEvolution/infer-prestart-no-wrap.ll +++ b/llvm/test/Analysis/ScalarEvolution/infer-prestart-no-wrap.ll @@ -53,7 +53,7 @@ define void @infer.sext.1(i32 %start, i1* %c) { ; CHECK: %idx.sext = sext i32 %idx to i64 ; CHECK-NEXT: --> {(2 + (sext i32 (4 * %start) to i64)),+,2}<nsw><%loop> %idx.inc = add nsw i32 %idx, 2 - %condition = load i1* %c + %condition = load i1, i1* %c br i1 %condition, label %exit, label %loop exit: @@ -73,7 +73,7 @@ define void @infer.sext.2(i1* %c, i8 %start) { ; CHECK: %idx.sext = sext i8 %idx to i16 ; CHECK-NEXT: --> {(1 + (sext i8 %start to i16)),+,1}<nsw><%loop> %idx.inc = add nsw i8 %idx, 1 - %condition = load volatile i1* %c + %condition = load volatile i1, i1* %c br i1 %condition, label %exit, label %loop exit: @@ -93,7 +93,7 @@ define void @infer.zext.1(i1* %c, i8 %start) { ; CHECK: %idx.zext = zext i8 %idx to i16 ; CHECK-NEXT: --> {(1 + (zext i8 %start to i16)),+,1}<nuw><%loop> %idx.inc = add nuw i8 %idx, 1 - %condition = load volatile i1* %c + %condition = load volatile i1, i1* %c br i1 %condition, label %exit, label %loop exit: diff --git a/llvm/test/Analysis/ScalarEvolution/load-with-range-metadata.ll b/llvm/test/Analysis/ScalarEvolution/load-with-range-metadata.ll index 32c1074b174..f26c8d56754 100644 --- a/llvm/test/Analysis/ScalarEvolution/load-with-range-metadata.ll +++ b/llvm/test/Analysis/ScalarEvolution/load-with-range-metadata.ll @@ -3,7 +3,7 @@ define i32 @slt_trip_count_with_range(i32 *%ptr0, i32 *%ptr1) { ; CHECK-LABEL: slt_trip_count_with_range entry: - %limit = load i32* %ptr0, !range !0 + %limit = load i32, i32* %ptr0, !range !0 br label %loop loop: @@ -20,7 +20,7 @@ define i32 @slt_trip_count_with_range(i32 *%ptr0, i32 *%ptr1) { define i32 @ult_trip_count_with_range(i32 *%ptr0, i32 *%ptr1) { ; CHECK-LABEL: ult_trip_count_with_range entry: - %limit = load i32* %ptr0, !range !0 + %limit = load i32, i32* %ptr0, !range !0 br label %loop loop: diff --git a/llvm/test/Analysis/ScalarEvolution/load.ll b/llvm/test/Analysis/ScalarEvolution/load.ll index 8b460a806cb..f73e7f1c201 100644 --- a/llvm/test/Analysis/ScalarEvolution/load.ll +++ b/llvm/test/Analysis/ScalarEvolution/load.ll @@ -17,10 +17,10 @@ for.body: ; preds = %entry, %for.body ; CHECK: --> %sum.04{{ *}}Exits: 2450 %i.03 = phi i32 [ 0, %entry ], [ %inc, %for.body ] %arrayidx = getelementptr inbounds [50 x i32], [50 x i32]* @arr1, i32 0, i32 %i.03 - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 ; CHECK: --> %0{{ *}}Exits: 50 %arrayidx1 = getelementptr inbounds [50 x i32], [50 x i32]* @arr2, i32 0, i32 %i.03 - %1 = load i32* %arrayidx1, align 4 + %1 = load i32, i32* %arrayidx1, align 4 ; CHECK: --> %1{{ *}}Exits: 0 %add = add i32 %0, %sum.04 %add2 = add i32 %add, %1 @@ -52,10 +52,10 @@ for.body: ; preds = %entry, %for.body %n.01 = phi %struct.ListNode* [ bitcast ({ %struct.ListNode*, i32, [4 x i8] }* @node5 to %struct.ListNode*), %entry ], [ %1, %for.body ] ; CHECK: --> %n.01{{ *}}Exits: @node1 %i = getelementptr inbounds %struct.ListNode, %struct.ListNode* %n.01, i64 0, i32 1 - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %add = add nsw i32 %0, %sum.02 %next = getelementptr inbounds %struct.ListNode, %struct.ListNode* %n.01, i64 0, i32 0 - %1 = load %struct.ListNode** %next, align 8 + %1 = load %struct.ListNode*, %struct.ListNode** %next, align 8 ; CHECK: --> %1{{ *}}Exits: 0 %cmp = icmp eq %struct.ListNode* %1, null br i1 %cmp, label %for.end, label %for.body diff --git a/llvm/test/Analysis/ScalarEvolution/max-trip-count.ll b/llvm/test/Analysis/ScalarEvolution/max-trip-count.ll index 4faedde8757..abebea7d74c 100644 --- a/llvm/test/Analysis/ScalarEvolution/max-trip-count.ll +++ b/llvm/test/Analysis/ScalarEvolution/max-trip-count.ll @@ -174,7 +174,7 @@ for.body.i: ; preds = %for.cond.i, %entry for.cond.i: ; preds = %for.body.i store i32 %add.i.i, i32* @a, align 4 - %ld = load volatile i32* @b + %ld = load volatile i32, i32* @b %cmp.i = icmp ne i32 %ld, 0 br i1 %cmp.i, label %for.body.i, label %bar.exit diff --git a/llvm/test/Analysis/ScalarEvolution/min-max-exprs.ll b/llvm/test/Analysis/ScalarEvolution/min-max-exprs.ll index b9ede6f7e44..892fc23fe6b 100644 --- a/llvm/test/Analysis/ScalarEvolution/min-max-exprs.ll +++ b/llvm/test/Analysis/ScalarEvolution/min-max-exprs.ll @@ -35,7 +35,7 @@ bb2: ; preds = %bb1 ; CHECK: select i1 %tmp4, i64 %tmp5, i64 %tmp6 ; CHECK-NEXT: --> (-1 + (-1 * ((-1 + (-1 * (sext i32 {3,+,1}<nw><%bb1> to i64))) smax (-1 + (-1 * (sext i32 %N to i64)))))) %tmp11 = getelementptr inbounds i32, i32* %A, i64 %tmp9 - %tmp12 = load i32* %tmp11, align 4 + %tmp12 = load i32, i32* %tmp11, align 4 %tmp13 = shl nsw i32 %tmp12, 1 %tmp14 = icmp sge i32 3, %i.0 %tmp17 = add nsw i64 %i.0.1, -3 diff --git a/llvm/test/Analysis/ScalarEvolution/nsw-offset-assume.ll b/llvm/test/Analysis/ScalarEvolution/nsw-offset-assume.ll index 246f9ad1abc..bef1070bbcb 100644 --- a/llvm/test/Analysis/ScalarEvolution/nsw-offset-assume.ll +++ b/llvm/test/Analysis/ScalarEvolution/nsw-offset-assume.ll @@ -28,10 +28,10 @@ bb: ; preds = %bb.nph, %bb1 ; CHECK: --> {%d,+,16}<nsw><%bb> %2 = getelementptr inbounds double, double* %d, i64 %1 ; <double*> [#uses=1] - %3 = load double* %2, align 8 ; <double> [#uses=1] + %3 = load double, double* %2, align 8 ; <double> [#uses=1] %4 = sext i32 %i.01 to i64 ; <i64> [#uses=1] %5 = getelementptr inbounds double, double* %q, i64 %4 ; <double*> [#uses=1] - %6 = load double* %5, align 8 ; <double> [#uses=1] + %6 = load double, double* %5, align 8 ; <double> [#uses=1] %7 = or i32 %i.01, 1 ; <i32> [#uses=1] ; CHECK: %8 = sext i32 %7 to i64 @@ -54,7 +54,7 @@ bb: ; preds = %bb.nph, %bb1 ; CHECK: {(8 + %q),+,16}<nsw><%bb> %t9 = getelementptr inbounds double, double* %q, i64 %t8 ; <double*> [#uses=1] - %10 = load double* %9, align 8 ; <double> [#uses=1] + %10 = load double, double* %9, align 8 ; <double> [#uses=1] %11 = fadd double %6, %10 ; <double> [#uses=1] %12 = fadd double %11, 3.200000e+00 ; <double> [#uses=1] %13 = fmul double %3, %12 ; <double> [#uses=1] diff --git a/llvm/test/Analysis/ScalarEvolution/nsw-offset.ll b/llvm/test/Analysis/ScalarEvolution/nsw-offset.ll index 7b8de519429..127bb19bff0 100644 --- a/llvm/test/Analysis/ScalarEvolution/nsw-offset.ll +++ b/llvm/test/Analysis/ScalarEvolution/nsw-offset.ll @@ -26,10 +26,10 @@ bb: ; preds = %bb.nph, %bb1 ; CHECK: --> {%d,+,16}<nsw><%bb> %2 = getelementptr inbounds double, double* %d, i64 %1 ; <double*> [#uses=1] - %3 = load double* %2, align 8 ; <double> [#uses=1] + %3 = load double, double* %2, align 8 ; <double> [#uses=1] %4 = sext i32 %i.01 to i64 ; <i64> [#uses=1] %5 = getelementptr inbounds double, double* %q, i64 %4 ; <double*> [#uses=1] - %6 = load double* %5, align 8 ; <double> [#uses=1] + %6 = load double, double* %5, align 8 ; <double> [#uses=1] %7 = or i32 %i.01, 1 ; <i32> [#uses=1] ; CHECK: %8 = sext i32 %7 to i64 @@ -52,7 +52,7 @@ bb: ; preds = %bb.nph, %bb1 ; CHECK: {(8 + %q),+,16}<nsw><%bb> %t9 = getelementptr inbounds double, double* %q, i64 %t8 ; <double*> [#uses=1] - %10 = load double* %9, align 8 ; <double> [#uses=1] + %10 = load double, double* %9, align 8 ; <double> [#uses=1] %11 = fadd double %6, %10 ; <double> [#uses=1] %12 = fadd double %11, 3.200000e+00 ; <double> [#uses=1] %13 = fmul double %3, %12 ; <double> [#uses=1] diff --git a/llvm/test/Analysis/ScalarEvolution/nsw.ll b/llvm/test/Analysis/ScalarEvolution/nsw.ll index 024b2804c06..0a3c535b049 100644 --- a/llvm/test/Analysis/ScalarEvolution/nsw.ll +++ b/llvm/test/Analysis/ScalarEvolution/nsw.ll @@ -7,7 +7,7 @@ target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64" ; CHECK: Classifying expressions for: @test1 define void @test1(double* %p) nounwind { entry: - %tmp = load double* %p, align 8 ; <double> [#uses=1] + %tmp = load double, double* %p, align 8 ; <double> [#uses=1] %tmp1 = fcmp ogt double %tmp, 2.000000e+00 ; <i1> [#uses=1] br i1 %tmp1, label %bb.nph, label %return @@ -20,7 +20,7 @@ bb: ; preds = %bb1, %bb.nph ; CHECK-NEXT: --> {0,+,1}<nuw><nsw><%bb> %tmp2 = sext i32 %i.01 to i64 ; <i64> [#uses=1] %tmp3 = getelementptr double, double* %p, i64 %tmp2 ; <double*> [#uses=1] - %tmp4 = load double* %tmp3, align 8 ; <double> [#uses=1] + %tmp4 = load double, double* %tmp3, align 8 ; <double> [#uses=1] %tmp5 = fmul double %tmp4, 9.200000e+00 ; <double> [#uses=1] %tmp6 = sext i32 %i.01 to i64 ; <i64> [#uses=1] %tmp7 = getelementptr double, double* %p, i64 %tmp6 ; <double*> [#uses=1] @@ -39,7 +39,7 @@ bb1: ; preds = %bb %tmp9 = getelementptr double, double* %p, i64 %phitmp ; <double*> [#uses=1] ; CHECK: %tmp9 ; CHECK-NEXT: --> {(8 + %p),+,8}<%bb> - %tmp10 = load double* %tmp9, align 8 ; <double> [#uses=1] + %tmp10 = load double, double* %tmp9, align 8 ; <double> [#uses=1] %tmp11 = fcmp ogt double %tmp10, 2.000000e+00 ; <i1> [#uses=1] br i1 %tmp11, label %bb, label %bb1.return_crit_edge diff --git a/llvm/test/Analysis/ScalarEvolution/pr22179.ll b/llvm/test/Analysis/ScalarEvolution/pr22179.ll index d9fb5104436..5dc41920e86 100644 --- a/llvm/test/Analysis/ScalarEvolution/pr22179.ll +++ b/llvm/test/Analysis/ScalarEvolution/pr22179.ll @@ -14,7 +14,7 @@ define i32 @main() { loop: %storemerge1 = phi i8 [ 0, %0 ], [ %inc, %loop ] - %m = load volatile i32* getelementptr inbounds (%struct.S* @b, i64 0, i32 0), align 4 + %m = load volatile i32, i32* getelementptr inbounds (%struct.S* @b, i64 0, i32 0), align 4 %inc = add nuw i8 %storemerge1, 1 ; CHECK: %inc = add nuw i8 %storemerge1, 1 ; CHECK-NEXT: --> {1,+,1}<nuw><%loop> diff --git a/llvm/test/Analysis/ScalarEvolution/pr22674.ll b/llvm/test/Analysis/ScalarEvolution/pr22674.ll index 6b7a143f11e..1bc7fd327d2 100644 --- a/llvm/test/Analysis/ScalarEvolution/pr22674.ll +++ b/llvm/test/Analysis/ScalarEvolution/pr22674.ll @@ -45,9 +45,9 @@ cond.false: ; preds = %for.end, %for.inc, _ZNK4llvm12AttributeSet3endEj.exit: ; preds = %for.end %second.i.i.i = getelementptr inbounds %"struct.std::pair.241.2040.3839.6152.6923.7694.8465.9493.10007.10264.18507", %"struct.std::pair.241.2040.3839.6152.6923.7694.8465.9493.10007.10264.18507"* undef, i32 %I.099.lcssa129, i32 1 - %0 = load %"class.llvm::AttributeSetNode.230.2029.3828.6141.6912.7683.8454.9482.9996.10253.18506"** %second.i.i.i, align 4, !tbaa !2 + %0 = load %"class.llvm::AttributeSetNode.230.2029.3828.6141.6912.7683.8454.9482.9996.10253.18506"*, %"class.llvm::AttributeSetNode.230.2029.3828.6141.6912.7683.8454.9482.9996.10253.18506"** %second.i.i.i, align 4, !tbaa !2 %NumAttrs.i.i.i = getelementptr inbounds %"class.llvm::AttributeSetNode.230.2029.3828.6141.6912.7683.8454.9482.9996.10253.18506", %"class.llvm::AttributeSetNode.230.2029.3828.6141.6912.7683.8454.9482.9996.10253.18506"* %0, i32 0, i32 1 - %1 = load i32* %NumAttrs.i.i.i, align 4, !tbaa !8 + %1 = load i32, i32* %NumAttrs.i.i.i, align 4, !tbaa !8 %add.ptr.i.i.i55 = getelementptr inbounds %"class.llvm::Attribute.222.2021.3820.6133.6904.7675.8446.9474.9988.10245.18509", %"class.llvm::Attribute.222.2021.3820.6133.6904.7675.8446.9474.9988.10245.18509"* undef, i32 %1 br i1 undef, label %return, label %for.body11 @@ -58,7 +58,7 @@ for.cond9: ; preds = %_ZNK4llvm9Attribute for.body11: ; preds = %for.cond9, %_ZNK4llvm12AttributeSet3endEj.exit %I5.096 = phi %"class.llvm::Attribute.222.2021.3820.6133.6904.7675.8446.9474.9988.10245.18509"* [ %incdec.ptr, %for.cond9 ], [ undef, %_ZNK4llvm12AttributeSet3endEj.exit ] %2 = bitcast %"class.llvm::Attribute.222.2021.3820.6133.6904.7675.8446.9474.9988.10245.18509"* %I5.096 to i32* - %3 = load i32* %2, align 4, !tbaa !10 + %3 = load i32, i32* %2, align 4, !tbaa !10 %tobool.i59 = icmp eq i32 %3, 0 br i1 %tobool.i59, label %cond.false21, label %_ZNK4llvm9Attribute15isEnumAttributeEv.exit diff --git a/llvm/test/Analysis/ScalarEvolution/scev-aa.ll b/llvm/test/Analysis/ScalarEvolution/scev-aa.ll index 9a3b9cd228e..e2123f44f16 100644 --- a/llvm/test/Analysis/ScalarEvolution/scev-aa.ll +++ b/llvm/test/Analysis/ScalarEvolution/scev-aa.ll @@ -22,8 +22,8 @@ bb: %pi = getelementptr double, double* %p, i64 %i %i.next = add i64 %i, 1 %pi.next = getelementptr double, double* %p, i64 %i.next - %x = load double* %pi - %y = load double* %pi.next + %x = load double, double* %pi + %y = load double, double* %pi.next %z = fmul double %x, %y store double %z, double* %pi %exitcond = icmp eq i64 %i.next, %n @@ -61,15 +61,15 @@ bb: %pi.j = getelementptr double, double* %p, i64 %e %f = add i64 %i.next, %j %pi.next.j = getelementptr double, double* %p, i64 %f - %x = load double* %pi.j - %y = load double* %pi.next.j + %x = load double, double* %pi.j + %y = load double, double* %pi.next.j %z = fmul double %x, %y store double %z, double* %pi.j %o = add i64 %j, 91 %g = add i64 %i, %o %pi.j.next = getelementptr double, double* %p, i64 %g - %a = load double* %pi.j.next + %a = load double, double* %pi.j.next %b = fmul double %x, %a store double %b, double* %pi.j.next @@ -118,15 +118,15 @@ bb: %pi.j = getelementptr double, double* %p, i64 %e %f = add i64 %i.next, %j %pi.next.j = getelementptr double, double* %p, i64 %f - %x = load double* %pi.j - %y = load double* %pi.next.j + %x = load double, double* %pi.j + %y = load double, double* %pi.next.j %z = fmul double %x, %y store double %z, double* %pi.j %o = add i64 %j, %n %g = add i64 %i, %o %pi.j.next = getelementptr double, double* %p, i64 %g - %a = load double* %pi.j.next + %a = load double, double* %pi.j.next %b = fmul double %x, %a store double %b, double* %pi.j.next @@ -202,7 +202,7 @@ for.body: ; preds = %entry, %for.body %inc = add nsw i64 %i, 1 ; <i64> [#uses=2] %arrayidx = getelementptr inbounds i64, i64* %p, i64 %inc store i64 0, i64* %arrayidx - %tmp6 = load i64* %p ; <i64> [#uses=1] + %tmp6 = load i64, i64* %p ; <i64> [#uses=1] %cmp = icmp slt i64 %inc, %tmp6 ; <i1> [#uses=1] br i1 %cmp, label %for.body, label %for.end diff --git a/llvm/test/Analysis/ScalarEvolution/scev-prestart-nowrap.ll b/llvm/test/Analysis/ScalarEvolution/scev-prestart-nowrap.ll index 3ca32bdd9a1..77f3482f03c 100644 --- a/llvm/test/Analysis/ScalarEvolution/scev-prestart-nowrap.ll +++ b/llvm/test/Analysis/ScalarEvolution/scev-prestart-nowrap.ll @@ -66,7 +66,7 @@ define i64 @bad.1(i32 %start, i32 %low.limit, i32 %high.limit, i1* %unknown) { br i1 %break.early, label %continue.1, label %early.exit continue.1: - %cond = load volatile i1* %unknown + %cond = load volatile i1, i1* %unknown %idx.inc = add nsw i32 %idx, 1 br i1 %cond, label %loop, label %continue diff --git a/llvm/test/Analysis/ScalarEvolution/sext-iv-0.ll b/llvm/test/Analysis/ScalarEvolution/sext-iv-0.ll index f5d54556d24..47f02715c0d 100644 --- a/llvm/test/Analysis/ScalarEvolution/sext-iv-0.ll +++ b/llvm/test/Analysis/ScalarEvolution/sext-iv-0.ll @@ -24,7 +24,7 @@ bb1: ; preds = %bb1, %bb1.thread ; CHECK: %2 ; CHECK-NEXT: --> {-128,+,1}<nsw><%bb1> Exits: 127 %3 = getelementptr double, double* %x, i64 %2 ; <double*> [#uses=1] - %4 = load double* %3, align 8 ; <double> [#uses=1] + %4 = load double, double* %3, align 8 ; <double> [#uses=1] %5 = fmul double %4, 3.900000e+00 ; <double> [#uses=1] %6 = sext i8 %0 to i64 ; <i64> [#uses=1] %7 = getelementptr double, double* %x, i64 %6 ; <double*> [#uses=1] diff --git a/llvm/test/Analysis/ScalarEvolution/sext-iv-1.ll b/llvm/test/Analysis/ScalarEvolution/sext-iv-1.ll index 07f055e4d17..575b744a1a4 100644 --- a/llvm/test/Analysis/ScalarEvolution/sext-iv-1.ll +++ b/llvm/test/Analysis/ScalarEvolution/sext-iv-1.ll @@ -24,7 +24,7 @@ bb1: ; preds = %bb1, %bb1.thread %1 = trunc i64 %i.0.reg2mem.0 to i9 ; <i8> [#uses=1] %2 = sext i9 %1 to i64 ; <i64> [#uses=1] %3 = getelementptr double, double* %x, i64 %2 ; <double*> [#uses=1] - %4 = load double* %3, align 8 ; <double> [#uses=1] + %4 = load double, double* %3, align 8 ; <double> [#uses=1] %5 = fmul double %4, 3.900000e+00 ; <double> [#uses=1] %6 = sext i7 %0 to i64 ; <i64> [#uses=1] %7 = getelementptr double, double* %x, i64 %6 ; <double*> [#uses=1] @@ -47,7 +47,7 @@ bb1: ; preds = %bb1, %bb1.thread %1 = trunc i64 %i.0.reg2mem.0 to i9 ; <i8> [#uses=1] %2 = sext i9 %1 to i64 ; <i64> [#uses=1] %3 = getelementptr double, double* %x, i64 %2 ; <double*> [#uses=1] - %4 = load double* %3, align 8 ; <double> [#uses=1] + %4 = load double, double* %3, align 8 ; <double> [#uses=1] %5 = fmul double %4, 3.900000e+00 ; <double> [#uses=1] %6 = sext i8 %0 to i64 ; <i64> [#uses=1] %7 = getelementptr double, double* %x, i64 %6 ; <double*> [#uses=1] @@ -70,7 +70,7 @@ bb1: ; preds = %bb1, %bb1.thread %1 = trunc i64 %i.0.reg2mem.0 to i9 ; <i8> [#uses=1] %2 = sext i9 %1 to i64 ; <i64> [#uses=1] %3 = getelementptr double, double* %x, i64 %2 ; <double*> [#uses=1] - %4 = load double* %3, align 8 ; <double> [#uses=1] + %4 = load double, double* %3, align 8 ; <double> [#uses=1] %5 = fmul double %4, 3.900000e+00 ; <double> [#uses=1] %6 = sext i8 %0 to i64 ; <i64> [#uses=1] %7 = getelementptr double, double* %x, i64 %6 ; <double*> [#uses=1] @@ -93,7 +93,7 @@ bb1: ; preds = %bb1, %bb1.thread %1 = trunc i64 %i.0.reg2mem.0 to i9 ; <i8> [#uses=1] %2 = sext i9 %1 to i64 ; <i64> [#uses=1] %3 = getelementptr double, double* %x, i64 %2 ; <double*> [#uses=1] - %4 = load double* %3, align 8 ; <double> [#uses=1] + %4 = load double, double* %3, align 8 ; <double> [#uses=1] %5 = fmul double %4, 3.900000e+00 ; <double> [#uses=1] %6 = sext i8 %0 to i64 ; <i64> [#uses=1] %7 = getelementptr double, double* %x, i64 %6 ; <double*> [#uses=1] diff --git a/llvm/test/Analysis/ScalarEvolution/sext-iv-2.ll b/llvm/test/Analysis/ScalarEvolution/sext-iv-2.ll index e580cc18d98..6e075739b43 100644 --- a/llvm/test/Analysis/ScalarEvolution/sext-iv-2.ll +++ b/llvm/test/Analysis/ScalarEvolution/sext-iv-2.ll @@ -56,7 +56,7 @@ bb4.bb5_crit_edge: ; preds = %bb4 br label %bb5 bb5: ; preds = %bb4.bb5_crit_edge, %entry - %tmp12 = load i32* getelementptr ([32 x [256 x i32]]* @table, i64 0, i64 9, i64 132), align 16 ; <i32> [#uses=1] + %tmp12 = load i32, i32* getelementptr ([32 x [256 x i32]]* @table, i64 0, i64 9, i64 132), align 16 ; <i32> [#uses=1] %tmp13 = icmp eq i32 %tmp12, -1116 ; <i1> [#uses=1] br i1 %tmp13, label %bb7, label %bb6 diff --git a/llvm/test/Analysis/ScalarEvolution/sle.ll b/llvm/test/Analysis/ScalarEvolution/sle.ll index c31f750cddb..f24c4807114 100644 --- a/llvm/test/Analysis/ScalarEvolution/sle.ll +++ b/llvm/test/Analysis/ScalarEvolution/sle.ll @@ -15,7 +15,7 @@ entry: for.body: ; preds = %for.body, %entry %i = phi i64 [ %i.next, %for.body ], [ 0, %entry ] ; <i64> [#uses=2] %arrayidx = getelementptr double, double* %p, i64 %i ; <double*> [#uses=2] - %t4 = load double* %arrayidx ; <double> [#uses=1] + %t4 = load double, double* %arrayidx ; <double> [#uses=1] %mul = fmul double %t4, 2.200000e+00 ; <double> [#uses=1] store double %mul, double* %arrayidx %i.next = add nsw i64 %i, 1 ; <i64> [#uses=2] diff --git a/llvm/test/Analysis/ScalarEvolution/trip-count11.ll b/llvm/test/Analysis/ScalarEvolution/trip-count11.ll index 3faa95176e7..b0a2c4094a5 100644 --- a/llvm/test/Analysis/ScalarEvolution/trip-count11.ll +++ b/llvm/test/Analysis/ScalarEvolution/trip-count11.ll @@ -21,7 +21,7 @@ for.cond: ; preds = %for.inc, %entry for.inc: ; preds = %for.cond %idxprom = sext i32 %i.0 to i64 %arrayidx = getelementptr inbounds [8 x i32], [8 x i32]* @foo.a, i64 0, i64 %idxprom - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %add = add nsw i32 %sum.0, %0 %inc = add nsw i32 %i.0, 1 br label %for.cond @@ -44,7 +44,7 @@ for.cond: ; preds = %for.inc, %entry for.inc: ; preds = %for.cond %idxprom = sext i32 %i.0 to i64 %arrayidx = getelementptr inbounds [8 x i32], [8 x i32] addrspace(1)* @foo.a_as1, i64 0, i64 %idxprom - %0 = load i32 addrspace(1)* %arrayidx, align 4 + %0 = load i32, i32 addrspace(1)* %arrayidx, align 4 %add = add nsw i32 %sum.0, %0 %inc = add nsw i32 %i.0, 1 br label %for.cond diff --git a/llvm/test/Analysis/ScalarEvolution/trip-count12.ll b/llvm/test/Analysis/ScalarEvolution/trip-count12.ll index 3fd16b23df6..d0086ee2e6a 100644 --- a/llvm/test/Analysis/ScalarEvolution/trip-count12.ll +++ b/llvm/test/Analysis/ScalarEvolution/trip-count12.ll @@ -17,7 +17,7 @@ for.body: ; preds = %for.body, %for.body %len.addr.04 = phi i32 [ %sub, %for.body ], [ %len, %for.body.preheader ] %res.03 = phi i32 [ %add, %for.body ], [ 0, %for.body.preheader ] %incdec.ptr = getelementptr inbounds i16, i16* %p.addr.05, i32 1 - %0 = load i16* %p.addr.05, align 2 + %0 = load i16, i16* %p.addr.05, align 2 %conv = zext i16 %0 to i32 %add = add i32 %conv, %res.03 %sub = add nsw i32 %len.addr.04, -2 diff --git a/llvm/test/Analysis/ScalarEvolution/trip-count4.ll b/llvm/test/Analysis/ScalarEvolution/trip-count4.ll index 6c1ed89989b..966ffd26fd2 100644 --- a/llvm/test/Analysis/ScalarEvolution/trip-count4.ll +++ b/llvm/test/Analysis/ScalarEvolution/trip-count4.ll @@ -13,7 +13,7 @@ loop: ; preds = %loop, %entry %s0 = shl i64 %indvar, 8 ; <i64> [#uses=1] %indvar.i8 = ashr i64 %s0, 8 ; <i64> [#uses=1] %t0 = getelementptr double, double* %d, i64 %indvar.i8 ; <double*> [#uses=2] - %t1 = load double* %t0 ; <double> [#uses=1] + %t1 = load double, double* %t0 ; <double> [#uses=1] %t2 = fmul double %t1, 1.000000e-01 ; <double> [#uses=1] store double %t2, double* %t0 %indvar.next = sub i64 %indvar, 1 ; <i64> [#uses=2] diff --git a/llvm/test/Analysis/ScalarEvolution/trip-count5.ll b/llvm/test/Analysis/ScalarEvolution/trip-count5.ll index 564a75a7458..dc02fedd134 100644 --- a/llvm/test/Analysis/ScalarEvolution/trip-count5.ll +++ b/llvm/test/Analysis/ScalarEvolution/trip-count5.ll @@ -9,7 +9,7 @@ target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f3 define float @t(float* %pTmp1, float* %peakWeight, float* %nrgReducePeakrate, i32 %bim) nounwind { entry: - %tmp3 = load float* %peakWeight, align 4 ; <float> [#uses=2] + %tmp3 = load float, float* %peakWeight, align 4 ; <float> [#uses=2] %tmp2538 = icmp sgt i32 %bim, 0 ; <i1> [#uses=1] br i1 %tmp2538, label %bb.nph, label %bb4 @@ -22,12 +22,12 @@ bb: ; preds = %bb1, %bb.nph %peakCount.034 = phi float [ %tmp19, %bb1 ], [ %tmp3, %bb.nph ] ; <float> [#uses=1] %tmp6 = sext i32 %hiPart.035 to i64 ; <i64> [#uses=1] %tmp7 = getelementptr float, float* %pTmp1, i64 %tmp6 ; <float*> [#uses=1] - %tmp8 = load float* %tmp7, align 4 ; <float> [#uses=1] + %tmp8 = load float, float* %tmp7, align 4 ; <float> [#uses=1] %tmp10 = fadd float %tmp8, %distERBhi.036 ; <float> [#uses=3] %tmp12 = add i32 %hiPart.035, 1 ; <i32> [#uses=3] %tmp15 = sext i32 %tmp12 to i64 ; <i64> [#uses=1] %tmp16 = getelementptr float, float* %peakWeight, i64 %tmp15 ; <float*> [#uses=1] - %tmp17 = load float* %tmp16, align 4 ; <float> [#uses=1] + %tmp17 = load float, float* %tmp16, align 4 ; <float> [#uses=1] %tmp19 = fadd float %tmp17, %peakCount.034 ; <float> [#uses=2] br label %bb1 diff --git a/llvm/test/Analysis/ScalarEvolution/trip-count6.ll b/llvm/test/Analysis/ScalarEvolution/trip-count6.ll index 9cba1101a6f..7980bbdcb12 100644 --- a/llvm/test/Analysis/ScalarEvolution/trip-count6.ll +++ b/llvm/test/Analysis/ScalarEvolution/trip-count6.ll @@ -13,7 +13,7 @@ bb: ; preds = %bb4, %entry %mode.0 = phi i8 [ 0, %entry ], [ %indvar.next, %bb4 ] ; <i8> [#uses=4] zext i8 %mode.0 to i32 ; <i32>:1 [#uses=1] getelementptr [4 x i32], [4 x i32]* @mode_table, i32 0, i32 %1 ; <i32*>:2 [#uses=1] - load i32* %2, align 4 ; <i32>:3 [#uses=1] + load i32, i32* %2, align 4 ; <i32>:3 [#uses=1] icmp eq i32 %3, %0 ; <i1>:4 [#uses=1] br i1 %4, label %bb1, label %bb2 diff --git a/llvm/test/Analysis/ScalarEvolution/trip-count7.ll b/llvm/test/Analysis/ScalarEvolution/trip-count7.ll index a4eb72f0737..bbe76c490cd 100644 --- a/llvm/test/Analysis/ScalarEvolution/trip-count7.ll +++ b/llvm/test/Analysis/ScalarEvolution/trip-count7.ll @@ -73,7 +73,7 @@ bb.i: ; preds = %bb7.i store i32 0, i32* %q, align 4 %tmp1 = sext i32 %tmp to i64 ; <i64> [#uses=1] %tmp2 = getelementptr [9 x i32], [9 x i32]* %a, i64 0, i64 %tmp1 ; <i32*> [#uses=1] - %tmp3 = load i32* %tmp2, align 4 ; <i32> [#uses=1] + %tmp3 = load i32, i32* %tmp2, align 4 ; <i32> [#uses=1] %tmp4 = icmp eq i32 %tmp3, 0 ; <i1> [#uses=1] br i1 %tmp4, label %bb.i.bb7.i.backedge_crit_edge, label %bb1.i @@ -81,7 +81,7 @@ bb1.i: ; preds = %bb.i %tmp5 = add i32 %j.0.i, 2 ; <i32> [#uses=1] %tmp6 = sext i32 %tmp5 to i64 ; <i64> [#uses=1] %tmp7 = getelementptr [17 x i32], [17 x i32]* %b, i64 0, i64 %tmp6 ; <i32*> [#uses=1] - %tmp8 = load i32* %tmp7, align 4 ; <i32> [#uses=1] + %tmp8 = load i32, i32* %tmp7, align 4 ; <i32> [#uses=1] %tmp9 = icmp eq i32 %tmp8, 0 ; <i1> [#uses=1] br i1 %tmp9, label %bb1.i.bb7.i.backedge_crit_edge, label %bb2.i @@ -89,7 +89,7 @@ bb2.i: ; preds = %bb1.i %tmp10 = sub i32 7, %j.0.i ; <i32> [#uses=1] %tmp11 = sext i32 %tmp10 to i64 ; <i64> [#uses=1] %tmp12 = getelementptr [15 x i32], [15 x i32]* %c, i64 0, i64 %tmp11 ; <i32*> [#uses=1] - %tmp13 = load i32* %tmp12, align 4 ; <i32> [#uses=1] + %tmp13 = load i32, i32* %tmp12, align 4 ; <i32> [#uses=1] %tmp14 = icmp eq i32 %tmp13, 0 ; <i1> [#uses=1] br i1 %tmp14, label %bb2.i.bb7.i.backedge_crit_edge, label %bb3.i @@ -108,7 +108,7 @@ bb3.i: ; preds = %bb2.i %tmp23 = getelementptr [15 x i32], [15 x i32]* %c, i64 0, i64 %tmp22 ; <i32*> [#uses=1] store i32 0, i32* %tmp23, align 4 call void @Try(i32 2, i32* %q, i32* %b9, i32* %a10, i32* %c11, i32* %x1.sub) nounwind - %tmp24 = load i32* %q, align 4 ; <i32> [#uses=1] + %tmp24 = load i32, i32* %q, align 4 ; <i32> [#uses=1] %tmp25 = icmp eq i32 %tmp24, 0 ; <i1> [#uses=1] br i1 %tmp25, label %bb5.i, label %bb3.i.bb7.i.backedge_crit_edge @@ -131,7 +131,7 @@ bb7.i.backedge: ; preds = %bb3.i.bb7.i.backedge_crit_edge, %bb2.i.bb7.i.backedg bb7.i: ; preds = %bb7.i.backedge, %newFuncRoot %j.0.i = phi i32 [ 0, %newFuncRoot ], [ %tmp, %bb7.i.backedge ] ; <i32> [#uses=8] - %tmp34 = load i32* %q, align 4 ; <i32> [#uses=1] + %tmp34 = load i32, i32* %q, align 4 ; <i32> [#uses=1] %tmp35 = icmp eq i32 %tmp34, 0 ; <i1> [#uses=1] %tmp36 = icmp ne i32 %j.0.i, 8 ; <i1> [#uses=1] %tmp37 = and i1 %tmp35, %tmp36 ; <i1> [#uses=1] diff --git a/llvm/test/Analysis/ScalarEvolution/zext-signed-addrec.ll b/llvm/test/Analysis/ScalarEvolution/zext-signed-addrec.ll index 43698204a72..9201ffca44e 100644 --- a/llvm/test/Analysis/ScalarEvolution/zext-signed-addrec.ll +++ b/llvm/test/Analysis/ScalarEvolution/zext-signed-addrec.ll @@ -15,16 +15,16 @@ target triple = "x86_64-unknown-linux-gnu" ; CHECK-LABEL: foo define i32 @foo() { entry: - %.pr = load i32* @b, align 4 + %.pr = load i32, i32* @b, align 4 %cmp10 = icmp slt i32 %.pr, 1 br i1 %cmp10, label %for.cond1.preheader.lr.ph, label %entry.for.end9_crit_edge entry.for.end9_crit_edge: ; preds = %entry - %.pre = load i32* @c, align 4 + %.pre = load i32, i32* @c, align 4 br label %for.end9 for.cond1.preheader.lr.ph: ; preds = %entry - %0 = load i32* @a, align 4 + %0 = load i32, i32* @a, align 4 %tobool = icmp eq i32 %0, 0 br i1 %tobool, label %for.cond1.preheader.for.cond1.preheader.split_crit_edge, label %return.loopexit.split |