diff options
Diffstat (limited to 'llvm/test/Transforms/LoopVectorize')
113 files changed, 771 insertions, 771 deletions
diff --git a/llvm/test/Transforms/LoopVectorize/12-12-11-if-conv.ll b/llvm/test/Transforms/LoopVectorize/12-12-11-if-conv.ll index aa2e6183781..d5e020c36e9 100644 --- a/llvm/test/Transforms/LoopVectorize/12-12-11-if-conv.ll +++ b/llvm/test/Transforms/LoopVectorize/12-12-11-if-conv.ll @@ -15,7 +15,7 @@ entry: for.body: ; preds = %entry, %if.end %indvars.iv = phi i64 [ %indvars.iv.next, %if.end ], [ 0, %entry ] %arrayidx = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %tobool = icmp eq i32 %0, 0 br i1 %tobool, label %if.end, label %if.then diff --git a/llvm/test/Transforms/LoopVectorize/AArch64/aarch64-unroll.ll b/llvm/test/Transforms/LoopVectorize/AArch64/aarch64-unroll.ll index 83e39a1fb7b..a689f44e912 100644 --- a/llvm/test/Transforms/LoopVectorize/AArch64/aarch64-unroll.ll +++ b/llvm/test/Transforms/LoopVectorize/AArch64/aarch64-unroll.ll @@ -23,9 +23,9 @@ for.body.preheader: ; preds = %entry for.body: ; preds = %for.body.preheader, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %for.body.preheader ] %arrayidx = getelementptr inbounds i32, i32* %a, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds i32, i32* %b, i64 %indvars.iv - %1 = load i32* %arrayidx2, align 4 + %1 = load i32, i32* %arrayidx2, align 4 %add = add nsw i32 %1, %0 %arrayidx4 = getelementptr inbounds i32, i32* %c, i64 %indvars.iv store i32 %add, i32* %arrayidx4, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/AArch64/arbitrary-induction-step.ll b/llvm/test/Transforms/LoopVectorize/AArch64/arbitrary-induction-step.ll index a7a78c75a99..4cd703f6458 100644 --- a/llvm/test/Transforms/LoopVectorize/AArch64/arbitrary-induction-step.ll +++ b/llvm/test/Transforms/LoopVectorize/AArch64/arbitrary-induction-step.ll @@ -11,8 +11,8 @@ target triple = "aarch64--linux-gnueabi" ; } ; CHECK-LABEL: @ind_plus2( -; CHECK: load <4 x i32>* -; CHECK: load <4 x i32>* +; CHECK: load <4 x i32>, <4 x i32>* +; CHECK: load <4 x i32>, <4 x i32>* ; CHECK: mul nsw <4 x i32> ; CHECK: mul nsw <4 x i32> ; CHECK: add nsw <4 x i32> @@ -21,7 +21,7 @@ target triple = "aarch64--linux-gnueabi" ; CHECK: icmp eq i64 %index.next, 512 ; FORCE-VEC-LABEL: @ind_plus2( -; FORCE-VEC: %wide.load = load <2 x i32>* +; FORCE-VEC: %wide.load = load <2 x i32>, <2 x i32>* ; FORCE-VEC: mul nsw <2 x i32> ; FORCE-VEC: add nsw <2 x i32> ; FORCE-VEC: %index.next = add i64 %index, 2 @@ -35,7 +35,7 @@ for.body: ; preds = %entry, %for.body %i = phi i32 [ 0, %entry ], [ %add1, %for.body ] %sum = phi i32 [ 0, %entry ], [ %add, %for.body ] %inc.ptr = getelementptr inbounds i32, i32* %A.addr, i64 1 - %0 = load i32* %A.addr, align 4 + %0 = load i32, i32* %A.addr, align 4 %mul = mul nsw i32 %0, %i %add = add nsw i32 %mul, %sum %add1 = add nsw i32 %i, 2 @@ -55,8 +55,8 @@ for.end: ; preds = %for.body ; } ; CHECK-LABEL: @ind_minus2( -; CHECK: load <4 x i32>* -; CHECK: load <4 x i32>* +; CHECK: load <4 x i32>, <4 x i32>* +; CHECK: load <4 x i32>, <4 x i32>* ; CHECK: mul nsw <4 x i32> ; CHECK: mul nsw <4 x i32> ; CHECK: add nsw <4 x i32> @@ -65,7 +65,7 @@ for.end: ; preds = %for.body ; CHECK: icmp eq i64 %index.next, 512 ; FORCE-VEC-LABEL: @ind_minus2( -; FORCE-VEC: %wide.load = load <2 x i32>* +; FORCE-VEC: %wide.load = load <2 x i32>, <2 x i32>* ; FORCE-VEC: mul nsw <2 x i32> ; FORCE-VEC: add nsw <2 x i32> ; FORCE-VEC: %index.next = add i64 %index, 2 @@ -79,7 +79,7 @@ for.body: ; preds = %entry, %for.body %i = phi i32 [ 1024, %entry ], [ %sub, %for.body ] %sum = phi i32 [ 0, %entry ], [ %add, %for.body ] %inc.ptr = getelementptr inbounds i32, i32* %A.addr, i64 1 - %0 = load i32* %A.addr, align 4 + %0 = load i32, i32* %A.addr, align 4 %mul = mul nsw i32 %0, %i %add = add nsw i32 %mul, %sum %sub = add nsw i32 %i, -2 @@ -102,10 +102,10 @@ for.end: ; preds = %for.body ; } ; CHECK-LABEL: @ptr_ind_plus2( -; CHECK: load i32* -; CHECK: load i32* -; CHECK: load i32* -; CHECK: load i32* +; CHECK: load i32, i32* +; CHECK: load i32, i32* +; CHECK: load i32, i32* +; CHECK: load i32, i32* ; CHECK: mul nsw i32 ; CHECK: mul nsw i32 ; CHECK: add nsw i32 @@ -114,13 +114,13 @@ for.end: ; preds = %for.body ; CHECK: %21 = icmp eq i64 %index.next, 1024 ; FORCE-VEC-LABEL: @ptr_ind_plus2( -; FORCE-VEC: load i32* +; FORCE-VEC: load i32, i32* ; FORCE-VEC: insertelement <2 x i32> -; FORCE-VEC: load i32* +; FORCE-VEC: load i32, i32* ; FORCE-VEC: insertelement <2 x i32> -; FORCE-VEC: load i32* +; FORCE-VEC: load i32, i32* ; FORCE-VEC: insertelement <2 x i32> -; FORCE-VEC: load i32* +; FORCE-VEC: load i32, i32* ; FORCE-VEC: insertelement <2 x i32> ; FORCE-VEC: mul nsw <2 x i32> ; FORCE-VEC: add nsw <2 x i32> @@ -135,9 +135,9 @@ for.body: ; preds = %for.body, %entry %sum = phi i32 [ 0, %entry ], [ %add, %for.body ] %i = phi i32 [ 0, %entry ], [ %inc, %for.body ] %inc.ptr = getelementptr inbounds i32, i32* %A.addr, i64 1 - %0 = load i32* %A.addr, align 4 + %0 = load i32, i32* %A.addr, align 4 %inc.ptr1 = getelementptr inbounds i32, i32* %A.addr, i64 2 - %1 = load i32* %inc.ptr, align 4 + %1 = load i32, i32* %inc.ptr, align 4 %mul = mul nsw i32 %1, %0 %add = add nsw i32 %mul, %sum %inc = add nsw i32 %i, 1 diff --git a/llvm/test/Transforms/LoopVectorize/AArch64/arm64-unroll.ll b/llvm/test/Transforms/LoopVectorize/AArch64/arm64-unroll.ll index 159aaf488da..395b468c509 100644 --- a/llvm/test/Transforms/LoopVectorize/AArch64/arm64-unroll.ll +++ b/llvm/test/Transforms/LoopVectorize/AArch64/arm64-unroll.ll @@ -23,9 +23,9 @@ for.body.preheader: ; preds = %entry for.body: ; preds = %for.body.preheader, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %for.body.preheader ] %arrayidx = getelementptr inbounds i32, i32* %a, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds i32, i32* %b, i64 %indvars.iv - %1 = load i32* %arrayidx2, align 4 + %1 = load i32, i32* %arrayidx2, align 4 %add = add nsw i32 %1, %0 %arrayidx4 = getelementptr inbounds i32, i32* %c, i64 %indvars.iv store i32 %add, i32* %arrayidx4, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/AArch64/gather-cost.ll b/llvm/test/Transforms/LoopVectorize/AArch64/gather-cost.ll index 302ec796734..46b8ef1a09b 100644 --- a/llvm/test/Transforms/LoopVectorize/AArch64/gather-cost.ll +++ b/llvm/test/Transforms/LoopVectorize/AArch64/gather-cost.ll @@ -31,23 +31,23 @@ for.body: %add = add i64 %v.055, %offset %mul = mul i64 %add, 3 %arrayidx = getelementptr inbounds [1536 x float], [1536 x float]* @src_data, i64 0, i64 %mul - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds [512 x float], [512 x float]* @kernel, i64 0, i64 %v.055 - %1 = load float* %arrayidx2, align 4 + %1 = load float, float* %arrayidx2, align 4 %mul3 = fmul fast float %0, %1 %arrayidx4 = getelementptr inbounds [512 x float], [512 x float]* @kernel2, i64 0, i64 %v.055 - %2 = load float* %arrayidx4, align 4 + %2 = load float, float* %arrayidx4, align 4 %mul5 = fmul fast float %mul3, %2 %arrayidx6 = getelementptr inbounds [512 x float], [512 x float]* @kernel3, i64 0, i64 %v.055 - %3 = load float* %arrayidx6, align 4 + %3 = load float, float* %arrayidx6, align 4 %mul7 = fmul fast float %mul5, %3 %arrayidx8 = getelementptr inbounds [512 x float], [512 x float]* @kernel4, i64 0, i64 %v.055 - %4 = load float* %arrayidx8, align 4 + %4 = load float, float* %arrayidx8, align 4 %mul9 = fmul fast float %mul7, %4 %add10 = fadd fast float %r.057, %mul9 %arrayidx.sum = add i64 %mul, 1 %arrayidx11 = getelementptr inbounds [1536 x float], [1536 x float]* @src_data, i64 0, i64 %arrayidx.sum - %5 = load float* %arrayidx11, align 4 + %5 = load float, float* %arrayidx11, align 4 %mul13 = fmul fast float %1, %5 %mul15 = fmul fast float %2, %mul13 %mul17 = fmul fast float %3, %mul15 @@ -55,7 +55,7 @@ for.body: %add20 = fadd fast float %g.056, %mul19 %arrayidx.sum52 = add i64 %mul, 2 %arrayidx21 = getelementptr inbounds [1536 x float], [1536 x float]* @src_data, i64 0, i64 %arrayidx.sum52 - %6 = load float* %arrayidx21, align 4 + %6 = load float, float* %arrayidx21, align 4 %mul23 = fmul fast float %1, %6 %mul25 = fmul fast float %2, %mul23 %mul27 = fmul fast float %3, %mul25 diff --git a/llvm/test/Transforms/LoopVectorize/AArch64/sdiv-pow2.ll b/llvm/test/Transforms/LoopVectorize/AArch64/sdiv-pow2.ll index d1ca199b82c..f3c6548c68e 100644 --- a/llvm/test/Transforms/LoopVectorize/AArch64/sdiv-pow2.ll +++ b/llvm/test/Transforms/LoopVectorize/AArch64/sdiv-pow2.ll @@ -7,7 +7,7 @@ target triple = "aarch64--linux-gnu" @Foo = common global %struct.anon zeroinitializer, align 4 ; CHECK-LABEL: @foo( -; CHECK: load <4 x i32>* +; CHECK: load <4 x i32>, <4 x i32>* ; CHECK: sdiv <4 x i32> ; CHECK: store <4 x i32> @@ -18,7 +18,7 @@ entry: for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds %struct.anon, %struct.anon* @Foo, i64 0, i32 2, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %div = sdiv i32 %0, 2 %arrayidx2 = getelementptr inbounds %struct.anon, %struct.anon* @Foo, i64 0, i32 0, i64 %indvars.iv store i32 %div, i32* %arrayidx2, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/ARM/arm-unroll.ll b/llvm/test/Transforms/LoopVectorize/ARM/arm-unroll.ll index 96e13842106..7b09913636f 100644 --- a/llvm/test/Transforms/LoopVectorize/ARM/arm-unroll.ll +++ b/llvm/test/Transforms/LoopVectorize/ARM/arm-unroll.ll @@ -21,7 +21,7 @@ define i32 @foo(i32* nocapture %A, i32 %n) nounwind readonly ssp { %i.02 = phi i32 [ %5, %.lr.ph ], [ 0, %0 ] %sum.01 = phi i32 [ %4, %.lr.ph ], [ 0, %0 ] %2 = getelementptr inbounds i32, i32* %A, i32 %i.02 - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = add nsw i32 %3, %sum.01 %5 = add nsw i32 %i.02, 1 %exitcond = icmp eq i32 %5, %n @@ -49,7 +49,7 @@ define i32 @register_limit(i32* nocapture %A, i32 %n) { %sum.05 = phi i32 [ %9, %.lr.ph ], [ 0, %0 ] %sum.06 = phi i32 [ %10, %.lr.ph ], [ 0, %0 ] %2 = getelementptr inbounds i32, i32* %A, i32 %i.02 - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = add nsw i32 %3, %sum.01 %5 = add nsw i32 %i.02, 1 %6 = add nsw i32 %3, %sum.02 diff --git a/llvm/test/Transforms/LoopVectorize/ARM/gather-cost.ll b/llvm/test/Transforms/LoopVectorize/ARM/gather-cost.ll index 1b5e45e089f..f14a8cc6ca2 100644 --- a/llvm/test/Transforms/LoopVectorize/ARM/gather-cost.ll +++ b/llvm/test/Transforms/LoopVectorize/ARM/gather-cost.ll @@ -34,23 +34,23 @@ for.body: %add = add i32 %v.055, %offset %mul = mul i32 %add, 3 %arrayidx = getelementptr inbounds [1536 x float], [1536 x float]* @src_data, i32 0, i32 %mul - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds [512 x float], [512 x float]* @kernel, i32 0, i32 %v.055 - %1 = load float* %arrayidx2, align 4 + %1 = load float, float* %arrayidx2, align 4 %mul3 = fmul fast float %0, %1 %arrayidx4 = getelementptr inbounds [512 x float], [512 x float]* @kernel2, i32 0, i32 %v.055 - %2 = load float* %arrayidx4, align 4 + %2 = load float, float* %arrayidx4, align 4 %mul5 = fmul fast float %mul3, %2 %arrayidx6 = getelementptr inbounds [512 x float], [512 x float]* @kernel3, i32 0, i32 %v.055 - %3 = load float* %arrayidx6, align 4 + %3 = load float, float* %arrayidx6, align 4 %mul7 = fmul fast float %mul5, %3 %arrayidx8 = getelementptr inbounds [512 x float], [512 x float]* @kernel4, i32 0, i32 %v.055 - %4 = load float* %arrayidx8, align 4 + %4 = load float, float* %arrayidx8, align 4 %mul9 = fmul fast float %mul7, %4 %add10 = fadd fast float %r.057, %mul9 %arrayidx.sum = add i32 %mul, 1 %arrayidx11 = getelementptr inbounds [1536 x float], [1536 x float]* @src_data, i32 0, i32 %arrayidx.sum - %5 = load float* %arrayidx11, align 4 + %5 = load float, float* %arrayidx11, align 4 %mul13 = fmul fast float %1, %5 %mul15 = fmul fast float %2, %mul13 %mul17 = fmul fast float %3, %mul15 @@ -58,7 +58,7 @@ for.body: %add20 = fadd fast float %g.056, %mul19 %arrayidx.sum52 = add i32 %mul, 2 %arrayidx21 = getelementptr inbounds [1536 x float], [1536 x float]* @src_data, i32 0, i32 %arrayidx.sum52 - %6 = load float* %arrayidx21, align 4 + %6 = load float, float* %arrayidx21, align 4 %mul23 = fmul fast float %1, %6 %mul25 = fmul fast float %2, %mul23 %mul27 = fmul fast float %3, %mul25 diff --git a/llvm/test/Transforms/LoopVectorize/ARM/gcc-examples.ll b/llvm/test/Transforms/LoopVectorize/ARM/gcc-examples.ll index ae61da02220..783156d7399 100644 --- a/llvm/test/Transforms/LoopVectorize/ARM/gcc-examples.ll +++ b/llvm/test/Transforms/LoopVectorize/ARM/gcc-examples.ll @@ -19,9 +19,9 @@ define void @example1() nounwind uwtable ssp { ; <label>:1 ; preds = %1, %0 %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %2 = getelementptr inbounds [2048 x i32], [2048 x i32]* @b, i64 0, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = getelementptr inbounds [2048 x i32], [2048 x i32]* @c, i64 0, i64 %indvars.iv - %5 = load i32* %4, align 4 + %5 = load i32, i32* %4, align 4 %6 = add nsw i32 %5, %3 %7 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv store i32 %6, i32* %7, align 4 @@ -45,7 +45,7 @@ define void @example10b(i16* noalias nocapture %sa, i16* noalias nocapture %sb, ; <label>:1 ; preds = %1, %0 %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %2 = getelementptr inbounds i16, i16* %sb, i64 %indvars.iv - %3 = load i16* %2, align 2 + %3 = load i16, i16* %2, align 2 %4 = sext i16 %3 to i32 %5 = getelementptr inbounds i32, i32* %ia, i64 %indvars.iv store i32 %4, i32* %5, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/ARM/mul-cast-vect.ll b/llvm/test/Transforms/LoopVectorize/ARM/mul-cast-vect.ll index d2e3de279f7..e88fcca1225 100644 --- a/llvm/test/Transforms/LoopVectorize/ARM/mul-cast-vect.ll +++ b/llvm/test/Transforms/LoopVectorize/ARM/mul-cast-vect.ll @@ -16,9 +16,9 @@ target triple = "armv7--linux-gnueabihf" define void @direct(%T432* %loadaddr, %T432* %loadaddr2, %T432* %storeaddr) { ; COST: function 'direct': - %v0 = load %T432* %loadaddr + %v0 = load %T432, %T432* %loadaddr ; ASM: vld1.64 - %v1 = load %T432* %loadaddr2 + %v1 = load %T432, %T432* %loadaddr2 ; ASM: vld1.64 %r3 = mul %T432 %v0, %v1 ; COST: cost of 2 for instruction: {{.*}} mul <4 x i32> @@ -30,9 +30,9 @@ define void @direct(%T432* %loadaddr, %T432* %loadaddr2, %T432* %storeaddr) { define void @ups1632(%T416* %loadaddr, %T416* %loadaddr2, %T432* %storeaddr) { ; COST: function 'ups1632': - %v0 = load %T416* %loadaddr + %v0 = load %T416, %T416* %loadaddr ; ASM: vldr - %v1 = load %T416* %loadaddr2 + %v1 = load %T416, %T416* %loadaddr2 ; ASM: vldr %r1 = sext %T416 %v0 to %T432 %r2 = sext %T416 %v1 to %T432 @@ -47,9 +47,9 @@ define void @ups1632(%T416* %loadaddr, %T416* %loadaddr2, %T432* %storeaddr) { define void @upu1632(%T416* %loadaddr, %T416* %loadaddr2, %T432* %storeaddr) { ; COST: function 'upu1632': - %v0 = load %T416* %loadaddr + %v0 = load %T416, %T416* %loadaddr ; ASM: vldr - %v1 = load %T416* %loadaddr2 + %v1 = load %T416, %T416* %loadaddr2 ; ASM: vldr %r1 = zext %T416 %v0 to %T432 %r2 = zext %T416 %v1 to %T432 @@ -64,9 +64,9 @@ define void @upu1632(%T416* %loadaddr, %T416* %loadaddr2, %T432* %storeaddr) { define void @ups3264(%T232* %loadaddr, %T232* %loadaddr2, %T264* %storeaddr) { ; COST: function 'ups3264': - %v0 = load %T232* %loadaddr + %v0 = load %T232, %T232* %loadaddr ; ASM: vldr - %v1 = load %T232* %loadaddr2 + %v1 = load %T232, %T232* %loadaddr2 ; ASM: vldr %r3 = mul %T232 %v0, %v1 ; ASM: vmul.i32 @@ -81,9 +81,9 @@ define void @ups3264(%T232* %loadaddr, %T232* %loadaddr2, %T264* %storeaddr) { define void @upu3264(%T232* %loadaddr, %T232* %loadaddr2, %T264* %storeaddr) { ; COST: function 'upu3264': - %v0 = load %T232* %loadaddr + %v0 = load %T232, %T232* %loadaddr ; ASM: vldr - %v1 = load %T232* %loadaddr2 + %v1 = load %T232, %T232* %loadaddr2 ; ASM: vldr %r3 = mul %T232 %v0, %v1 ; ASM: vmul.i32 @@ -98,9 +98,9 @@ define void @upu3264(%T232* %loadaddr, %T232* %loadaddr2, %T264* %storeaddr) { define void @dn3216(%T432* %loadaddr, %T432* %loadaddr2, %T416* %storeaddr) { ; COST: function 'dn3216': - %v0 = load %T432* %loadaddr + %v0 = load %T432, %T432* %loadaddr ; ASM: vld1.64 - %v1 = load %T432* %loadaddr2 + %v1 = load %T432, %T432* %loadaddr2 ; ASM: vld1.64 %r3 = mul %T432 %v0, %v1 ; ASM: vmul.i32 diff --git a/llvm/test/Transforms/LoopVectorize/ARM/width-detect.ll b/llvm/test/Transforms/LoopVectorize/ARM/width-detect.ll index f970e922810..66d2556dfb8 100644 --- a/llvm/test/Transforms/LoopVectorize/ARM/width-detect.ll +++ b/llvm/test/Transforms/LoopVectorize/ARM/width-detect.ll @@ -14,7 +14,7 @@ define float @foo_F32(float* nocapture %A, i32 %n) nounwind uwtable readonly ssp %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 0, %0 ] %prod.01 = phi float [ %4, %.lr.ph ], [ 0.000000e+00, %0 ] %2 = getelementptr inbounds float, float* %A, i64 %indvars.iv - %3 = load float* %2, align 8 + %3 = load float, float* %2, align 8 %4 = fmul fast float %prod.01, %3 %indvars.iv.next = add i64 %indvars.iv, 1 %lftr.wideiv = trunc i64 %indvars.iv.next to i32 @@ -37,7 +37,7 @@ define signext i8 @foo_I8(i8* nocapture %A, i32 %n) nounwind uwtable readonly ss %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 0, %0 ] %red.01 = phi i8 [ %4, %.lr.ph ], [ 0, %0 ] %2 = getelementptr inbounds i8, i8* %A, i64 %indvars.iv - %3 = load i8* %2, align 1 + %3 = load i8, i8* %2, align 1 %4 = xor i8 %3, %red.01 %indvars.iv.next = add i64 %indvars.iv, 1 %lftr.wideiv = trunc i64 %indvars.iv.next to i32 diff --git a/llvm/test/Transforms/LoopVectorize/PowerPC/small-loop-rdx.ll b/llvm/test/Transforms/LoopVectorize/PowerPC/small-loop-rdx.ll index bc043475a64..2898af2986d 100644 --- a/llvm/test/Transforms/LoopVectorize/PowerPC/small-loop-rdx.ll +++ b/llvm/test/Transforms/LoopVectorize/PowerPC/small-loop-rdx.ll @@ -30,7 +30,7 @@ for.body: ; preds = %for.body, %for.body %indvars.iv = phi i64 [ 0, %for.body.lr.ph ], [ %indvars.iv.next, %for.body ] %redx.05 = phi double [ 0.000000e+00, %for.body.lr.ph ], [ %add, %for.body ] %arrayidx = getelementptr inbounds double, double* %arr, i64 %indvars.iv - %1 = load double* %arrayidx, align 8 + %1 = load double, double* %arrayidx, align 8 %add = fadd fast double %1, %redx.05 %indvars.iv.next = add i64 %indvars.iv, 1 %lftr.wideiv = trunc i64 %indvars.iv to i32 diff --git a/llvm/test/Transforms/LoopVectorize/PowerPC/vsx-tsvc-s173.ll b/llvm/test/Transforms/LoopVectorize/PowerPC/vsx-tsvc-s173.ll index 27a11028b5b..65b3919585e 100644 --- a/llvm/test/Transforms/LoopVectorize/PowerPC/vsx-tsvc-s173.ll +++ b/llvm/test/Transforms/LoopVectorize/PowerPC/vsx-tsvc-s173.ll @@ -9,7 +9,7 @@ target triple = "powerpc64-unknown-linux-gnu" define signext i32 @s173() #0 { entry: - %0 = load i32* @ntimes, align 4 + %0 = load i32, i32* @ntimes, align 4 %cmp21 = icmp sgt i32 %0, 0 br i1 %cmp21, label %for.cond1.preheader, label %for.end12 @@ -20,9 +20,9 @@ for.cond1.preheader: ; preds = %for.end, %entry for.body3: ; preds = %for.body3, %for.cond1.preheader %indvars.iv = phi i64 [ 0, %for.cond1.preheader ], [ %indvars.iv.next, %for.body3 ] %arrayidx = getelementptr inbounds %struct.GlobalData, %struct.GlobalData* @global_data, i64 0, i32 0, i64 %indvars.iv - %1 = load float* %arrayidx, align 4 + %1 = load float, float* %arrayidx, align 4 %arrayidx5 = getelementptr inbounds %struct.GlobalData, %struct.GlobalData* @global_data, i64 0, i32 3, i64 %indvars.iv - %2 = load float* %arrayidx5, align 4 + %2 = load float, float* %arrayidx5, align 4 %add = fadd float %1, %2 %3 = add nsw i64 %indvars.iv, 16000 %arrayidx8 = getelementptr inbounds %struct.GlobalData, %struct.GlobalData* @global_data, i64 0, i32 0, i64 %3 @@ -33,7 +33,7 @@ for.body3: ; preds = %for.body3, %for.con for.end: ; preds = %for.body3 %inc11 = add nsw i32 %nl.022, 1 - %4 = load i32* @ntimes, align 4 + %4 = load i32, i32* @ntimes, align 4 %mul = mul nsw i32 %4, 10 %cmp = icmp slt i32 %inc11, %mul br i1 %cmp, label %for.cond1.preheader, label %for.end12 @@ -42,7 +42,7 @@ for.end12: ; preds = %for.end, %entry ret i32 0 ; CHECK-LABEL: @s173 -; CHECK: load <4 x float>* +; CHECK: load <4 x float>, <4 x float>* ; CHECK: add i64 %index, 16000 ; CHECK: ret i32 0 } diff --git a/llvm/test/Transforms/LoopVectorize/X86/already-vectorized.ll b/llvm/test/Transforms/LoopVectorize/X86/already-vectorized.ll index dcf2c6e70c4..248d6dc90bd 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/already-vectorized.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/already-vectorized.ll @@ -22,7 +22,7 @@ for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %red.05 = phi i32 [ 0, %entry ], [ %add, %for.body ] %arrayidx = getelementptr inbounds [255 x i32], [255 x i32]* @a, i64 0, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %add = add nsw i32 %0, %red.05 %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1 %exitcond = icmp eq i64 %indvars.iv.next, 255 diff --git a/llvm/test/Transforms/LoopVectorize/X86/assume.ll b/llvm/test/Transforms/LoopVectorize/X86/assume.ll index c036bba07ce..4fd378d1a0a 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/assume.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/assume.ll @@ -23,7 +23,7 @@ entry: for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds float, float* %b, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp1 = fcmp ogt float %0, 1.000000e+02 tail call void @llvm.assume(i1 %cmp1) %add = fadd float %0, 1.000000e+00 @@ -49,12 +49,12 @@ attributes #1 = { nounwind } define void @test2(%struct.data* nocapture readonly %d) #0 { entry: %b = getelementptr inbounds %struct.data, %struct.data* %d, i64 0, i32 1 - %0 = load float** %b, align 8 + %0 = load float*, float** %b, align 8 %ptrint = ptrtoint float* %0 to i64 %maskedptr = and i64 %ptrint, 31 %maskcond = icmp eq i64 %maskedptr, 0 %a = getelementptr inbounds %struct.data, %struct.data* %d, i64 0, i32 0 - %1 = load float** %a, align 8 + %1 = load float*, float** %a, align 8 %ptrint2 = ptrtoint float* %1 to i64 %maskedptr3 = and i64 %ptrint2, 31 %maskcond4 = icmp eq i64 %maskedptr3, 0 @@ -85,7 +85,7 @@ for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] tail call void @llvm.assume(i1 %maskcond) %arrayidx = getelementptr inbounds float, float* %0, i64 %indvars.iv - %2 = load float* %arrayidx, align 4 + %2 = load float, float* %arrayidx, align 4 %add = fadd float %2, 1.000000e+00 tail call void @llvm.assume(i1 %maskcond4) %arrayidx5 = getelementptr inbounds float, float* %1, i64 %indvars.iv diff --git a/llvm/test/Transforms/LoopVectorize/X86/avx1.ll b/llvm/test/Transforms/LoopVectorize/X86/avx1.ll index 95088df5c3f..37977c43ac3 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/avx1.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/avx1.ll @@ -13,7 +13,7 @@ define i32 @read_mod_write_single_ptr(float* nocapture %a, i32 %n) nounwind uwta .lr.ph: ; preds = %0, %.lr.ph %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 0, %0 ] %2 = getelementptr inbounds float, float* %a, i64 %indvars.iv - %3 = load float* %2, align 4 + %3 = load float, float* %2, align 4 %4 = fmul float %3, 3.000000e+00 store float %4, float* %2, align 4 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -36,7 +36,7 @@ define i32 @read_mod_i64(i64* nocapture %a, i32 %n) nounwind uwtable ssp { .lr.ph: ; preds = %0, %.lr.ph %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 0, %0 ] %2 = getelementptr inbounds i64, i64* %a, i64 %indvars.iv - %3 = load i64* %2, align 4 + %3 = load i64, i64* %2, align 4 %4 = add i64 %3, 3 store i64 %4, i64* %2, align 4 %indvars.iv.next = add i64 %indvars.iv, 1 diff --git a/llvm/test/Transforms/LoopVectorize/X86/constant-vector-operand.ll b/llvm/test/Transforms/LoopVectorize/X86/constant-vector-operand.ll index 106bd84fc7c..d75b1d940ae 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/constant-vector-operand.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/constant-vector-operand.ll @@ -16,7 +16,7 @@ entry: for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32]* @B, i64 0, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %shl = ashr i32 %0, 3 %arrayidx2 = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 0, i64 %indvars.iv store i32 %shl, i32* %arrayidx2, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/X86/cost-model.ll b/llvm/test/Transforms/LoopVectorize/X86/cost-model.ll index 190e130b609..013657102e6 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/cost-model.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/cost-model.ll @@ -22,12 +22,12 @@ for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %0 = shl nsw i64 %indvars.iv, 1 %arrayidx = getelementptr inbounds [2048 x i32], [2048 x i32]* @c, i64 0, i64 %0 - %1 = load i32* %arrayidx, align 8 + %1 = load i32, i32* %arrayidx, align 8 %idxprom1 = sext i32 %1 to i64 %arrayidx2 = getelementptr inbounds [2048 x i32], [2048 x i32]* @b, i64 0, i64 %idxprom1 - %2 = load i32* %arrayidx2, align 4 + %2 = load i32, i32* %arrayidx2, align 4 %arrayidx4 = getelementptr inbounds [2048 x i32], [2048 x i32]* @d, i64 0, i64 %indvars.iv - %3 = load i32* %arrayidx4, align 4 + %3 = load i32, i32* %arrayidx4, align 4 %idxprom5 = sext i32 %3 to i64 %arrayidx6 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %idxprom5 store i32 %2, i32* %arrayidx6, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/X86/fp32_to_uint32-cost-model.ll b/llvm/test/Transforms/LoopVectorize/X86/fp32_to_uint32-cost-model.ll index 627ae0076b0..4a56d6b5ebf 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/fp32_to_uint32-cost-model.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/fp32_to_uint32-cost-model.ll @@ -21,7 +21,7 @@ for.body.preheader: ; preds = %entry for.body: ; preds = %for.body.preheader, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %for.body.preheader ] %arrayidx = getelementptr inbounds [10000 x float], [10000 x float]* @float_array, i64 0, i64 %indvars.iv - %1 = load float* %arrayidx, align 4 + %1 = load float, float* %arrayidx, align 4 %conv = fptoui float %1 to i32 %arrayidx2 = getelementptr inbounds [10000 x i32], [10000 x i32]* @unsigned_array, i64 0, i64 %indvars.iv store i32 %conv, i32* %arrayidx2, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/X86/fp64_to_uint32-cost-model.ll b/llvm/test/Transforms/LoopVectorize/X86/fp64_to_uint32-cost-model.ll index 7c1dfe39236..c066afcfa63 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/fp64_to_uint32-cost-model.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/fp64_to_uint32-cost-model.ll @@ -13,7 +13,7 @@ target triple = "x86_64-apple-macosx" define void @convert() { entry: - %0 = load i32* @n, align 4 + %0 = load i32, i32* @n, align 4 %cmp4 = icmp eq i32 %0, 0 br i1 %cmp4, label %for.end, label %for.body.preheader @@ -23,7 +23,7 @@ for.body.preheader: ; preds = %entry for.body: ; preds = %for.body.preheader, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %for.body.preheader ] %arrayidx = getelementptr inbounds [10000 x double], [10000 x double]* @double_array, i64 0, i64 %indvars.iv - %1 = load double* %arrayidx, align 8 + %1 = load double, double* %arrayidx, align 8 %conv = fptoui double %1 to i32 %arrayidx2 = getelementptr inbounds [10000 x i32], [10000 x i32]* @unsigned_array, i64 0, i64 %indvars.iv store i32 %conv, i32* %arrayidx2, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/X86/fp_to_sint8-cost-model.ll b/llvm/test/Transforms/LoopVectorize/X86/fp_to_sint8-cost-model.ll index 106c9d66623..b3a0710545d 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/fp_to_sint8-cost-model.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/fp_to_sint8-cost-model.ll @@ -12,7 +12,7 @@ entry: for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds float, float* %b, i64 %indvars.iv - %tmp = load float* %arrayidx, align 4 + %tmp = load float, float* %arrayidx, align 4 %conv = fptosi float %tmp to i8 %arrayidx2 = getelementptr inbounds i8, i8* %a, i64 %indvars.iv store i8 %conv, i8* %arrayidx2, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/X86/gather-cost.ll b/llvm/test/Transforms/LoopVectorize/X86/gather-cost.ll index 8c9cb653da0..f0e6c8f205c 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/gather-cost.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/gather-cost.ll @@ -32,23 +32,23 @@ for.body: %add = add i64 %v.055, %offset %mul = mul i64 %add, 3 %arrayidx = getelementptr inbounds [1536 x float], [1536 x float]* @src_data, i64 0, i64 %mul - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds [512 x float], [512 x float]* @kernel, i64 0, i64 %v.055 - %1 = load float* %arrayidx2, align 4 + %1 = load float, float* %arrayidx2, align 4 %mul3 = fmul fast float %0, %1 %arrayidx4 = getelementptr inbounds [512 x float], [512 x float]* @kernel2, i64 0, i64 %v.055 - %2 = load float* %arrayidx4, align 4 + %2 = load float, float* %arrayidx4, align 4 %mul5 = fmul fast float %mul3, %2 %arrayidx6 = getelementptr inbounds [512 x float], [512 x float]* @kernel3, i64 0, i64 %v.055 - %3 = load float* %arrayidx6, align 4 + %3 = load float, float* %arrayidx6, align 4 %mul7 = fmul fast float %mul5, %3 %arrayidx8 = getelementptr inbounds [512 x float], [512 x float]* @kernel4, i64 0, i64 %v.055 - %4 = load float* %arrayidx8, align 4 + %4 = load float, float* %arrayidx8, align 4 %mul9 = fmul fast float %mul7, %4 %add10 = fadd fast float %r.057, %mul9 %arrayidx.sum = add i64 %mul, 1 %arrayidx11 = getelementptr inbounds [1536 x float], [1536 x float]* @src_data, i64 0, i64 %arrayidx.sum - %5 = load float* %arrayidx11, align 4 + %5 = load float, float* %arrayidx11, align 4 %mul13 = fmul fast float %1, %5 %mul15 = fmul fast float %2, %mul13 %mul17 = fmul fast float %3, %mul15 @@ -56,7 +56,7 @@ for.body: %add20 = fadd fast float %g.056, %mul19 %arrayidx.sum52 = add i64 %mul, 2 %arrayidx21 = getelementptr inbounds [1536 x float], [1536 x float]* @src_data, i64 0, i64 %arrayidx.sum52 - %6 = load float* %arrayidx21, align 4 + %6 = load float, float* %arrayidx21, align 4 %mul23 = fmul fast float %1, %6 %mul25 = fmul fast float %2, %mul23 %mul27 = fmul fast float %3, %mul25 diff --git a/llvm/test/Transforms/LoopVectorize/X86/gcc-examples.ll b/llvm/test/Transforms/LoopVectorize/X86/gcc-examples.ll index eb846622262..c581f4bf2a6 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/gcc-examples.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/gcc-examples.ll @@ -29,9 +29,9 @@ define void @example1() nounwind uwtable ssp { ; <label>:1 ; preds = %1, %0 %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %2 = getelementptr inbounds [2048 x i32], [2048 x i32]* @b, i64 0, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = getelementptr inbounds [2048 x i32], [2048 x i32]* @c, i64 0, i64 %indvars.iv - %5 = load i32* %4, align 4 + %5 = load i32, i32* %4, align 4 %6 = add nsw i32 %5, %3 %7 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv store i32 %6, i32* %7, align 4 @@ -62,7 +62,7 @@ define void @example10b(i16* noalias nocapture %sa, i16* noalias nocapture %sb, ; <label>:1 ; preds = %1, %0 %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %2 = getelementptr inbounds i16, i16* %sb, i64 %indvars.iv - %3 = load i16* %2, align 2 + %3 = load i16, i16* %2, align 2 %4 = sext i16 %3 to i32 %5 = getelementptr inbounds i32, i32* %ia, i64 %indvars.iv store i32 %4, i32* %5, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/X86/illegal-parallel-loop-uniform-write.ll b/llvm/test/Transforms/LoopVectorize/X86/illegal-parallel-loop-uniform-write.ll index f2163b0bf81..cbba5300b9c 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/illegal-parallel-loop-uniform-write.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/illegal-parallel-loop-uniform-write.ll @@ -15,7 +15,7 @@ entry: for.end.us: ; preds = %for.body3.us %arrayidx9.us = getelementptr inbounds i32, i32* %b, i64 %indvars.iv33 - %0 = load i32* %arrayidx9.us, align 4, !llvm.mem.parallel_loop_access !3 + %0 = load i32, i32* %arrayidx9.us, align 4, !llvm.mem.parallel_loop_access !3 %add10.us = add nsw i32 %0, 3 store i32 %add10.us, i32* %arrayidx9.us, align 4, !llvm.mem.parallel_loop_access !3 %indvars.iv.next34 = add i64 %indvars.iv33, 1 @@ -29,7 +29,7 @@ for.body3.us: ; preds = %for.body3.us, %for. %add4.us = add i32 %add.us, %1 %idxprom.us = sext i32 %add4.us to i64 %arrayidx.us = getelementptr inbounds i32, i32* %a, i64 %idxprom.us - %2 = load i32* %arrayidx.us, align 4, !llvm.mem.parallel_loop_access !3 + %2 = load i32, i32* %arrayidx.us, align 4, !llvm.mem.parallel_loop_access !3 %add5.us = add nsw i32 %2, 1 store i32 %add5.us, i32* %arrayidx7.us, align 4, !llvm.mem.parallel_loop_access !3 %indvars.iv.next30 = add i64 %indvars.iv29, 1 diff --git a/llvm/test/Transforms/LoopVectorize/X86/masked_load_store.ll b/llvm/test/Transforms/LoopVectorize/X86/masked_load_store.ll index c3ee6f81a79..c3175b02036 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/masked_load_store.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/masked_load_store.ll @@ -46,34 +46,34 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 10000 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %idxprom = sext i32 %1 to i64 - %2 = load i32** %trigger.addr, align 8 + %2 = load i32*, i32** %trigger.addr, align 8 %arrayidx = getelementptr inbounds i32, i32* %2, i64 %idxprom - %3 = load i32* %arrayidx, align 4 + %3 = load i32, i32* %arrayidx, align 4 %cmp1 = icmp slt i32 %3, 100 br i1 %cmp1, label %if.then, label %if.end if.then: ; preds = %for.body - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %idxprom2 = sext i32 %4 to i64 - %5 = load i32** %B.addr, align 8 + %5 = load i32*, i32** %B.addr, align 8 %arrayidx3 = getelementptr inbounds i32, i32* %5, i64 %idxprom2 - %6 = load i32* %arrayidx3, align 4 - %7 = load i32* %i, align 4 + %6 = load i32, i32* %arrayidx3, align 4 + %7 = load i32, i32* %i, align 4 %idxprom4 = sext i32 %7 to i64 - %8 = load i32** %trigger.addr, align 8 + %8 = load i32*, i32** %trigger.addr, align 8 %arrayidx5 = getelementptr inbounds i32, i32* %8, i64 %idxprom4 - %9 = load i32* %arrayidx5, align 4 + %9 = load i32, i32* %arrayidx5, align 4 %add = add nsw i32 %6, %9 - %10 = load i32* %i, align 4 + %10 = load i32, i32* %i, align 4 %idxprom6 = sext i32 %10 to i64 - %11 = load i32** %A.addr, align 8 + %11 = load i32*, i32** %A.addr, align 8 %arrayidx7 = getelementptr inbounds i32, i32* %11, i64 %idxprom6 store i32 %add, i32* %arrayidx7, align 4 br label %if.end @@ -82,7 +82,7 @@ if.end: ; preds = %if.then, %for.body br label %for.inc for.inc: ; preds = %if.end - %12 = load i32* %i, align 4 + %12 = load i32, i32* %i, align 4 %inc = add nsw i32 %12, 1 store i32 %inc, i32* %i, align 4 br label %for.cond @@ -130,35 +130,35 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 10000 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %idxprom = sext i32 %1 to i64 - %2 = load i32** %trigger.addr, align 8 + %2 = load i32*, i32** %trigger.addr, align 8 %arrayidx = getelementptr inbounds i32, i32* %2, i64 %idxprom - %3 = load i32* %arrayidx, align 4 + %3 = load i32, i32* %arrayidx, align 4 %cmp1 = icmp slt i32 %3, 100 br i1 %cmp1, label %if.then, label %if.end if.then: ; preds = %for.body - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %idxprom2 = sext i32 %4 to i64 - %5 = load float** %B.addr, align 8 + %5 = load float*, float** %B.addr, align 8 %arrayidx3 = getelementptr inbounds float, float* %5, i64 %idxprom2 - %6 = load float* %arrayidx3, align 4 - %7 = load i32* %i, align 4 + %6 = load float, float* %arrayidx3, align 4 + %7 = load i32, i32* %i, align 4 %idxprom4 = sext i32 %7 to i64 - %8 = load i32** %trigger.addr, align 8 + %8 = load i32*, i32** %trigger.addr, align 8 %arrayidx5 = getelementptr inbounds i32, i32* %8, i64 %idxprom4 - %9 = load i32* %arrayidx5, align 4 + %9 = load i32, i32* %arrayidx5, align 4 %conv = sitofp i32 %9 to float %add = fadd float %6, %conv - %10 = load i32* %i, align 4 + %10 = load i32, i32* %i, align 4 %idxprom6 = sext i32 %10 to i64 - %11 = load float** %A.addr, align 8 + %11 = load float*, float** %A.addr, align 8 %arrayidx7 = getelementptr inbounds float, float* %11, i64 %idxprom6 store float %add, float* %arrayidx7, align 4 br label %if.end @@ -167,7 +167,7 @@ if.end: ; preds = %if.then, %for.body br label %for.inc for.inc: ; preds = %if.end - %12 = load i32* %i, align 4 + %12 = load i32, i32* %i, align 4 %inc = add nsw i32 %12, 1 store i32 %inc, i32* %i, align 4 br label %for.cond @@ -218,35 +218,35 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 10000 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %idxprom = sext i32 %1 to i64 - %2 = load i32** %trigger.addr, align 8 + %2 = load i32*, i32** %trigger.addr, align 8 %arrayidx = getelementptr inbounds i32, i32* %2, i64 %idxprom - %3 = load i32* %arrayidx, align 4 + %3 = load i32, i32* %arrayidx, align 4 %cmp1 = icmp slt i32 %3, 100 br i1 %cmp1, label %if.then, label %if.end if.then: ; preds = %for.body - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %idxprom2 = sext i32 %4 to i64 - %5 = load double** %B.addr, align 8 + %5 = load double*, double** %B.addr, align 8 %arrayidx3 = getelementptr inbounds double, double* %5, i64 %idxprom2 - %6 = load double* %arrayidx3, align 8 - %7 = load i32* %i, align 4 + %6 = load double, double* %arrayidx3, align 8 + %7 = load i32, i32* %i, align 4 %idxprom4 = sext i32 %7 to i64 - %8 = load i32** %trigger.addr, align 8 + %8 = load i32*, i32** %trigger.addr, align 8 %arrayidx5 = getelementptr inbounds i32, i32* %8, i64 %idxprom4 - %9 = load i32* %arrayidx5, align 4 + %9 = load i32, i32* %arrayidx5, align 4 %conv = sitofp i32 %9 to double %add = fadd double %6, %conv - %10 = load i32* %i, align 4 + %10 = load i32, i32* %i, align 4 %idxprom6 = sext i32 %10 to i64 - %11 = load double** %A.addr, align 8 + %11 = load double*, double** %A.addr, align 8 %arrayidx7 = getelementptr inbounds double, double* %11, i64 %idxprom6 store double %add, double* %arrayidx7, align 8 br label %if.end @@ -255,7 +255,7 @@ if.end: ; preds = %if.then, %for.body br label %for.inc for.inc: ; preds = %if.end - %12 = load i32* %i, align 4 + %12 = load i32, i32* %i, align 4 %inc = add nsw i32 %12, 1 store i32 %inc, i32* %i, align 4 br label %for.cond @@ -297,36 +297,36 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 10000 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %idxprom = sext i32 %1 to i64 - %2 = load i32** %trigger.addr, align 8 + %2 = load i32*, i32** %trigger.addr, align 8 %arrayidx = getelementptr inbounds i32, i32* %2, i64 %idxprom - %3 = load i32* %arrayidx, align 4 + %3 = load i32, i32* %arrayidx, align 4 %cmp1 = icmp slt i32 %3, 100 br i1 %cmp1, label %if.then, label %if.end if.then: ; preds = %for.body - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %mul = mul nsw i32 %4, 2 %idxprom2 = sext i32 %mul to i64 - %5 = load double** %B.addr, align 8 + %5 = load double*, double** %B.addr, align 8 %arrayidx3 = getelementptr inbounds double, double* %5, i64 %idxprom2 - %6 = load double* %arrayidx3, align 8 - %7 = load i32* %i, align 4 + %6 = load double, double* %arrayidx3, align 8 + %7 = load i32, i32* %i, align 4 %idxprom4 = sext i32 %7 to i64 - %8 = load i32** %trigger.addr, align 8 + %8 = load i32*, i32** %trigger.addr, align 8 %arrayidx5 = getelementptr inbounds i32, i32* %8, i64 %idxprom4 - %9 = load i32* %arrayidx5, align 4 + %9 = load i32, i32* %arrayidx5, align 4 %conv = sitofp i32 %9 to double %add = fadd double %6, %conv - %10 = load i32* %i, align 4 + %10 = load i32, i32* %i, align 4 %idxprom6 = sext i32 %10 to i64 - %11 = load double** %A.addr, align 8 + %11 = load double*, double** %A.addr, align 8 %arrayidx7 = getelementptr inbounds double, double* %11, i64 %idxprom6 store double %add, double* %arrayidx7, align 8 br label %if.end @@ -335,7 +335,7 @@ if.end: ; preds = %if.then, %for.body br label %for.inc for.inc: ; preds = %if.end - %12 = load i32* %i, align 4 + %12 = load i32, i32* %i, align 4 %inc = add nsw i32 %12, 1 store i32 %inc, i32* %i, align 4 br label %for.cond @@ -373,34 +373,34 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 10000 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %idxprom = sext i32 %1 to i64 - %2 = load i32** %trigger.addr, align 8 + %2 = load i32*, i32** %trigger.addr, align 8 %arrayidx = getelementptr inbounds i32, i32* %2, i64 %idxprom - %3 = load i32* %arrayidx, align 4 + %3 = load i32, i32* %arrayidx, align 4 %cmp1 = icmp slt i32 %3, 100 br i1 %cmp1, label %if.then, label %if.end if.then: ; preds = %for.body - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %idxprom2 = sext i32 %4 to i64 - %5 = load i32** %B.addr, align 8 + %5 = load i32*, i32** %B.addr, align 8 %arrayidx3 = getelementptr inbounds i32, i32* %5, i64 %idxprom2 - %6 = load i32* %arrayidx3, align 4 - %7 = load i32* %i, align 4 + %6 = load i32, i32* %arrayidx3, align 4 + %7 = load i32, i32* %i, align 4 %idxprom4 = sext i32 %7 to i64 - %8 = load i32** %trigger.addr, align 8 + %8 = load i32*, i32** %trigger.addr, align 8 %arrayidx5 = getelementptr inbounds i32, i32* %8, i64 %idxprom4 - %9 = load i32* %arrayidx5, align 4 + %9 = load i32, i32* %arrayidx5, align 4 %add = add nsw i32 %6, %9 - %10 = load i32* %i, align 4 + %10 = load i32, i32* %i, align 4 %idxprom6 = sext i32 %10 to i64 - %11 = load i32** %A.addr, align 8 + %11 = load i32*, i32** %A.addr, align 8 %arrayidx7 = getelementptr inbounds i32, i32* %11, i64 %idxprom6 store i32 sdiv (i32 1, i32 zext (i1 icmp eq (i32** getelementptr inbounds ([1 x i32*]* @a, i64 0, i64 1), i32** @c) to i32)), i32* %arrayidx7, align 4 br label %if.end @@ -409,7 +409,7 @@ if.end: ; preds = %if.then, %for.body br label %for.inc for.inc: ; preds = %if.end - %12 = load i32* %i, align 4 + %12 = load i32, i32* %i, align 4 %inc = add nsw i32 %12, 1 store i32 %inc, i32* %i, align 4 br label %for.cond @@ -459,29 +459,29 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp sge i32 %0, 0 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %idxprom = sext i32 %1 to i64 - %2 = load i32** %trigger.addr, align 8 + %2 = load i32*, i32** %trigger.addr, align 8 %arrayidx = getelementptr inbounds i32, i32* %2, i64 %idxprom - %3 = load i32* %arrayidx, align 4 + %3 = load i32, i32* %arrayidx, align 4 %cmp1 = icmp sgt i32 %3, 0 br i1 %cmp1, label %if.then, label %if.end if.then: ; preds = %for.body - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %idxprom2 = sext i32 %4 to i64 - %5 = load double** %in.addr, align 8 + %5 = load double*, double** %in.addr, align 8 %arrayidx3 = getelementptr inbounds double, double* %5, i64 %idxprom2 - %6 = load double* %arrayidx3, align 8 + %6 = load double, double* %arrayidx3, align 8 %add = fadd double %6, 5.000000e-01 - %7 = load i32* %i, align 4 + %7 = load i32, i32* %i, align 4 %idxprom4 = sext i32 %7 to i64 - %8 = load double** %out.addr, align 8 + %8 = load double*, double** %out.addr, align 8 %arrayidx5 = getelementptr inbounds double, double* %8, i64 %idxprom4 store double %add, double* %arrayidx5, align 8 br label %if.end @@ -490,7 +490,7 @@ if.end: ; preds = %if.then, %for.body br label %for.inc for.inc: ; preds = %if.end - %9 = load i32* %i, align 4 + %9 = load i32, i32* %i, align 4 %dec = add nsw i32 %9, -1 store i32 %dec, i32* %i, align 4 br label %for.cond diff --git a/llvm/test/Transforms/LoopVectorize/X86/metadata-enable.ll b/llvm/test/Transforms/LoopVectorize/X86/metadata-enable.ll index bdce3ad3d04..ba8e11e5874 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/metadata-enable.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/metadata-enable.ll @@ -55,7 +55,7 @@ entry: for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds i32, i32* %b, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %add = add nsw i32 %0, %N %arrayidx2 = getelementptr inbounds i32, i32* %a, i64 %indvars.iv store i32 %add, i32* %arrayidx2, align 4 @@ -64,7 +64,7 @@ for.body: ; preds = %for.body, %entry br i1 %exitcond, label %for.end, label %for.body, !llvm.loop !0 for.end: ; preds = %for.body - %1 = load i32* %a, align 4 + %1 = load i32, i32* %a, align 4 ret i32 %1 } @@ -106,7 +106,7 @@ entry: for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds i32, i32* %b, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %add = add nsw i32 %0, %N %arrayidx2 = getelementptr inbounds i32, i32* %a, i64 %indvars.iv store i32 %add, i32* %arrayidx2, align 4 @@ -115,7 +115,7 @@ for.body: ; preds = %for.body, %entry br i1 %exitcond, label %for.end, label %for.body for.end: ; preds = %for.body - %1 = load i32* %a, align 4 + %1 = load i32, i32* %a, align 4 ret i32 %1 } @@ -157,7 +157,7 @@ entry: for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds i32, i32* %b, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %add = add nsw i32 %0, %N %arrayidx2 = getelementptr inbounds i32, i32* %a, i64 %indvars.iv store i32 %add, i32* %arrayidx2, align 4 @@ -166,7 +166,7 @@ for.body: ; preds = %for.body, %entry br i1 %exitcond, label %for.end, label %for.body, !llvm.loop !2 for.end: ; preds = %for.body - %1 = load i32* %a, align 4 + %1 = load i32, i32* %a, align 4 ret i32 %1 } diff --git a/llvm/test/Transforms/LoopVectorize/X86/min-trip-count-switch.ll b/llvm/test/Transforms/LoopVectorize/X86/min-trip-count-switch.ll index 3207025f167..bb972c4488a 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/min-trip-count-switch.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/min-trip-count-switch.ll @@ -11,7 +11,7 @@ entry: for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds float, float* %a, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %add = fadd float %0, 1.000000e+00 store float %add, float* %arrayidx, align 4 %indvars.iv.next = add i64 %indvars.iv, 1 diff --git a/llvm/test/Transforms/LoopVectorize/X86/no-vector.ll b/llvm/test/Transforms/LoopVectorize/X86/no-vector.ll index 1ac18f27f83..4b464b01267 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/no-vector.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/no-vector.ll @@ -9,7 +9,7 @@ for.body: ; preds = %entry, %for.body %i.06 = phi i32 [ %inc, %for.body ], [ 0, %entry ] %r.05 = phi i32 [ %xor, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds i8, i8* %s, i32 %i.06 - %0 = load i8* %arrayidx, align 1 + %0 = load i8, i8* %arrayidx, align 1 %conv = sext i8 %0 to i32 %xor = xor i32 %conv, %r.05 %inc = add nsw i32 %i.06, 1 diff --git a/llvm/test/Transforms/LoopVectorize/X86/parallel-loops-after-reg2mem.ll b/llvm/test/Transforms/LoopVectorize/X86/parallel-loops-after-reg2mem.ll index a7f636290ba..631361c9d47 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/parallel-loops-after-reg2mem.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/parallel-loops-after-reg2mem.ll @@ -17,28 +17,28 @@ entry: br label %for.body for.body: ; preds = %for.body.for.body_crit_edge, %entry - %indvars.iv.reload = load i64* %indvars.iv.reg2mem + %indvars.iv.reload = load i64, i64* %indvars.iv.reg2mem %arrayidx = getelementptr inbounds i32, i32* %b, i64 %indvars.iv.reload - %0 = load i32* %arrayidx, align 4, !llvm.mem.parallel_loop_access !3 + %0 = load i32, i32* %arrayidx, align 4, !llvm.mem.parallel_loop_access !3 %arrayidx2 = getelementptr inbounds i32, i32* %a, i64 %indvars.iv.reload - %1 = load i32* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !3 + %1 = load i32, i32* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !3 %idxprom3 = sext i32 %1 to i64 %arrayidx4 = getelementptr inbounds i32, i32* %a, i64 %idxprom3 store i32 %0, i32* %arrayidx4, align 4, !llvm.mem.parallel_loop_access !3 %indvars.iv.next = add i64 %indvars.iv.reload, 1 ; A new store without the parallel metadata here: store i64 %indvars.iv.next, i64* %indvars.iv.next.reg2mem - %indvars.iv.next.reload1 = load i64* %indvars.iv.next.reg2mem + %indvars.iv.next.reload1 = load i64, i64* %indvars.iv.next.reg2mem %arrayidx6 = getelementptr inbounds i32, i32* %b, i64 %indvars.iv.next.reload1 - %2 = load i32* %arrayidx6, align 4, !llvm.mem.parallel_loop_access !3 + %2 = load i32, i32* %arrayidx6, align 4, !llvm.mem.parallel_loop_access !3 store i32 %2, i32* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !3 - %indvars.iv.next.reload = load i64* %indvars.iv.next.reg2mem + %indvars.iv.next.reload = load i64, i64* %indvars.iv.next.reg2mem %lftr.wideiv = trunc i64 %indvars.iv.next.reload to i32 %exitcond = icmp eq i32 %lftr.wideiv, 512 br i1 %exitcond, label %for.end, label %for.body.for.body_crit_edge, !llvm.loop !3 for.body.for.body_crit_edge: ; preds = %for.body - %indvars.iv.next.reload2 = load i64* %indvars.iv.next.reg2mem + %indvars.iv.next.reload2 = load i64, i64* %indvars.iv.next.reg2mem store i64 %indvars.iv.next.reload2, i64* %indvars.iv.reg2mem br label %for.body diff --git a/llvm/test/Transforms/LoopVectorize/X86/parallel-loops.ll b/llvm/test/Transforms/LoopVectorize/X86/parallel-loops.ll index 83bed435de4..53061edc35c 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/parallel-loops.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/parallel-loops.ll @@ -21,15 +21,15 @@ entry: for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds i32, i32* %b, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds i32, i32* %a, i64 %indvars.iv - %1 = load i32* %arrayidx2, align 4 + %1 = load i32, i32* %arrayidx2, align 4 %idxprom3 = sext i32 %1 to i64 %arrayidx4 = getelementptr inbounds i32, i32* %a, i64 %idxprom3 store i32 %0, i32* %arrayidx4, align 4 %indvars.iv.next = add i64 %indvars.iv, 1 %arrayidx6 = getelementptr inbounds i32, i32* %b, i64 %indvars.iv.next - %2 = load i32* %arrayidx6, align 4 + %2 = load i32, i32* %arrayidx6, align 4 store i32 %2, i32* %arrayidx2, align 4 %lftr.wideiv = trunc i64 %indvars.iv.next to i32 %exitcond = icmp eq i32 %lftr.wideiv, 512 @@ -51,9 +51,9 @@ entry: for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds i32, i32* %b, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4, !llvm.mem.parallel_loop_access !3 + %0 = load i32, i32* %arrayidx, align 4, !llvm.mem.parallel_loop_access !3 %arrayidx2 = getelementptr inbounds i32, i32* %a, i64 %indvars.iv - %1 = load i32* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !3 + %1 = load i32, i32* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !3 %idxprom3 = sext i32 %1 to i64 %arrayidx4 = getelementptr inbounds i32, i32* %a, i64 %idxprom3 ; This store might have originated from inlining a function with a parallel @@ -61,7 +61,7 @@ for.body: ; preds = %for.body, %entry store i32 %0, i32* %arrayidx4, align 4, !llvm.mem.parallel_loop_access !5 %indvars.iv.next = add i64 %indvars.iv, 1 %arrayidx6 = getelementptr inbounds i32, i32* %b, i64 %indvars.iv.next - %2 = load i32* %arrayidx6, align 4, !llvm.mem.parallel_loop_access !3 + %2 = load i32, i32* %arrayidx6, align 4, !llvm.mem.parallel_loop_access !3 store i32 %2, i32* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !3 %lftr.wideiv = trunc i64 %indvars.iv.next to i32 %exitcond = icmp eq i32 %lftr.wideiv, 512 @@ -84,9 +84,9 @@ entry: for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds i32, i32* %b, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4, !llvm.mem.parallel_loop_access !6 + %0 = load i32, i32* %arrayidx, align 4, !llvm.mem.parallel_loop_access !6 %arrayidx2 = getelementptr inbounds i32, i32* %a, i64 %indvars.iv - %1 = load i32* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !6 + %1 = load i32, i32* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !6 %idxprom3 = sext i32 %1 to i64 %arrayidx4 = getelementptr inbounds i32, i32* %a, i64 %idxprom3 ; This refers to the loop marked with !7 which we are not in at the moment. @@ -94,7 +94,7 @@ for.body: ; preds = %for.body, %entry store i32 %0, i32* %arrayidx4, align 4, !llvm.mem.parallel_loop_access !7 %indvars.iv.next = add i64 %indvars.iv, 1 %arrayidx6 = getelementptr inbounds i32, i32* %b, i64 %indvars.iv.next - %2 = load i32* %arrayidx6, align 4, !llvm.mem.parallel_loop_access !6 + %2 = load i32, i32* %arrayidx6, align 4, !llvm.mem.parallel_loop_access !6 store i32 %2, i32* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !6 %lftr.wideiv = trunc i64 %indvars.iv.next to i32 %exitcond = icmp eq i32 %lftr.wideiv, 512 diff --git a/llvm/test/Transforms/LoopVectorize/X86/powof2div.ll b/llvm/test/Transforms/LoopVectorize/X86/powof2div.ll index af3cfe06ef9..6bc738a7d14 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/powof2div.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/powof2div.ll @@ -7,7 +7,7 @@ target triple = "x86_64-unknown-linux-gnu" @Foo = common global %struct.anon zeroinitializer, align 4 ;CHECK-LABEL: @foo( -;CHECK: load <4 x i32>* +;CHECK: load <4 x i32>, <4 x i32>* ;CHECK: sdiv <4 x i32> ;CHECK: store <4 x i32> @@ -18,7 +18,7 @@ entry: for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds %struct.anon, %struct.anon* @Foo, i64 0, i32 2, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %div = sdiv i32 %0, 2 %arrayidx2 = getelementptr inbounds %struct.anon, %struct.anon* @Foo, i64 0, i32 0, i64 %indvars.iv store i32 %div, i32* %arrayidx2, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/X86/reduction-crash.ll b/llvm/test/Transforms/LoopVectorize/X86/reduction-crash.ll index ce2aa8963e2..3741b95d985 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/reduction-crash.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/reduction-crash.ll @@ -14,14 +14,14 @@ bb: br label %bb2 bb2: ; preds = %bb - %tmp = load double* null, align 8 + %tmp = load double, double* null, align 8 br i1 undef, label %bb3, label %bb12 bb3: ; preds = %bb3, %bb2 %tmp4 = phi double [ %tmp9, %bb3 ], [ %tmp, %bb2 ] %tmp5 = phi i32 [ %tmp8, %bb3 ], [ 0, %bb2 ] %tmp6 = getelementptr inbounds [16 x double], [16 x double]* undef, i32 0, i32 %tmp5 - %tmp7 = load double* %tmp6, align 4 + %tmp7 = load double, double* %tmp6, align 4 %tmp8 = add nsw i32 %tmp5, 1 %tmp9 = fadd fast double %tmp4, undef %tmp10 = getelementptr inbounds float, float* %arg, i32 %tmp5 diff --git a/llvm/test/Transforms/LoopVectorize/X86/small-size.ll b/llvm/test/Transforms/LoopVectorize/X86/small-size.ll index 38e3c086347..47c262b11b4 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/small-size.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/small-size.ll @@ -31,9 +31,9 @@ define void @example1() optsize { ; <label>:1 ; preds = %1, %0 %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %2 = getelementptr inbounds [2048 x i32], [2048 x i32]* @b, i64 0, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = getelementptr inbounds [2048 x i32], [2048 x i32]* @c, i64 0, i64 %indvars.iv - %5 = load i32* %4, align 4 + %5 = load i32, i32* %4, align 4 %6 = add nsw i32 %5, %3 %7 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv store i32 %6, i32* %7, align 4 @@ -77,9 +77,9 @@ define void @example2(i32 %n, i32 %x) optsize { %.02 = phi i32 [ %4, %.lr.ph ], [ %n, %.preheader ] %4 = add nsw i32 %.02, -1 %5 = getelementptr inbounds [2048 x i32], [2048 x i32]* @b, i64 0, i64 %indvars.iv - %6 = load i32* %5, align 4 + %6 = load i32, i32* %5, align 4 %7 = getelementptr inbounds [2048 x i32], [2048 x i32]* @c, i64 0, i64 %indvars.iv - %8 = load i32* %7, align 4 + %8 = load i32, i32* %7, align 4 %9 = and i32 %8, %6 %10 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv store i32 %9, i32* %10, align 4 @@ -105,7 +105,7 @@ define void @example3(i32 %n, i32* noalias nocapture %p, i32* noalias nocapture %.023 = phi i32* [ %3, %.lr.ph ], [ %q, %0 ] %2 = add nsw i32 %.05, -1 %3 = getelementptr inbounds i32, i32* %.023, i64 1 - %4 = load i32* %.023, align 16 + %4 = load i32, i32* %.023, align 16 %5 = getelementptr inbounds i32, i32* %.014, i64 1 store i32 %4, i32* %.014, align 16 %6 = icmp eq i32 %2, 0 @@ -129,7 +129,7 @@ define void @example4(i32 %n, i32* noalias nocapture %p, i32* noalias nocapture %.023 = phi i32* [ %3, %.lr.ph ], [ %q, %0 ] %2 = add nsw i32 %.05, -1 %3 = getelementptr inbounds i32, i32* %.023, i64 1 - %4 = load i32* %.023, align 16 + %4 = load i32, i32* %.023, align 16 %5 = getelementptr inbounds i32, i32* %.014, i64 1 store i32 %4, i32* %.014, align 16 %6 = icmp eq i32 %2, 0 @@ -153,7 +153,7 @@ define void @example23(i16* nocapture %src, i32* nocapture %dst) optsize { %.013 = phi i32* [ %dst, %0 ], [ %6, %1 ] %i.02 = phi i32 [ 0, %0 ], [ %7, %1 ] %2 = getelementptr inbounds i16, i16* %.04, i64 1 - %3 = load i16* %.04, align 2 + %3 = load i16, i16* %.04, align 2 %4 = zext i16 %3 to i32 %5 = shl nuw nsw i32 %4, 7 %6 = getelementptr inbounds i32, i32* %.013, i64 1 @@ -179,7 +179,7 @@ define void @example23b(i16* noalias nocapture %src, i32* noalias nocapture %dst %.013 = phi i32* [ %dst, %0 ], [ %6, %1 ] %i.02 = phi i32 [ 0, %0 ], [ %7, %1 ] %2 = getelementptr inbounds i16, i16* %.04, i64 1 - %3 = load i16* %.04, align 2 + %3 = load i16, i16* %.04, align 2 %4 = zext i16 %3 to i32 %5 = shl nuw nsw i32 %4, 7 %6 = getelementptr inbounds i32, i32* %.013, i64 1 diff --git a/llvm/test/Transforms/LoopVectorize/X86/tripcount.ll b/llvm/test/Transforms/LoopVectorize/X86/tripcount.ll index 9984e2d04e8..c0bbb92c2c5 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/tripcount.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/tripcount.ll @@ -23,7 +23,7 @@ for.body.preheader: for.body: %i.07 = phi i32 [ %inc, %for.body ], [ 0, %for.body.preheader ] %arrayidx = getelementptr inbounds [0 x i32], [0 x i32]* @big, i32 0, i32 %i.07 - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %neg = xor i32 %0, -1 store i32 %neg, i32* %arrayidx, align 4 %inc = add nsw i32 %i.07, 1 diff --git a/llvm/test/Transforms/LoopVectorize/X86/uint64_to_fp64-cost-model.ll b/llvm/test/Transforms/LoopVectorize/X86/uint64_to_fp64-cost-model.ll index 40421d4cff0..38af11c443d 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/uint64_to_fp64-cost-model.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/uint64_to_fp64-cost-model.ll @@ -13,7 +13,7 @@ entry: for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds i64, i64* %a, i64 %indvars.iv - %tmp = load i64* %arrayidx, align 4 + %tmp = load i64, i64* %arrayidx, align 4 %conv = uitofp i64 %tmp to double %arrayidx2 = getelementptr inbounds double, double* %b, i64 %indvars.iv store double %conv, double* %arrayidx2, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/X86/unroll-pm.ll b/llvm/test/Transforms/LoopVectorize/X86/unroll-pm.ll index 50fdf187499..52914b6a7c6 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/unroll-pm.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/unroll-pm.ll @@ -18,7 +18,7 @@ define i32 @bar(i32* nocapture %A, i32 %n) nounwind uwtable ssp { .lr.ph: ; preds = %0, %.lr.ph %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 0, %0 ] %2 = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = add nsw i32 %3, 6 store i32 %4, i32* %2, align 4 %indvars.iv.next = add i64 %indvars.iv, 1 diff --git a/llvm/test/Transforms/LoopVectorize/X86/unroll-small-loops.ll b/llvm/test/Transforms/LoopVectorize/X86/unroll-small-loops.ll index eff6ca49edc..4411da3f0a9 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/unroll-small-loops.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/unroll-small-loops.ll @@ -16,8 +16,8 @@ target triple = "x86_64-apple-macosx10.8.0" ; CHECK-VECTOR: ret ; ; CHECK-SCALAR-LABEL: @foo( -; CHECK-SCALAR: load i32* -; CHECK-SCALAR-NOT: load i32* +; CHECK-SCALAR: load i32, i32* +; CHECK-SCALAR-NOT: load i32, i32* ; CHECK-SCALAR: store i32 ; CHECK-SCALAR-NOT: store i32 ; CHECK-SCALAR: ret @@ -27,7 +27,7 @@ define i32 @foo(i32* nocapture %A) nounwind uwtable ssp { ; <label>:1 ; preds = %1, %0 %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %2 = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = add nsw i32 %3, 6 store i32 %4, i32* %2, align 4 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -58,7 +58,7 @@ define i32 @bar(i32* nocapture %A, i32 %n) nounwind uwtable ssp { .lr.ph: ; preds = %0, %.lr.ph %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 0, %0 ] %2 = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = add nsw i32 %3, 6 store i32 %4, i32* %2, align 4 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -87,7 +87,7 @@ entry: for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds float, float* %B, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %mul = fmul float %0, %N %arrayidx2 = getelementptr inbounds float, float* %A, i64 %indvars.iv store float %mul, float* %arrayidx2, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/X86/unroll_selection.ll b/llvm/test/Transforms/LoopVectorize/X86/unroll_selection.ll index 360d66d3571..71b829071e2 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/unroll_selection.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/unroll_selection.ll @@ -17,7 +17,7 @@ define void @reg_pressure(double* nocapture %A, i32 %n) nounwind uwtable ssp { ; <label>:2 ; preds = %2, %0 %indvars.iv = phi i64 [ %indvars.iv.next, %2 ], [ %1, %0 ] %3 = getelementptr inbounds double, double* %A, i64 %indvars.iv - %4 = load double* %3, align 8 + %4 = load double, double* %3, align 8 %5 = fadd double %4, 3.000000e+00 %6 = fmul double %4, 2.000000e+00 %7 = fadd double %5, %6 @@ -59,7 +59,7 @@ define void @small_loop(i16* nocapture %A, i64 %n) nounwind uwtable ssp { .lr.ph: ; preds = %0, %.lr.ph %i.01 = phi i64 [ %5, %.lr.ph ], [ 0, %0 ] %2 = getelementptr inbounds i16, i16* %A, i64 %i.01 - %3 = load i16* %2, align 2 + %3 = load i16, i16* %2, align 2 %4 = xor i16 %3, 3 store i16 %4, i16* %2, align 2 %5 = add i64 %i.01, 1 diff --git a/llvm/test/Transforms/LoopVectorize/X86/vect.omp.force.ll b/llvm/test/Transforms/LoopVectorize/X86/vect.omp.force.ll index e6a0d475161..10307568b85 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/vect.omp.force.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/vect.omp.force.ll @@ -36,7 +36,7 @@ for.body.preheader: for.body: %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %for.body.preheader ] %arrayidx = getelementptr inbounds float, float* %B, i64 %indvars.iv - %0 = load float* %arrayidx, align 4, !llvm.mem.parallel_loop_access !1 + %0 = load float, float* %arrayidx, align 4, !llvm.mem.parallel_loop_access !1 %call = tail call float @llvm.sin.f32(float %0) %arrayidx2 = getelementptr inbounds float, float* %A, i64 %indvars.iv store float %call, float* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !1 @@ -70,7 +70,7 @@ for.body.preheader: for.body: %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %for.body.preheader ] %arrayidx = getelementptr inbounds float, float* %B, i64 %indvars.iv - %0 = load float* %arrayidx, align 4, !llvm.mem.parallel_loop_access !3 + %0 = load float, float* %arrayidx, align 4, !llvm.mem.parallel_loop_access !3 %call = tail call float @llvm.sin.f32(float %0) %arrayidx2 = getelementptr inbounds float, float* %A, i64 %indvars.iv store float %call, float* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !3 diff --git a/llvm/test/Transforms/LoopVectorize/X86/vect.omp.force.small-tc.ll b/llvm/test/Transforms/LoopVectorize/X86/vect.omp.force.small-tc.ll index f4bc1600c3c..8d139ac7e5a 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/vect.omp.force.small-tc.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/vect.omp.force.small-tc.ll @@ -30,9 +30,9 @@ entry: for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds float, float* %B, i64 %indvars.iv - %0 = load float* %arrayidx, align 4, !llvm.mem.parallel_loop_access !1 + %0 = load float, float* %arrayidx, align 4, !llvm.mem.parallel_loop_access !1 %arrayidx2 = getelementptr inbounds float, float* %A, i64 %indvars.iv - %1 = load float* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !1 + %1 = load float, float* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !1 %add = fadd fast float %0, %1 store float %add, float* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !1 %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1 @@ -56,9 +56,9 @@ entry: for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds float, float* %B, i64 %indvars.iv - %0 = load float* %arrayidx, align 4, !llvm.mem.parallel_loop_access !3 + %0 = load float, float* %arrayidx, align 4, !llvm.mem.parallel_loop_access !3 %arrayidx2 = getelementptr inbounds float, float* %A, i64 %indvars.iv - %1 = load float* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !3 + %1 = load float, float* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !3 %add = fadd fast float %0, %1 store float %add, float* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !3 %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1 diff --git a/llvm/test/Transforms/LoopVectorize/X86/vector-scalar-select-cost.ll b/llvm/test/Transforms/LoopVectorize/X86/vector-scalar-select-cost.ll index 86cebf342a6..5efabe16d32 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/vector-scalar-select-cost.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/vector-scalar-select-cost.ll @@ -16,9 +16,9 @@ define void @scalarselect(i1 %cond) { ; <label>:1 %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %2 = getelementptr inbounds [2048 x i32], [2048 x i32]* @b, i64 0, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = getelementptr inbounds [2048 x i32], [2048 x i32]* @c, i64 0, i64 %indvars.iv - %5 = load i32* %4, align 4 + %5 = load i32, i32* %4, align 4 %6 = add nsw i32 %5, %3 %7 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv @@ -43,9 +43,9 @@ define void @vectorselect(i1 %cond) { ; <label>:1 %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %2 = getelementptr inbounds [2048 x i32], [2048 x i32]* @b, i64 0, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = getelementptr inbounds [2048 x i32], [2048 x i32]* @c, i64 0, i64 %indvars.iv - %5 = load i32* %4, align 4 + %5 = load i32, i32* %4, align 4 %6 = add nsw i32 %5, %3 %7 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv %8 = icmp ult i64 %indvars.iv, 8 diff --git a/llvm/test/Transforms/LoopVectorize/X86/vector_ptr_load_store.ll b/llvm/test/Transforms/LoopVectorize/X86/vector_ptr_load_store.ll index 048c2603bbd..6cd3c9c3bc0 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/vector_ptr_load_store.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/vector_ptr_load_store.ll @@ -19,7 +19,7 @@ target triple = "x86_64-apple-macosx10.8.0" ; CHECK: test_consecutive_store ; CHECK: The Widest type: 64 bits define void @test_consecutive_store(%0**, %0**, %0** nocapture) nounwind ssp uwtable align 2 { - %4 = load %0** %2, align 8 + %4 = load %0*, %0** %2, align 8 %5 = icmp eq %0** %0, %1 br i1 %5, label %12, label %6 @@ -62,7 +62,7 @@ define void @test_nonconsecutive_store() nounwind ssp uwtable { ; <label>:3 ; preds = %3, %1 %4 = phi i64 [ 0, %1 ], [ %11, %3 ] %5 = getelementptr inbounds [2048 x i16], [2048 x i16]* @q, i64 0, i64 %4 - %6 = load i16* %5, align 2 + %6 = load i16, i16* %5, align 2 %7 = sext i16 %6 to i64 %8 = add i64 %7, 1 %9 = inttoptr i64 %8 to i32* @@ -101,7 +101,7 @@ define i8 @test_consecutive_ptr_load() nounwind readonly ssp uwtable { %2 = phi i64 [ 0, %0 ], [ %10, %1 ] %3 = phi i8 [ 0, %0 ], [ %9, %1 ] %4 = getelementptr inbounds [1024 x i32*], [1024 x i32*]* @ia, i32 0, i64 %2 - %5 = load i32** %4, align 4 + %5 = load i32*, i32** %4, align 4 %6 = ptrtoint i32* %5 to i64 %7 = trunc i64 %6 to i8 %8 = add i8 %3, 1 @@ -129,7 +129,7 @@ define void @test_nonconsecutive_ptr_load() nounwind ssp uwtable { %4 = phi i64 [ 0, %1 ], [ %10, %3 ] %5 = getelementptr inbounds [2048 x [8 x i32*]], [2048 x [8 x i32*]]* @p2, i64 0, i64 %4, i64 %2 %6 = getelementptr inbounds [2048 x i16], [2048 x i16]* @q2, i64 0, i64 %4 - %7 = load i32** %5, align 2 + %7 = load i32*, i32** %5, align 2 %8 = ptrtoint i32* %7 to i64 %9 = trunc i64 %8 to i16 store i16 %9, i16* %6, align 8 diff --git a/llvm/test/Transforms/LoopVectorize/X86/vectorization-remarks-missed.ll b/llvm/test/Transforms/LoopVectorize/X86/vectorization-remarks-missed.ll index 92eba528c0d..2ab0ee3ea30 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/vectorization-remarks-missed.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/vectorization-remarks-missed.ll @@ -98,10 +98,10 @@ for.body.preheader: ; preds = %entry for.body: ; preds = %for.body.preheader, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %for.body.preheader ] %arrayidx = getelementptr inbounds i32, i32* %B, i64 %indvars.iv, !dbg !35 - %0 = load i32* %arrayidx, align 4, !dbg !35, !tbaa !18 + %0 = load i32, i32* %arrayidx, align 4, !dbg !35, !tbaa !18 %idxprom1 = sext i32 %0 to i64, !dbg !35 %arrayidx2 = getelementptr inbounds i32, i32* %A, i64 %idxprom1, !dbg !35 - %1 = load i32* %arrayidx2, align 4, !dbg !35, !tbaa !18 + %1 = load i32, i32* %arrayidx2, align 4, !dbg !35, !tbaa !18 %arrayidx4 = getelementptr inbounds i32, i32* %A, i64 %indvars.iv, !dbg !35 store i32 %1, i32* %arrayidx4, align 4, !dbg !35, !tbaa !18 %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1, !dbg !32 diff --git a/llvm/test/Transforms/LoopVectorize/X86/vectorization-remarks.ll b/llvm/test/Transforms/LoopVectorize/X86/vectorization-remarks.ll index cf64283b248..10e27c10cc0 100644 --- a/llvm/test/Transforms/LoopVectorize/X86/vectorization-remarks.ll +++ b/llvm/test/Transforms/LoopVectorize/X86/vectorization-remarks.ll @@ -27,10 +27,10 @@ for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %add8 = phi i32 [ 0, %entry ], [ %add, %for.body ], !dbg !19 %arrayidx = getelementptr inbounds [16 x i8], [16 x i8]* %cb, i64 0, i64 %indvars.iv, !dbg !19 - %0 = load i8* %arrayidx, align 1, !dbg !19, !tbaa !21 + %0 = load i8, i8* %arrayidx, align 1, !dbg !19, !tbaa !21 %conv = sext i8 %0 to i32, !dbg !19 %arrayidx2 = getelementptr inbounds [16 x i8], [16 x i8]* %cc, i64 0, i64 %indvars.iv, !dbg !19 - %1 = load i8* %arrayidx2, align 1, !dbg !19, !tbaa !21 + %1 = load i8, i8* %arrayidx2, align 1, !dbg !19, !tbaa !21 %conv3 = sext i8 %1 to i32, !dbg !19 %sub = sub i32 %conv, %conv3, !dbg !19 %add = add nsw i32 %sub, %add8, !dbg !19 diff --git a/llvm/test/Transforms/LoopVectorize/align.ll b/llvm/test/Transforms/LoopVectorize/align.ll index f12dbde416d..7ee401deef4 100644 --- a/llvm/test/Transforms/LoopVectorize/align.ll +++ b/llvm/test/Transforms/LoopVectorize/align.ll @@ -6,8 +6,8 @@ target triple = "x86_64-apple-macosx10.8.0" ; Make sure we output the abi alignment if no alignment is specified. ;CHECK-LABEL: @align -;CHECK: load <4 x i32>* {{.*}} align 4 -;CHECK: load <4 x i32>* {{.*}} align 4 +;CHECK: load <4 x i32>, <4 x i32>* {{.*}} align 4 +;CHECK: load <4 x i32>, <4 x i32>* {{.*}} align 4 ;CHECK: store <4 x i32> {{.*}} align 4 define void @align(i32* %a, i32* %b, i32* %c) nounwind uwtable ssp { @@ -16,9 +16,9 @@ define void @align(i32* %a, i32* %b, i32* %c) nounwind uwtable ssp { ; <label>:1 ; preds = %1, %0 %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %2 = getelementptr inbounds i32, i32* %b, i64 %indvars.iv - %3 = load i32* %2 + %3 = load i32, i32* %2 %4 = getelementptr inbounds i32, i32* %c, i64 %indvars.iv - %5 = load i32* %4 + %5 = load i32, i32* %4 %6 = add nsw i32 %5, %3 %7 = getelementptr inbounds i32, i32* %a, i64 %indvars.iv store i32 %6, i32* %7 diff --git a/llvm/test/Transforms/LoopVectorize/bzip_reverse_loops.ll b/llvm/test/Transforms/LoopVectorize/bzip_reverse_loops.ll index 707b5b9fd32..f1efb256391 100644 --- a/llvm/test/Transforms/LoopVectorize/bzip_reverse_loops.ll +++ b/llvm/test/Transforms/LoopVectorize/bzip_reverse_loops.ll @@ -17,7 +17,7 @@ do.body: ; preds = %cond.end, %entry %n.addr.0 = phi i32 [ %n, %entry ], [ %dec, %cond.end ] %p.addr.0 = phi i16* [ %p, %entry ], [ %incdec.ptr, %cond.end ] %incdec.ptr = getelementptr inbounds i16, i16* %p.addr.0, i64 -1 - %0 = load i16* %incdec.ptr, align 2 + %0 = load i16, i16* %incdec.ptr, align 2 %conv = zext i16 %0 to i32 %cmp = icmp ult i32 %conv, %size br i1 %cmp, label %cond.end, label %cond.true @@ -52,7 +52,7 @@ do.body: ; preds = %do.body, %entry %n.addr.0 = phi i32 [ %n, %entry ], [ %dec, %do.body ] %p.0 = phi i32* [ %a, %entry ], [ %incdec.ptr, %do.body ] %incdec.ptr = getelementptr inbounds i32, i32* %p.0, i64 -1 - %0 = load i32* %incdec.ptr, align 4 + %0 = load i32, i32* %incdec.ptr, align 4 %cmp = icmp slt i32 %0, %wsize %sub = sub nsw i32 %0, %wsize %cond = select i1 %cmp, i32 0, i32 %sub diff --git a/llvm/test/Transforms/LoopVectorize/calloc.ll b/llvm/test/Transforms/LoopVectorize/calloc.ll index 3ac3b7b6106..a41e51787f3 100644 --- a/llvm/test/Transforms/LoopVectorize/calloc.ll +++ b/llvm/test/Transforms/LoopVectorize/calloc.ll @@ -23,7 +23,7 @@ for.body: ; preds = %for.body, %for.body %i.030 = phi i64 [ 0, %for.body.lr.ph ], [ %inc, %for.body ] %shr = lshr i64 %i.030, 1 %arrayidx = getelementptr inbounds i8, i8* %bytes, i64 %shr - %1 = load i8* %arrayidx, align 1 + %1 = load i8, i8* %arrayidx, align 1 %conv = zext i8 %1 to i32 %and = shl i64 %i.030, 2 %neg = and i64 %and, 4 diff --git a/llvm/test/Transforms/LoopVectorize/conditional-assignment.ll b/llvm/test/Transforms/LoopVectorize/conditional-assignment.ll index 7c523201b61..15750aad6b8 100644 --- a/llvm/test/Transforms/LoopVectorize/conditional-assignment.ll +++ b/llvm/test/Transforms/LoopVectorize/conditional-assignment.ll @@ -13,7 +13,7 @@ entry: for.body: ; preds = %for.inc, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.inc ] %arrayidx = getelementptr inbounds i32, i32* %indices, i64 %indvars.iv, !dbg !12 - %0 = load i32* %arrayidx, align 4, !dbg !12, !tbaa !14 + %0 = load i32, i32* %arrayidx, align 4, !dbg !12, !tbaa !14 %cmp1 = icmp eq i32 %0, 1024, !dbg !12 br i1 %cmp1, label %if.then, label %for.inc, !dbg !12 diff --git a/llvm/test/Transforms/LoopVectorize/control-flow.ll b/llvm/test/Transforms/LoopVectorize/control-flow.ll index ee73110cf6d..c95d55cef18 100644 --- a/llvm/test/Transforms/LoopVectorize/control-flow.ll +++ b/llvm/test/Transforms/LoopVectorize/control-flow.ll @@ -31,7 +31,7 @@ for.body.preheader: ; preds = %entry for.body: ; preds = %for.body.preheader, %if.else %indvars.iv = phi i64 [ %indvars.iv.next, %if.else ], [ 0, %for.body.preheader ] %arrayidx = getelementptr inbounds i32, i32* %A, i64 %indvars.iv, !dbg !12 - %0 = load i32* %arrayidx, align 4, !dbg !12, !tbaa !15 + %0 = load i32, i32* %arrayidx, align 4, !dbg !12, !tbaa !15 %cmp1 = icmp sgt i32 %0, 10, !dbg !12 br i1 %cmp1, label %end.loopexit, label %if.else, !dbg !12 diff --git a/llvm/test/Transforms/LoopVectorize/cpp-new-array.ll b/llvm/test/Transforms/LoopVectorize/cpp-new-array.ll index cc41e5c47d8..22896d33428 100644 --- a/llvm/test/Transforms/LoopVectorize/cpp-new-array.ll +++ b/llvm/test/Transforms/LoopVectorize/cpp-new-array.ll @@ -25,10 +25,10 @@ for.body: ; preds = %entry, %for.body %i.01 = phi i32 [ 0, %entry ], [ %inc, %for.body ] %idxprom = sext i32 %i.01 to i64 %arrayidx = getelementptr inbounds float, float* %0, i64 %idxprom - %3 = load float* %arrayidx, align 4 + %3 = load float, float* %arrayidx, align 4 %idxprom5 = sext i32 %i.01 to i64 %arrayidx6 = getelementptr inbounds float, float* %1, i64 %idxprom5 - %4 = load float* %arrayidx6, align 4 + %4 = load float, float* %arrayidx6, align 4 %add = fadd float %3, %4 %idxprom7 = sext i32 %i.01 to i64 %arrayidx8 = getelementptr inbounds float, float* %2, i64 %idxprom7 @@ -38,7 +38,7 @@ for.body: ; preds = %entry, %for.body br i1 %cmp, label %for.body, label %for.end for.end: ; preds = %for.body - %5 = load float* %2, align 4 + %5 = load float, float* %2, align 4 %conv10 = fptosi float %5 to i32 ret i32 %conv10 } diff --git a/llvm/test/Transforms/LoopVectorize/dbg.value.ll b/llvm/test/Transforms/LoopVectorize/dbg.value.ll index f3e75d84393..bf9a7967ecc 100644 --- a/llvm/test/Transforms/LoopVectorize/dbg.value.ll +++ b/llvm/test/Transforms/LoopVectorize/dbg.value.ll @@ -18,9 +18,9 @@ for.body: ;CHECK: load <4 x i32> %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32]* @B, i64 0, i64 %indvars.iv, !dbg !19 - %0 = load i32* %arrayidx, align 4, !dbg !19 + %0 = load i32, i32* %arrayidx, align 4, !dbg !19 %arrayidx2 = getelementptr inbounds [1024 x i32], [1024 x i32]* @C, i64 0, i64 %indvars.iv, !dbg !19 - %1 = load i32* %arrayidx2, align 4, !dbg !19 + %1 = load i32, i32* %arrayidx2, align 4, !dbg !19 %add = add nsw i32 %1, %0, !dbg !19 %arrayidx4 = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 0, i64 %indvars.iv, !dbg !19 store i32 %add, i32* %arrayidx4, align 4, !dbg !19 diff --git a/llvm/test/Transforms/LoopVectorize/debugloc.ll b/llvm/test/Transforms/LoopVectorize/debugloc.ll index a30ca62752b..c443ae482f0 100644 --- a/llvm/test/Transforms/LoopVectorize/debugloc.ll +++ b/llvm/test/Transforms/LoopVectorize/debugloc.ll @@ -9,7 +9,7 @@ target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f3 ; CHECK: vector.body ; CHECK: index {{.*}}, !dbg ![[LOC]] ; CHECK: getelementptr inbounds i32, i32* %a, {{.*}}, !dbg ![[LOC2:[0-9]+]] -; CHECK: load <2 x i32>* {{.*}}, !dbg ![[LOC2]] +; CHECK: load <2 x i32>, <2 x i32>* {{.*}}, !dbg ![[LOC2]] ; CHECK: add <2 x i32> {{.*}}, !dbg ![[LOC2]] ; CHECK: add i64 %index, 2, !dbg ![[LOC]] ; CHECK: icmp eq i64 %index.next, %end.idx.rnd.down, !dbg ![[LOC]] @@ -33,7 +33,7 @@ for.body: ; preds = %for.body.lr.ph, %fo %indvars.iv = phi i64 [ 0, %for.body.lr.ph ], [ %indvars.iv.next, %for.body ] %sum.05 = phi i32 [ 0, %for.body.lr.ph ], [ %add, %for.body ] %arrayidx = getelementptr inbounds i32, i32* %a, i64 %indvars.iv, !dbg !22 - %0 = load i32* %arrayidx, align 4, !dbg !22 + %0 = load i32, i32* %arrayidx, align 4, !dbg !22 %add = add i32 %0, %sum.05, !dbg !22 tail call void @llvm.dbg.value(metadata i32 %add.lcssa, i64 0, metadata !15, metadata !{}), !dbg !22 %indvars.iv.next = add i64 %indvars.iv, 1, !dbg !21 diff --git a/llvm/test/Transforms/LoopVectorize/duplicated-metadata.ll b/llvm/test/Transforms/LoopVectorize/duplicated-metadata.ll index e82d2804207..9f7cdef51fe 100644 --- a/llvm/test/Transforms/LoopVectorize/duplicated-metadata.ll +++ b/llvm/test/Transforms/LoopVectorize/duplicated-metadata.ll @@ -13,7 +13,7 @@ entry: for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds float, float* %a, i64 %indvars.iv - %p = load float* %arrayidx, align 4 + %p = load float, float* %arrayidx, align 4 %mul = fmul float %p, 2.000000e+00 store float %mul, float* %arrayidx, align 4 %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1 diff --git a/llvm/test/Transforms/LoopVectorize/ee-crash.ll b/llvm/test/Transforms/LoopVectorize/ee-crash.ll index 7ed1c669b74..0c387346f86 100644 --- a/llvm/test/Transforms/LoopVectorize/ee-crash.ll +++ b/llvm/test/Transforms/LoopVectorize/ee-crash.ll @@ -18,7 +18,7 @@ entry: for.body.i: ; preds = %entry, %for.body.i %__init.addr.05.i = phi i32 [ %add.i, %for.body.i ], [ 0, %entry ] %__first.addr.04.i = phi i32* [ %incdec.ptr.i, %for.body.i ], [ %A, %entry ] - %0 = load i32* %__first.addr.04.i, align 4 + %0 = load i32, i32* %__first.addr.04.i, align 4 %q1 = extractelement <2 x i32> %q, i32 %n %q2 = add nsw i32 %0, %q1 %add.i = add nsw i32 %q2, %__init.addr.05.i diff --git a/llvm/test/Transforms/LoopVectorize/exact.ll b/llvm/test/Transforms/LoopVectorize/exact.ll index 90bad3a99c2..0a4e0dc17df 100644 --- a/llvm/test/Transforms/LoopVectorize/exact.ll +++ b/llvm/test/Transforms/LoopVectorize/exact.ll @@ -12,7 +12,7 @@ entry: for.body: %iv = phi i64 [ 0, %entry ], [ %iv.next, %for.body ] %arrayidx = getelementptr inbounds i32, i32* %x, i64 %iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %conv1 = lshr exact i32 %0, 1 store i32 %conv1, i32* %arrayidx, align 4 %iv.next = add nuw nsw i64 %iv, 1 diff --git a/llvm/test/Transforms/LoopVectorize/flags.ll b/llvm/test/Transforms/LoopVectorize/flags.ll index a4e392f01a4..fcbc8741741 100644 --- a/llvm/test/Transforms/LoopVectorize/flags.ll +++ b/llvm/test/Transforms/LoopVectorize/flags.ll @@ -15,7 +15,7 @@ define i32 @flags1(i32 %n, i32* nocapture %A) nounwind uwtable ssp { .lr.ph: ; preds = %0, %.lr.ph %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 9, %0 ] %2 = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = mul nsw i32 %3, 3 store i32 %4, i32* %2, align 4 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -40,7 +40,7 @@ define i32 @flags2(i32 %n, i32* nocapture %A) nounwind uwtable ssp { .lr.ph: ; preds = %0, %.lr.ph %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 9, %0 ] %2 = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = mul i32 %3, 3 store i32 %4, i32* %2, align 4 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -67,7 +67,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %q.04 = phi float [ 0.000000e+00, %entry ], [ %add, %for.body ] %arrayidx = getelementptr inbounds float, float* %s, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %add = fadd fast float %q.04, %0 %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1 %exitcond = icmp eq i64 %indvars.iv.next, 256 diff --git a/llvm/test/Transforms/LoopVectorize/float-reduction.ll b/llvm/test/Transforms/LoopVectorize/float-reduction.ll index 1401bd9d7c7..1310b279f61 100644 --- a/llvm/test/Transforms/LoopVectorize/float-reduction.ll +++ b/llvm/test/Transforms/LoopVectorize/float-reduction.ll @@ -13,7 +13,7 @@ for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %sum.04 = phi float [ 0.000000e+00, %entry ], [ %add, %for.body ] %arrayidx = getelementptr inbounds float, float* %A, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %add = fadd fast float %sum.04, %0 %indvars.iv.next = add i64 %indvars.iv, 1 %lftr.wideiv = trunc i64 %indvars.iv.next to i32 @@ -35,7 +35,7 @@ for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %sum.04 = phi float [ 0.000000e+00, %entry ], [ %sub, %for.body ] %arrayidx = getelementptr inbounds float, float* %A, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %sub = fsub fast float %sum.04, %0 %indvars.iv.next = add i64 %indvars.iv, 1 %lftr.wideiv = trunc i64 %indvars.iv.next to i32 diff --git a/llvm/test/Transforms/LoopVectorize/funcall.ll b/llvm/test/Transforms/LoopVectorize/funcall.ll index 81158eeb483..35c2dfca4b2 100644 --- a/llvm/test/Transforms/LoopVectorize/funcall.ll +++ b/llvm/test/Transforms/LoopVectorize/funcall.ll @@ -17,7 +17,7 @@ entry: for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds double, double* %d, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %1 = tail call double @llvm.pow.f64(double %0, double %t) store double %1, double* %arrayidx, align 8 %indvars.iv.next = add i64 %indvars.iv, 1 diff --git a/llvm/test/Transforms/LoopVectorize/gcc-examples.ll b/llvm/test/Transforms/LoopVectorize/gcc-examples.ll index 9bc5cb78e03..18809018615 100644 --- a/llvm/test/Transforms/LoopVectorize/gcc-examples.ll +++ b/llvm/test/Transforms/LoopVectorize/gcc-examples.ll @@ -45,9 +45,9 @@ define void @example1() nounwind uwtable ssp { ; <label>:1 ; preds = %1, %0 %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %2 = getelementptr inbounds [2048 x i32], [2048 x i32]* @b, i64 0, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = getelementptr inbounds [2048 x i32], [2048 x i32]* @c, i64 0, i64 %indvars.iv - %5 = load i32* %4, align 4 + %5 = load i32, i32* %4, align 4 %6 = add nsw i32 %5, %3 %7 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv store i32 %6, i32* %7, align 4 @@ -96,9 +96,9 @@ define void @example2(i32 %n, i32 %x) nounwind uwtable ssp { %.02 = phi i32 [ %4, %.lr.ph ], [ %n, %.preheader ] %4 = add nsw i32 %.02, -1 %5 = getelementptr inbounds [2048 x i32], [2048 x i32]* @b, i64 0, i64 %indvars.iv - %6 = load i32* %5, align 4 + %6 = load i32, i32* %5, align 4 %7 = getelementptr inbounds [2048 x i32], [2048 x i32]* @c, i64 0, i64 %indvars.iv - %8 = load i32* %7, align 4 + %8 = load i32, i32* %7, align 4 %9 = and i32 %8, %6 %10 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv store i32 %9, i32* %10, align 4 @@ -129,7 +129,7 @@ define void @example3(i32 %n, i32* noalias nocapture %p, i32* noalias nocapture %.023 = phi i32* [ %3, %.lr.ph ], [ %q, %0 ] %2 = add nsw i32 %.05, -1 %3 = getelementptr inbounds i32, i32* %.023, i64 1 - %4 = load i32* %.023, align 16 + %4 = load i32, i32* %.023, align 16 %5 = getelementptr inbounds i32, i32* %.014, i64 1 store i32 %4, i32* %.014, align 16 %6 = icmp eq i32 %2, 0 @@ -162,7 +162,7 @@ define void @example4(i32 %n, i32* noalias nocapture %p, i32* noalias nocapture %.018 = phi i32* [ %8, %.lr.ph10 ], [ %p, %0 ] %.027 = phi i32* [ %5, %.lr.ph10 ], [ %q, %0 ] %5 = getelementptr inbounds i32, i32* %.027, i64 1 - %6 = load i32* %.027, align 16 + %6 = load i32, i32* %.027, align 16 %7 = add nsw i32 %6, 5 %8 = getelementptr inbounds i32, i32* %.018, i64 1 store i32 %7, i32* %.018, align 16 @@ -177,10 +177,10 @@ define void @example4(i32 %n, i32* noalias nocapture %p, i32* noalias nocapture %indvars.iv11 = phi i64 [ %indvars.iv.next12, %.lr.ph6 ], [ 0, %.preheader4 ] %indvars.iv.next12 = add i64 %indvars.iv11, 1 %11 = getelementptr inbounds [2048 x i32], [2048 x i32]* @b, i64 0, i64 %indvars.iv.next12 - %12 = load i32* %11, align 4 + %12 = load i32, i32* %11, align 4 %13 = add nsw i64 %indvars.iv11, 3 %14 = getelementptr inbounds [2048 x i32], [2048 x i32]* @c, i64 0, i64 %13 - %15 = load i32* %14, align 4 + %15 = load i32, i32* %14, align 4 %16 = add nsw i32 %15, %12 %17 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv11 store i32 %16, i32* %17, align 4 @@ -191,7 +191,7 @@ define void @example4(i32 %n, i32* noalias nocapture %p, i32* noalias nocapture .lr.ph: ; preds = %.preheader, %.lr.ph %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 0, %.preheader ] %18 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv - %19 = load i32* %18, align 4 + %19 = load i32, i32* %18, align 4 %20 = icmp sgt i32 %19, 4 %21 = select i1 %20, i32 4, i32 0 %22 = getelementptr inbounds [2048 x i32], [2048 x i32]* @b, i64 0, i64 %indvars.iv @@ -250,9 +250,9 @@ define i32 @example9() nounwind uwtable readonly ssp { %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %diff.01 = phi i32 [ 0, %0 ], [ %7, %1 ] %2 = getelementptr inbounds [1024 x i32], [1024 x i32]* @ub, i64 0, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = getelementptr inbounds [1024 x i32], [1024 x i32]* @uc, i64 0, i64 %indvars.iv - %5 = load i32* %4, align 4 + %5 = load i32, i32* %4, align 4 %6 = add i32 %3, %diff.01 %7 = sub i32 %6, %5 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -277,16 +277,16 @@ define void @example10a(i16* noalias nocapture %sa, i16* noalias nocapture %sb, ; <label>:1 ; preds = %1, %0 %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %2 = getelementptr inbounds i32, i32* %ib, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = getelementptr inbounds i32, i32* %ic, i64 %indvars.iv - %5 = load i32* %4, align 4 + %5 = load i32, i32* %4, align 4 %6 = add nsw i32 %5, %3 %7 = getelementptr inbounds i32, i32* %ia, i64 %indvars.iv store i32 %6, i32* %7, align 4 %8 = getelementptr inbounds i16, i16* %sb, i64 %indvars.iv - %9 = load i16* %8, align 2 + %9 = load i16, i16* %8, align 2 %10 = getelementptr inbounds i16, i16* %sc, i64 %indvars.iv - %11 = load i16* %10, align 2 + %11 = load i16, i16* %10, align 2 %12 = add i16 %11, %9 %13 = getelementptr inbounds i16, i16* %sa, i64 %indvars.iv store i16 %12, i16* %13, align 2 @@ -310,7 +310,7 @@ define void @example10b(i16* noalias nocapture %sa, i16* noalias nocapture %sb, ; <label>:1 ; preds = %1, %0 %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %2 = getelementptr inbounds i16, i16* %sb, i64 %indvars.iv - %3 = load i16* %2, align 2 + %3 = load i16, i16* %2, align 2 %4 = sext i16 %3 to i32 %5 = getelementptr inbounds i32, i32* %ia, i64 %indvars.iv store i32 %4, i32* %5, align 4 @@ -341,14 +341,14 @@ define void @example11() nounwind uwtable ssp { %2 = shl nsw i64 %indvars.iv, 1 %3 = or i64 %2, 1 %4 = getelementptr inbounds [2048 x i32], [2048 x i32]* @b, i64 0, i64 %3 - %5 = load i32* %4, align 4 + %5 = load i32, i32* %4, align 4 %6 = getelementptr inbounds [2048 x i32], [2048 x i32]* @c, i64 0, i64 %3 - %7 = load i32* %6, align 4 + %7 = load i32, i32* %6, align 4 %8 = mul nsw i32 %7, %5 %9 = getelementptr inbounds [2048 x i32], [2048 x i32]* @b, i64 0, i64 %2 - %10 = load i32* %9, align 8 + %10 = load i32, i32* %9, align 8 %11 = getelementptr inbounds [2048 x i32], [2048 x i32]* @c, i64 0, i64 %2 - %12 = load i32* %11, align 8 + %12 = load i32, i32* %11, align 8 %13 = mul nsw i32 %12, %10 %14 = sub nsw i32 %8, %13 %15 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv @@ -397,18 +397,18 @@ define void @example13(i32** nocapture %A, i32** nocapture %B, i32* nocapture %o .preheader: ; preds = %14, %0 %indvars.iv4 = phi i64 [ 0, %0 ], [ %indvars.iv.next5, %14 ] %1 = getelementptr inbounds i32*, i32** %A, i64 %indvars.iv4 - %2 = load i32** %1, align 8 + %2 = load i32*, i32** %1, align 8 %3 = getelementptr inbounds i32*, i32** %B, i64 %indvars.iv4 - %4 = load i32** %3, align 8 + %4 = load i32*, i32** %3, align 8 br label %5 ; <label>:5 ; preds = %.preheader, %5 %indvars.iv = phi i64 [ 0, %.preheader ], [ %indvars.iv.next, %5 ] %diff.02 = phi i32 [ 0, %.preheader ], [ %11, %5 ] %6 = getelementptr inbounds i32, i32* %2, i64 %indvars.iv - %7 = load i32* %6, align 4 + %7 = load i32, i32* %6, align 4 %8 = getelementptr inbounds i32, i32* %4, i64 %indvars.iv - %9 = load i32* %8, align 4 + %9 = load i32, i32* %8, align 4 %10 = add i32 %7, %diff.02 %11 = sub i32 %10, %9 %indvars.iv.next = add i64 %indvars.iv, 8 @@ -445,13 +445,13 @@ define void @example14(i32** nocapture %in, i32** nocapture %coeff, i32* nocaptu %indvars.iv = phi i64 [ 0, %.preheader ], [ %indvars.iv.next, %0 ] %sum.12 = phi i32 [ %sum.05, %.preheader ], [ %10, %0 ] %1 = getelementptr inbounds i32*, i32** %in, i64 %indvars.iv - %2 = load i32** %1, align 8 + %2 = load i32*, i32** %1, align 8 %3 = getelementptr inbounds i32, i32* %2, i64 %indvars.iv7 - %4 = load i32* %3, align 4 + %4 = load i32, i32* %3, align 4 %5 = getelementptr inbounds i32*, i32** %coeff, i64 %indvars.iv - %6 = load i32** %5, align 8 + %6 = load i32*, i32** %5, align 8 %7 = getelementptr inbounds i32, i32* %6, i64 %indvars.iv7 - %8 = load i32* %7, align 4 + %8 = load i32, i32* %7, align 4 %9 = mul nsw i32 %8, %4 %10 = add nsw i32 %9, %sum.12 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -479,13 +479,13 @@ define void @example14(i32** nocapture %in, i32** nocapture %coeff, i32* nocaptu %sum.12.1 = phi i32 [ %sum.05.1, %.preheader.1 ], [ %23, %12 ] %13 = add nsw i64 %indvars.iv.1, 1 %14 = getelementptr inbounds i32*, i32** %in, i64 %13 - %15 = load i32** %14, align 8 + %15 = load i32*, i32** %14, align 8 %16 = getelementptr inbounds i32, i32* %15, i64 %indvars.iv7.1 - %17 = load i32* %16, align 4 + %17 = load i32, i32* %16, align 4 %18 = getelementptr inbounds i32*, i32** %coeff, i64 %indvars.iv.1 - %19 = load i32** %18, align 8 + %19 = load i32*, i32** %18, align 8 %20 = getelementptr inbounds i32, i32* %19, i64 %indvars.iv7.1 - %21 = load i32* %20, align 4 + %21 = load i32, i32* %20, align 4 %22 = mul nsw i32 %21, %17 %23 = add nsw i32 %22, %sum.12.1 %lftr.wideiv.1 = trunc i64 %13 to i32 @@ -513,13 +513,13 @@ define void @example14(i32** nocapture %in, i32** nocapture %coeff, i32* nocaptu %sum.12.2 = phi i32 [ %sum.05.2, %.preheader.2 ], [ %37, %26 ] %27 = add nsw i64 %indvars.iv.2, 2 %28 = getelementptr inbounds i32*, i32** %in, i64 %27 - %29 = load i32** %28, align 8 + %29 = load i32*, i32** %28, align 8 %30 = getelementptr inbounds i32, i32* %29, i64 %indvars.iv7.2 - %31 = load i32* %30, align 4 + %31 = load i32, i32* %30, align 4 %32 = getelementptr inbounds i32*, i32** %coeff, i64 %indvars.iv.2 - %33 = load i32** %32, align 8 + %33 = load i32*, i32** %32, align 8 %34 = getelementptr inbounds i32, i32* %33, i64 %indvars.iv7.2 - %35 = load i32* %34, align 4 + %35 = load i32, i32* %34, align 4 %36 = mul nsw i32 %35, %31 %37 = add nsw i32 %36, %sum.12.2 %indvars.iv.next.2 = add i64 %indvars.iv.2, 1 @@ -548,13 +548,13 @@ define void @example14(i32** nocapture %in, i32** nocapture %coeff, i32* nocaptu %sum.12.3 = phi i32 [ %sum.05.3, %.preheader.3 ], [ %51, %40 ] %41 = add nsw i64 %indvars.iv.3, 3 %42 = getelementptr inbounds i32*, i32** %in, i64 %41 - %43 = load i32** %42, align 8 + %43 = load i32*, i32** %42, align 8 %44 = getelementptr inbounds i32, i32* %43, i64 %indvars.iv7.3 - %45 = load i32* %44, align 4 + %45 = load i32, i32* %44, align 4 %46 = getelementptr inbounds i32*, i32** %coeff, i64 %indvars.iv.3 - %47 = load i32** %46, align 8 + %47 = load i32*, i32** %46, align 8 %48 = getelementptr inbounds i32, i32* %47, i64 %indvars.iv7.3 - %49 = load i32* %48, align 4 + %49 = load i32, i32* %48, align 4 %50 = mul nsw i32 %49, %45 %51 = add nsw i32 %50, %sum.12.3 %indvars.iv.next.3 = add i64 %indvars.iv.3, 1 @@ -591,7 +591,7 @@ define i32 @example21(i32* nocapture %b, i32 %n) nounwind uwtable readonly ssp { %a.02 = phi i32 [ 0, %.lr.ph ], [ %6, %3 ] %indvars.iv.next = add i64 %indvars.iv, -1 %4 = getelementptr inbounds i32, i32* %b, i64 %indvars.iv.next - %5 = load i32* %4, align 4 + %5 = load i32, i32* %4, align 4 %6 = add nsw i32 %5, %a.02 %7 = trunc i64 %indvars.iv.next to i32 %8 = icmp sgt i32 %7, 0 @@ -613,7 +613,7 @@ define void @example23(i16* nocapture %src, i32* nocapture %dst) nounwind uwtabl %.013 = phi i32* [ %dst, %0 ], [ %6, %1 ] %i.02 = phi i32 [ 0, %0 ], [ %7, %1 ] %2 = getelementptr inbounds i16, i16* %.04, i64 1 - %3 = load i16* %.04, align 2 + %3 = load i16, i16* %.04, align 2 %4 = zext i16 %3 to i32 %5 = shl nuw nsw i32 %4, 7 %6 = getelementptr inbounds i32, i32* %.013, i64 1 @@ -635,9 +635,9 @@ define void @example24(i16 signext %x, i16 signext %y) nounwind uwtable ssp { ; <label>:1 ; preds = %1, %0 %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %2 = getelementptr inbounds [1024 x float], [1024 x float]* @fa, i64 0, i64 %indvars.iv - %3 = load float* %2, align 4 + %3 = load float, float* %2, align 4 %4 = getelementptr inbounds [1024 x float], [1024 x float]* @fb, i64 0, i64 %indvars.iv - %5 = load float* %4, align 4 + %5 = load float, float* %4, align 4 %6 = fcmp olt float %3, %5 %x.y = select i1 %6, i16 %x, i16 %y %7 = sext i16 %x.y to i32 @@ -662,14 +662,14 @@ define void @example25() nounwind uwtable ssp { ; <label>:1 ; preds = %1, %0 %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %2 = getelementptr inbounds [1024 x float], [1024 x float]* @da, i64 0, i64 %indvars.iv - %3 = load float* %2, align 4 + %3 = load float, float* %2, align 4 %4 = getelementptr inbounds [1024 x float], [1024 x float]* @db, i64 0, i64 %indvars.iv - %5 = load float* %4, align 4 + %5 = load float, float* %4, align 4 %6 = fcmp olt float %3, %5 %7 = getelementptr inbounds [1024 x float], [1024 x float]* @dc, i64 0, i64 %indvars.iv - %8 = load float* %7, align 4 + %8 = load float, float* %7, align 4 %9 = getelementptr inbounds [1024 x float], [1024 x float]* @dd, i64 0, i64 %indvars.iv - %10 = load float* %9, align 4 + %10 = load float, float* %9, align 4 %11 = fcmp olt float %8, %10 %12 = and i1 %6, %11 %13 = zext i1 %12 to i32 diff --git a/llvm/test/Transforms/LoopVectorize/global_alias.ll b/llvm/test/Transforms/LoopVectorize/global_alias.ll index ef232669f8f..b2ac4ca04ec 100644 --- a/llvm/test/Transforms/LoopVectorize/global_alias.ll +++ b/llvm/test/Transforms/LoopVectorize/global_alias.ll @@ -35,31 +35,31 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %arrayidx = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 2), i32 0, i32 %1 - %2 = load i32* %arrayidx, align 4 - %3 = load i32* %a.addr, align 4 + %2 = load i32, i32* %arrayidx, align 4 + %3 = load i32, i32* %a.addr, align 4 %add = add nsw i32 %2, %3 - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %arrayidx1 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %4 store i32 %add, i32* %arrayidx1, align 4 br label %for.inc for.inc: ; preds = %for.body - %5 = load i32* %i, align 4 + %5 = load i32, i32* %i, align 4 %inc = add nsw i32 %5, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %6 = load i32* %a.addr, align 4 + %6 = load i32, i32* %a.addr, align 4 %arrayidx2 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %6 - %7 = load i32* %arrayidx2, align 4 + %7 = load i32, i32* %arrayidx2, align 4 ret i32 %7 } @@ -83,32 +83,32 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 90 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %add = add nsw i32 %1, 10 %arrayidx = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 2), i32 0, i32 %add - %2 = load i32* %arrayidx, align 4 - %3 = load i32* %a.addr, align 4 + %2 = load i32, i32* %arrayidx, align 4 + %3 = load i32, i32* %a.addr, align 4 %add1 = add nsw i32 %2, %3 - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %arrayidx2 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %4 store i32 %add1, i32* %arrayidx2, align 4 br label %for.inc for.inc: ; preds = %for.body - %5 = load i32* %i, align 4 + %5 = load i32, i32* %i, align 4 %inc = add nsw i32 %5, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %6 = load i32* %a.addr, align 4 + %6 = load i32, i32* %a.addr, align 4 %arrayidx3 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %6 - %7 = load i32* %arrayidx3, align 4 + %7 = load i32, i32* %arrayidx3, align 4 ret i32 %7 } @@ -132,32 +132,32 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %arrayidx = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 2), i32 0, i32 %1 - %2 = load i32* %arrayidx, align 4 - %3 = load i32* %a.addr, align 4 + %2 = load i32, i32* %arrayidx, align 4 + %3 = load i32, i32* %a.addr, align 4 %add = add nsw i32 %2, %3 - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %add1 = add nsw i32 %4, 10 %arrayidx2 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %add1 store i32 %add, i32* %arrayidx2, align 4 br label %for.inc for.inc: ; preds = %for.body - %5 = load i32* %i, align 4 + %5 = load i32, i32* %i, align 4 %inc = add nsw i32 %5, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %6 = load i32* %a.addr, align 4 + %6 = load i32, i32* %a.addr, align 4 %arrayidx3 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %6 - %7 = load i32* %arrayidx3, align 4 + %7 = load i32, i32* %arrayidx3, align 4 ret i32 %7 } @@ -184,34 +184,34 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32** @PB, align 4 - %2 = load i32* %i, align 4 + %1 = load i32*, i32** @PB, align 4 + %2 = load i32, i32* %i, align 4 %add.ptr = getelementptr inbounds i32, i32* %1, i32 %2 - %3 = load i32* %add.ptr, align 4 - %4 = load i32* %a.addr, align 4 + %3 = load i32, i32* %add.ptr, align 4 + %4 = load i32, i32* %a.addr, align 4 %add = add nsw i32 %3, %4 - %5 = load i32** @PA, align 4 - %6 = load i32* %i, align 4 + %5 = load i32*, i32** @PA, align 4 + %6 = load i32, i32* %i, align 4 %add.ptr1 = getelementptr inbounds i32, i32* %5, i32 %6 store i32 %add, i32* %add.ptr1, align 4 br label %for.inc for.inc: ; preds = %for.body - %7 = load i32* %i, align 4 + %7 = load i32, i32* %i, align 4 %inc = add nsw i32 %7, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %8 = load i32** @PA, align 4 - %9 = load i32* %a.addr, align 4 + %8 = load i32*, i32** @PA, align 4 + %9 = load i32, i32* %a.addr, align 4 %add.ptr2 = getelementptr inbounds i32, i32* %8, i32 %9 - %10 = load i32* %add.ptr2, align 4 + %10 = load i32, i32* %add.ptr2, align 4 ret i32 %10 } @@ -237,37 +237,37 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 - %2 = load i32* %N, align 4 + %1 = load i32, i32* %i, align 4 + %2 = load i32, i32* %N, align 4 %arrayidx = getelementptr inbounds [100 x [100 x i32]], [100 x [100 x i32]]* getelementptr inbounds (%struct.anon.0* @Bar, i32 0, i32 2), i32 0, i32 %2 %arrayidx1 = getelementptr inbounds [100 x i32], [100 x i32]* %arrayidx, i32 0, i32 %1 - %3 = load i32* %arrayidx1, align 4 - %4 = load i32* %a.addr, align 4 + %3 = load i32, i32* %arrayidx1, align 4 + %4 = load i32, i32* %a.addr, align 4 %add = add nsw i32 %3, %4 - %5 = load i32* %i, align 4 - %6 = load i32* %N, align 4 + %5 = load i32, i32* %i, align 4 + %6 = load i32, i32* %N, align 4 %arrayidx2 = getelementptr inbounds [100 x [100 x i32]], [100 x [100 x i32]]* getelementptr inbounds (%struct.anon.0* @Bar, i32 0, i32 0), i32 0, i32 %6 %arrayidx3 = getelementptr inbounds [100 x i32], [100 x i32]* %arrayidx2, i32 0, i32 %5 store i32 %add, i32* %arrayidx3, align 4 br label %for.inc for.inc: ; preds = %for.body - %7 = load i32* %i, align 4 + %7 = load i32, i32* %i, align 4 %inc = add nsw i32 %7, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %8 = load i32* %a.addr, align 4 - %9 = load i32* %N, align 4 + %8 = load i32, i32* %a.addr, align 4 + %9 = load i32, i32* %N, align 4 %arrayidx4 = getelementptr inbounds [100 x [100 x i32]], [100 x [100 x i32]]* getelementptr inbounds (%struct.anon.0* @Bar, i32 0, i32 0), i32 0, i32 %9 %arrayidx5 = getelementptr inbounds [100 x i32], [100 x i32]* %arrayidx4, i32 0, i32 %8 - %10 = load i32* %arrayidx5, align 4 + %10 = load i32, i32* %arrayidx5, align 4 ret i32 %10 } @@ -293,38 +293,38 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 - %2 = load i32* %N, align 4 + %1 = load i32, i32* %i, align 4 + %2 = load i32, i32* %N, align 4 %add = add nsw i32 %2, 1 %arrayidx = getelementptr inbounds [100 x [100 x i32]], [100 x [100 x i32]]* getelementptr inbounds (%struct.anon.0* @Bar, i32 0, i32 0), i32 0, i32 %add %arrayidx1 = getelementptr inbounds [100 x i32], [100 x i32]* %arrayidx, i32 0, i32 %1 - %3 = load i32* %arrayidx1, align 4 - %4 = load i32* %a.addr, align 4 + %3 = load i32, i32* %arrayidx1, align 4 + %4 = load i32, i32* %a.addr, align 4 %add2 = add nsw i32 %3, %4 - %5 = load i32* %i, align 4 - %6 = load i32* %N, align 4 + %5 = load i32, i32* %i, align 4 + %6 = load i32, i32* %N, align 4 %arrayidx3 = getelementptr inbounds [100 x [100 x i32]], [100 x [100 x i32]]* getelementptr inbounds (%struct.anon.0* @Bar, i32 0, i32 0), i32 0, i32 %6 %arrayidx4 = getelementptr inbounds [100 x i32], [100 x i32]* %arrayidx3, i32 0, i32 %5 store i32 %add2, i32* %arrayidx4, align 4 br label %for.inc for.inc: ; preds = %for.body - %7 = load i32* %i, align 4 + %7 = load i32, i32* %i, align 4 %inc = add nsw i32 %7, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %8 = load i32* %a.addr, align 4 - %9 = load i32* %N, align 4 + %8 = load i32, i32* %a.addr, align 4 + %9 = load i32, i32* %N, align 4 %arrayidx5 = getelementptr inbounds [100 x [100 x i32]], [100 x [100 x i32]]* getelementptr inbounds (%struct.anon.0* @Bar, i32 0, i32 0), i32 0, i32 %9 %arrayidx6 = getelementptr inbounds [100 x i32], [100 x i32]* %arrayidx5, i32 0, i32 %8 - %10 = load i32* %arrayidx6, align 4 + %10 = load i32, i32* %arrayidx6, align 4 ret i32 %10 } @@ -347,19 +347,19 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %sub = sub nsw i32 100, %1 %sub1 = sub nsw i32 %sub, 1 %arrayidx = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 2), i32 0, i32 %sub1 - %2 = load i32* %arrayidx, align 4 - %3 = load i32* %a.addr, align 4 + %2 = load i32, i32* %arrayidx, align 4 + %3 = load i32, i32* %a.addr, align 4 %add = add nsw i32 %2, %3 - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %sub2 = sub nsw i32 100, %4 %sub3 = sub nsw i32 %sub2, 1 %arrayidx4 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %sub3 @@ -367,15 +367,15 @@ for.body: ; preds = %for.cond br label %for.inc for.inc: ; preds = %for.body - %5 = load i32* %i, align 4 + %5 = load i32, i32* %i, align 4 %inc = add nsw i32 %5, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %6 = load i32* %a.addr, align 4 + %6 = load i32, i32* %a.addr, align 4 %arrayidx5 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %6 - %7 = load i32* %arrayidx5, align 4 + %7 = load i32, i32* %arrayidx5, align 4 ret i32 %7 } @@ -399,19 +399,19 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 90 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %sub = sub nsw i32 100, %1 %sub1 = sub nsw i32 %sub, 10 %arrayidx = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 2), i32 0, i32 %sub1 - %2 = load i32* %arrayidx, align 4 - %3 = load i32* %a.addr, align 4 + %2 = load i32, i32* %arrayidx, align 4 + %3 = load i32, i32* %a.addr, align 4 %add = add nsw i32 %2, %3 - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %sub2 = sub nsw i32 100, %4 %sub3 = sub nsw i32 %sub2, 1 %arrayidx4 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %sub3 @@ -419,15 +419,15 @@ for.body: ; preds = %for.cond br label %for.inc for.inc: ; preds = %for.body - %5 = load i32* %i, align 4 + %5 = load i32, i32* %i, align 4 %inc = add nsw i32 %5, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %6 = load i32* %a.addr, align 4 + %6 = load i32, i32* %a.addr, align 4 %arrayidx5 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %6 - %7 = load i32* %arrayidx5, align 4 + %7 = load i32, i32* %arrayidx5, align 4 ret i32 %7 } @@ -451,19 +451,19 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %sub = sub nsw i32 100, %1 %sub1 = sub nsw i32 %sub, 1 %arrayidx = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 2), i32 0, i32 %sub1 - %2 = load i32* %arrayidx, align 4 - %3 = load i32* %a.addr, align 4 + %2 = load i32, i32* %arrayidx, align 4 + %3 = load i32, i32* %a.addr, align 4 %add = add nsw i32 %2, %3 - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %sub2 = sub nsw i32 100, %4 %sub3 = sub nsw i32 %sub2, 10 %arrayidx4 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %sub3 @@ -471,15 +471,15 @@ for.body: ; preds = %for.cond br label %for.inc for.inc: ; preds = %for.body - %5 = load i32* %i, align 4 + %5 = load i32, i32* %i, align 4 %inc = add nsw i32 %5, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %6 = load i32* %a.addr, align 4 + %6 = load i32, i32* %a.addr, align 4 %arrayidx5 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %6 - %7 = load i32* %arrayidx5, align 4 + %7 = load i32, i32* %arrayidx5, align 4 ret i32 %7 } @@ -506,23 +506,23 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32** @PB, align 4 + %1 = load i32*, i32** @PB, align 4 %add.ptr = getelementptr inbounds i32, i32* %1, i32 100 - %2 = load i32* %i, align 4 + %2 = load i32, i32* %i, align 4 %idx.neg = sub i32 0, %2 %add.ptr1 = getelementptr inbounds i32, i32* %add.ptr, i32 %idx.neg %add.ptr2 = getelementptr inbounds i32, i32* %add.ptr1, i32 -1 - %3 = load i32* %add.ptr2, align 4 - %4 = load i32* %a.addr, align 4 + %3 = load i32, i32* %add.ptr2, align 4 + %4 = load i32, i32* %a.addr, align 4 %add = add nsw i32 %3, %4 - %5 = load i32** @PA, align 4 + %5 = load i32*, i32** @PA, align 4 %add.ptr3 = getelementptr inbounds i32, i32* %5, i32 100 - %6 = load i32* %i, align 4 + %6 = load i32, i32* %i, align 4 %idx.neg4 = sub i32 0, %6 %add.ptr5 = getelementptr inbounds i32, i32* %add.ptr3, i32 %idx.neg4 %add.ptr6 = getelementptr inbounds i32, i32* %add.ptr5, i32 -1 @@ -530,16 +530,16 @@ for.body: ; preds = %for.cond br label %for.inc for.inc: ; preds = %for.body - %7 = load i32* %i, align 4 + %7 = load i32, i32* %i, align 4 %inc = add nsw i32 %7, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %8 = load i32** @PA, align 4 - %9 = load i32* %a.addr, align 4 + %8 = load i32*, i32** @PA, align 4 + %9 = load i32, i32* %a.addr, align 4 %add.ptr7 = getelementptr inbounds i32, i32* %8, i32 %9 - %10 = load i32* %add.ptr7, align 4 + %10 = load i32, i32* %add.ptr7, align 4 ret i32 %10 } @@ -565,41 +565,41 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %sub = sub nsw i32 100, %1 %sub1 = sub nsw i32 %sub, 1 - %2 = load i32* %N, align 4 + %2 = load i32, i32* %N, align 4 %arrayidx = getelementptr inbounds [100 x [100 x i32]], [100 x [100 x i32]]* getelementptr inbounds (%struct.anon.0* @Bar, i32 0, i32 2), i32 0, i32 %2 %arrayidx2 = getelementptr inbounds [100 x i32], [100 x i32]* %arrayidx, i32 0, i32 %sub1 - %3 = load i32* %arrayidx2, align 4 - %4 = load i32* %a.addr, align 4 + %3 = load i32, i32* %arrayidx2, align 4 + %4 = load i32, i32* %a.addr, align 4 %add = add nsw i32 %3, %4 - %5 = load i32* %i, align 4 + %5 = load i32, i32* %i, align 4 %sub3 = sub nsw i32 100, %5 %sub4 = sub nsw i32 %sub3, 1 - %6 = load i32* %N, align 4 + %6 = load i32, i32* %N, align 4 %arrayidx5 = getelementptr inbounds [100 x [100 x i32]], [100 x [100 x i32]]* getelementptr inbounds (%struct.anon.0* @Bar, i32 0, i32 0), i32 0, i32 %6 %arrayidx6 = getelementptr inbounds [100 x i32], [100 x i32]* %arrayidx5, i32 0, i32 %sub4 store i32 %add, i32* %arrayidx6, align 4 br label %for.inc for.inc: ; preds = %for.body - %7 = load i32* %i, align 4 + %7 = load i32, i32* %i, align 4 %inc = add nsw i32 %7, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %8 = load i32* %a.addr, align 4 - %9 = load i32* %N, align 4 + %8 = load i32, i32* %a.addr, align 4 + %9 = load i32, i32* %N, align 4 %arrayidx7 = getelementptr inbounds [100 x [100 x i32]], [100 x [100 x i32]]* getelementptr inbounds (%struct.anon.0* @Bar, i32 0, i32 0), i32 0, i32 %9 %arrayidx8 = getelementptr inbounds [100 x i32], [100 x i32]* %arrayidx7, i32 0, i32 %8 - %10 = load i32* %arrayidx8, align 4 + %10 = load i32, i32* %arrayidx8, align 4 ret i32 %10 } @@ -625,42 +625,42 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %sub = sub nsw i32 100, %1 %sub1 = sub nsw i32 %sub, 1 - %2 = load i32* %N, align 4 + %2 = load i32, i32* %N, align 4 %add = add nsw i32 %2, 1 %arrayidx = getelementptr inbounds [100 x [100 x i32]], [100 x [100 x i32]]* getelementptr inbounds (%struct.anon.0* @Bar, i32 0, i32 0), i32 0, i32 %add %arrayidx2 = getelementptr inbounds [100 x i32], [100 x i32]* %arrayidx, i32 0, i32 %sub1 - %3 = load i32* %arrayidx2, align 4 - %4 = load i32* %a.addr, align 4 + %3 = load i32, i32* %arrayidx2, align 4 + %4 = load i32, i32* %a.addr, align 4 %add3 = add nsw i32 %3, %4 - %5 = load i32* %i, align 4 + %5 = load i32, i32* %i, align 4 %sub4 = sub nsw i32 100, %5 %sub5 = sub nsw i32 %sub4, 1 - %6 = load i32* %N, align 4 + %6 = load i32, i32* %N, align 4 %arrayidx6 = getelementptr inbounds [100 x [100 x i32]], [100 x [100 x i32]]* getelementptr inbounds (%struct.anon.0* @Bar, i32 0, i32 0), i32 0, i32 %6 %arrayidx7 = getelementptr inbounds [100 x i32], [100 x i32]* %arrayidx6, i32 0, i32 %sub5 store i32 %add3, i32* %arrayidx7, align 4 br label %for.inc for.inc: ; preds = %for.body - %7 = load i32* %i, align 4 + %7 = load i32, i32* %i, align 4 %inc = add nsw i32 %7, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %8 = load i32* %a.addr, align 4 - %9 = load i32* %N, align 4 + %8 = load i32, i32* %a.addr, align 4 + %9 = load i32, i32* %N, align 4 %arrayidx8 = getelementptr inbounds [100 x [100 x i32]], [100 x [100 x i32]]* getelementptr inbounds (%struct.anon.0* @Bar, i32 0, i32 0), i32 0, i32 %9 %arrayidx9 = getelementptr inbounds [100 x i32], [100 x i32]* %arrayidx8, i32 0, i32 %8 - %10 = load i32* %arrayidx9, align 4 + %10 = load i32, i32* %arrayidx9, align 4 ret i32 %10 } @@ -684,32 +684,32 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %add = add nsw i32 %1, 4 %arrayidx = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %add - %2 = load i32* %arrayidx, align 4 - %3 = load i32* %a.addr, align 4 + %2 = load i32, i32* %arrayidx, align 4 + %3 = load i32, i32* %a.addr, align 4 %add1 = add nsw i32 %2, %3 - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %arrayidx2 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %4 store i32 %add1, i32* %arrayidx2, align 4 br label %for.inc for.inc: ; preds = %for.body - %5 = load i32* %i, align 4 + %5 = load i32, i32* %i, align 4 %inc = add nsw i32 %5, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %6 = load i32* %a.addr, align 4 + %6 = load i32, i32* %a.addr, align 4 %arrayidx3 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %6 - %7 = load i32* %arrayidx3, align 4 + %7 = load i32, i32* %arrayidx3, align 4 ret i32 %7 } @@ -733,19 +733,19 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %sub = sub nsw i32 100, %1 %sub1 = sub nsw i32 %sub, 5 %arrayidx = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %sub1 - %2 = load i32* %arrayidx, align 4 - %3 = load i32* %a.addr, align 4 + %2 = load i32, i32* %arrayidx, align 4 + %3 = load i32, i32* %a.addr, align 4 %add = add nsw i32 %2, %3 - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %sub2 = sub nsw i32 100, %4 %sub3 = sub nsw i32 %sub2, 1 %arrayidx4 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %sub3 @@ -753,15 +753,15 @@ for.body: ; preds = %for.cond br label %for.inc for.inc: ; preds = %for.body - %5 = load i32* %i, align 4 + %5 = load i32, i32* %i, align 4 %inc = add nsw i32 %5, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %6 = load i32* %a.addr, align 4 + %6 = load i32, i32* %a.addr, align 4 %arrayidx5 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %6 - %7 = load i32* %arrayidx5, align 4 + %7 = load i32, i32* %arrayidx5, align 4 ret i32 %7 } @@ -789,33 +789,33 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %sub = sub nsw i32 100, %1 %sub1 = sub nsw i32 %sub, 1 %arrayidx = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 2), i32 0, i32 %sub1 - %2 = load i32* %arrayidx, align 4 - %3 = load i32* %a.addr, align 4 + %2 = load i32, i32* %arrayidx, align 4 + %3 = load i32, i32* %a.addr, align 4 %add = add nsw i32 %2, %3 - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %arrayidx2 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %4 store i32 %add, i32* %arrayidx2, align 4 br label %for.inc for.inc: ; preds = %for.body - %5 = load i32* %i, align 4 + %5 = load i32, i32* %i, align 4 %inc = add nsw i32 %5, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %6 = load i32* %a.addr, align 4 + %6 = load i32, i32* %a.addr, align 4 %arrayidx3 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %6 - %7 = load i32* %arrayidx3, align 4 + %7 = load i32, i32* %arrayidx3, align 4 ret i32 %7 } @@ -839,17 +839,17 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %arrayidx = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 2), i32 0, i32 %1 - %2 = load i32* %arrayidx, align 4 - %3 = load i32* %a.addr, align 4 + %2 = load i32, i32* %arrayidx, align 4 + %3 = load i32, i32* %a.addr, align 4 %add = add nsw i32 %2, %3 - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %sub = sub nsw i32 100, %4 %sub1 = sub nsw i32 %sub, 1 %arrayidx2 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %sub1 @@ -857,15 +857,15 @@ for.body: ; preds = %for.cond br label %for.inc for.inc: ; preds = %for.body - %5 = load i32* %i, align 4 + %5 = load i32, i32* %i, align 4 %inc = add nsw i32 %5, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %6 = load i32* %a.addr, align 4 + %6 = load i32, i32* %a.addr, align 4 %arrayidx3 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %6 - %7 = load i32* %arrayidx3, align 4 + %7 = load i32, i32* %arrayidx3, align 4 ret i32 %7 } @@ -889,37 +889,37 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32** @PB, align 4 + %1 = load i32*, i32** @PB, align 4 %add.ptr = getelementptr inbounds i32, i32* %1, i32 100 - %2 = load i32* %i, align 4 + %2 = load i32, i32* %i, align 4 %idx.neg = sub i32 0, %2 %add.ptr1 = getelementptr inbounds i32, i32* %add.ptr, i32 %idx.neg %add.ptr2 = getelementptr inbounds i32, i32* %add.ptr1, i32 -1 - %3 = load i32* %add.ptr2, align 4 - %4 = load i32* %a.addr, align 4 + %3 = load i32, i32* %add.ptr2, align 4 + %4 = load i32, i32* %a.addr, align 4 %add = add nsw i32 %3, %4 - %5 = load i32** @PA, align 4 - %6 = load i32* %i, align 4 + %5 = load i32*, i32** @PA, align 4 + %6 = load i32, i32* %i, align 4 %add.ptr3 = getelementptr inbounds i32, i32* %5, i32 %6 store i32 %add, i32* %add.ptr3, align 4 br label %for.inc for.inc: ; preds = %for.body - %7 = load i32* %i, align 4 + %7 = load i32, i32* %i, align 4 %inc = add nsw i32 %7, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %8 = load i32** @PA, align 4 - %9 = load i32* %a.addr, align 4 + %8 = load i32*, i32** @PA, align 4 + %9 = load i32, i32* %a.addr, align 4 %add.ptr4 = getelementptr inbounds i32, i32* %8, i32 %9 - %10 = load i32* %add.ptr4, align 4 + %10 = load i32, i32* %add.ptr4, align 4 ret i32 %10 } @@ -946,34 +946,34 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %sub = sub nsw i32 100, %1 %sub1 = sub nsw i32 %sub, 1 %arrayidx = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 2), i32 0, i32 %sub1 - %2 = load i32* %arrayidx, align 4 - %3 = load i32* %a.addr, align 4 + %2 = load i32, i32* %arrayidx, align 4 + %3 = load i32, i32* %a.addr, align 4 %add = add nsw i32 %2, %3 - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %add2 = add nsw i32 %4, 10 %arrayidx3 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %add2 store i32 %add, i32* %arrayidx3, align 4 br label %for.inc for.inc: ; preds = %for.body - %5 = load i32* %i, align 4 + %5 = load i32, i32* %i, align 4 %inc = add nsw i32 %5, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %6 = load i32* %a.addr, align 4 + %6 = load i32, i32* %a.addr, align 4 %arrayidx4 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %6 - %7 = load i32* %arrayidx4, align 4 + %7 = load i32, i32* %arrayidx4, align 4 ret i32 %7 } @@ -996,33 +996,33 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %sub = sub nsw i32 100, %1 %sub1 = sub nsw i32 %sub, 10 %arrayidx = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 2), i32 0, i32 %sub1 - %2 = load i32* %arrayidx, align 4 - %3 = load i32* %a.addr, align 4 + %2 = load i32, i32* %arrayidx, align 4 + %3 = load i32, i32* %a.addr, align 4 %add = add nsw i32 %2, %3 - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %arrayidx2 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %4 store i32 %add, i32* %arrayidx2, align 4 br label %for.inc for.inc: ; preds = %for.body - %5 = load i32* %i, align 4 + %5 = load i32, i32* %i, align 4 %inc = add nsw i32 %5, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %6 = load i32* %a.addr, align 4 + %6 = load i32, i32* %a.addr, align 4 %arrayidx3 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %6 - %7 = load i32* %arrayidx3, align 4 + %7 = load i32, i32* %arrayidx3, align 4 ret i32 %7 } @@ -1045,33 +1045,33 @@ entry: br label %for.cond for.cond: ; preds = %for.inc, %entry - %0 = load i32* %i, align 4 + %0 = load i32, i32* %i, align 4 %cmp = icmp slt i32 %0, 100 br i1 %cmp, label %for.body, label %for.end for.body: ; preds = %for.cond - %1 = load i32* %i, align 4 + %1 = load i32, i32* %i, align 4 %sub = sub nsw i32 100, %1 %sub1 = sub nsw i32 %sub, 10 %arrayidx = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 2), i32 0, i32 %sub1 - %2 = load i32* %arrayidx, align 4 - %3 = load i32* %a.addr, align 4 + %2 = load i32, i32* %arrayidx, align 4 + %3 = load i32, i32* %a.addr, align 4 %add = add nsw i32 %2, %3 - %4 = load i32* %i, align 4 + %4 = load i32, i32* %i, align 4 %add2 = add nsw i32 %4, 10 %arrayidx3 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %add2 store i32 %add, i32* %arrayidx3, align 4 br label %for.inc for.inc: ; preds = %for.body - %5 = load i32* %i, align 4 + %5 = load i32, i32* %i, align 4 %inc = add nsw i32 %5, 1 store i32 %inc, i32* %i, align 4 br label %for.cond for.end: ; preds = %for.cond - %6 = load i32* %a.addr, align 4 + %6 = load i32, i32* %a.addr, align 4 %arrayidx4 = getelementptr inbounds [100 x i32], [100 x i32]* getelementptr inbounds (%struct.anon* @Foo, i32 0, i32 0), i32 0, i32 %6 - %7 = load i32* %arrayidx4, align 4 + %7 = load i32, i32* %arrayidx4, align 4 ret i32 %7 } diff --git a/llvm/test/Transforms/LoopVectorize/hoist-loads.ll b/llvm/test/Transforms/LoopVectorize/hoist-loads.ll index ae7f5dc3b79..a20b0f6cdf3 100644 --- a/llvm/test/Transforms/LoopVectorize/hoist-loads.ll +++ b/llvm/test/Transforms/LoopVectorize/hoist-loads.ll @@ -16,12 +16,12 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %if.end9 ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @A, i64 0, i64 %indvars.iv %arrayidx2 = getelementptr inbounds [1024 x float], [1024 x float]* @B, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx2, align 4 + %0 = load float, float* %arrayidx2, align 4 %cmp3 = fcmp oeq float %0, 0.000000e+00 br i1 %cmp3, label %if.end9, label %if.else if.else: - %1 = load float* %arrayidx, align 4 + %1 = load float, float* %arrayidx, align 4 br label %if.end9 if.end9: @@ -48,12 +48,12 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %if.end9 ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @A, i64 0, i64 %indvars.iv %arrayidx2 = getelementptr inbounds [1024 x float], [1024 x float]* @B, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx2, align 4 + %0 = load float, float* %arrayidx2, align 4 %cmp3 = fcmp oeq float %0, 0.000000e+00 br i1 %cmp3, label %if.end9, label %if.else if.else: - %1 = load float* %arrayidx, align 4 + %1 = load float, float* %arrayidx, align 4 br label %if.end9 if.end9: diff --git a/llvm/test/Transforms/LoopVectorize/i8-induction.ll b/llvm/test/Transforms/LoopVectorize/i8-induction.ll index 90e3ec00cde..d9e8a430be1 100644 --- a/llvm/test/Transforms/LoopVectorize/i8-induction.ll +++ b/llvm/test/Transforms/LoopVectorize/i8-induction.ll @@ -9,7 +9,7 @@ target triple = "x86_64-apple-macosx10.8.0" define void @f() nounwind uwtable ssp { scalar.ph: store i8 0, i8* inttoptr (i64 1 to i8*), align 1 - %0 = load i8* @a, align 1 + %0 = load i8, i8* @a, align 1 br label %for.body for.body: diff --git a/llvm/test/Transforms/LoopVectorize/if-conversion-edgemasks.ll b/llvm/test/Transforms/LoopVectorize/if-conversion-edgemasks.ll index f2d1f2596c2..8d435f51353 100644 --- a/llvm/test/Transforms/LoopVectorize/if-conversion-edgemasks.ll +++ b/llvm/test/Transforms/LoopVectorize/if-conversion-edgemasks.ll @@ -20,15 +20,15 @@ entry: br i1 %cmp88, label %for.body.lr.ph, label %for.end for.body.lr.ph: - %0 = load i32** @b, align 8 - %1 = load i32** @a, align 8 - %2 = load i32** @c, align 8 + %0 = load i32*, i32** @b, align 8 + %1 = load i32*, i32** @a, align 8 + %2 = load i32*, i32** @c, align 8 br label %for.body for.body: %indvars.iv = phi i64 [ 0, %for.body.lr.ph ], [ %indvars.iv.next, %_ZL3fn3ii.exit58 ] %arrayidx = getelementptr inbounds i32, i32* %0, i64 %indvars.iv - %3 = load i32* %arrayidx, align 4 %4 = trunc i64 %indvars.iv to i32 + %3 = load i32, i32* %arrayidx, align 4 %4 = trunc i64 %indvars.iv to i32 %and.i = and i32 %4, 1 %tobool.i.i = icmp eq i32 %and.i, 0 br i1 %tobool.i.i, label %if.end.i, label %if.then.i @@ -136,7 +136,7 @@ _ZL3fn3ii.exit: %p1.addr.0.i16.i = phi i32 [ %or.i14.i, %if.then.i15.i ], [ %p1.addr.3.i.i, %_Z3fn2iii.exit.i ] %arrayidx2 = getelementptr inbounds i32, i32* %1, i64 %indvars.iv store i32 %p1.addr.0.i16.i, i32* %arrayidx2, align 4 %arrayidx4 = getelementptr inbounds i32, i32* %0, i64 %indvars.iv - %10 = load i32* %arrayidx4, align 4 br i1 %tobool.i.i, label %_Z3fn1ii.exit.i26, label %if.then.i.i21 + %10 = load i32, i32* %arrayidx4, align 4 br i1 %tobool.i.i, label %_Z3fn1ii.exit.i26, label %if.then.i.i21 if.then.i.i21: %and.i.i18 = lshr i32 %10, 2 diff --git a/llvm/test/Transforms/LoopVectorize/if-conversion-nest.ll b/llvm/test/Transforms/LoopVectorize/if-conversion-nest.ll index 4a327e8fe4f..3a581ebf847 100644 --- a/llvm/test/Transforms/LoopVectorize/if-conversion-nest.ll +++ b/llvm/test/Transforms/LoopVectorize/if-conversion-nest.ll @@ -20,9 +20,9 @@ entry: for.body: %indvars.iv = phi i64 [ %indvars.iv.next, %if.end14 ], [ 0, %entry ] %arrayidx = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds i32, i32* %B, i64 %indvars.iv - %1 = load i32* %arrayidx2, align 4 + %1 = load i32, i32* %arrayidx2, align 4 %cmp3 = icmp sgt i32 %0, %1 br i1 %cmp3, label %if.then, label %if.end14 diff --git a/llvm/test/Transforms/LoopVectorize/if-conversion-reduction.ll b/llvm/test/Transforms/LoopVectorize/if-conversion-reduction.ll index 711ca7633a5..20333b94345 100644 --- a/llvm/test/Transforms/LoopVectorize/if-conversion-reduction.ll +++ b/llvm/test/Transforms/LoopVectorize/if-conversion-reduction.ll @@ -15,7 +15,7 @@ for.body: ; preds = %entry, %for.inc %indvars.iv = phi i64 [ %indvars.iv.next, %for.inc ], [ 0, %entry ] %sum.011 = phi i32 [ %sum.1, %for.inc ], [ 0, %entry ] %arrayidx = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %cmp1 = icmp sgt i32 %0, 30 br i1 %cmp1, label %if.then, label %for.inc diff --git a/llvm/test/Transforms/LoopVectorize/if-conversion.ll b/llvm/test/Transforms/LoopVectorize/if-conversion.ll index 7a3d825353e..3a6ac8b1747 100644 --- a/llvm/test/Transforms/LoopVectorize/if-conversion.ll +++ b/llvm/test/Transforms/LoopVectorize/if-conversion.ll @@ -36,9 +36,9 @@ for.body.lr.ph: for.body: %indvars.iv = phi i64 [ %0, %for.body.lr.ph ], [ %indvars.iv.next, %if.end ] %arrayidx = getelementptr inbounds i32, i32* %a, i64 %indvars.iv - %1 = load i32* %arrayidx, align 4 + %1 = load i32, i32* %arrayidx, align 4 %arrayidx4 = getelementptr inbounds i32, i32* %b, i64 %indvars.iv - %2 = load i32* %arrayidx4, align 4 + %2 = load i32, i32* %arrayidx4, align 4 %cmp5 = icmp sgt i32 %1, %2 br i1 %cmp5, label %if.then, label %if.end @@ -85,7 +85,7 @@ for.body: ; preds = %entry, %for.inc %indvars.iv = phi i64 [ %indvars.iv.next, %for.inc ], [ 0, %entry ] %sum.011 = phi i32 [ %sum.1, %for.inc ], [ 0, %entry ] %arrayidx = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %cmp1 = icmp sgt i32 %0, 30 br i1 %cmp1, label %if.then, label %for.inc diff --git a/llvm/test/Transforms/LoopVectorize/if-pred-stores.ll b/llvm/test/Transforms/LoopVectorize/if-pred-stores.ll index ee5f8bcce88..991d027ada5 100644 --- a/llvm/test/Transforms/LoopVectorize/if-pred-stores.ll +++ b/llvm/test/Transforms/LoopVectorize/if-pred-stores.ll @@ -41,8 +41,8 @@ entry: ; UNROLL: %[[IND1:[a-zA-Z0-9]+]] = add i64 %{{.*}}, 1 ; UNROLL: %[[v0:[a-zA-Z0-9]+]] = getelementptr inbounds i32, i32* %f, i64 %[[IND]] ; UNROLL: %[[v1:[a-zA-Z0-9]+]] = getelementptr inbounds i32, i32* %f, i64 %[[IND1]] -; UNROLL: %[[v2:[a-zA-Z0-9]+]] = load i32* %[[v0]], align 4 -; UNROLL: %[[v3:[a-zA-Z0-9]+]] = load i32* %[[v1]], align 4 +; UNROLL: %[[v2:[a-zA-Z0-9]+]] = load i32, i32* %[[v0]], align 4 +; UNROLL: %[[v3:[a-zA-Z0-9]+]] = load i32, i32* %[[v1]], align 4 ; UNROLL: %[[v4:[a-zA-Z0-9]+]] = icmp sgt i32 %[[v2]], 100 ; UNROLL: %[[v5:[a-zA-Z0-9]+]] = icmp sgt i32 %[[v3]], 100 ; UNROLL: %[[v6:[a-zA-Z0-9]+]] = add nsw i32 %[[v2]], 20 @@ -67,7 +67,7 @@ entry: for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.inc ] %arrayidx = getelementptr inbounds i32, i32* %f, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %cmp1 = icmp sgt i32 %0, 100 br i1 %cmp1, label %if.then, label %for.inc @@ -105,7 +105,7 @@ for.body14: %indvars.iv3 = phi i64 [ %indvars.iv.next4, %for.inc23 ], [ undef, %for.body9 ] %iNewChunks.120 = phi i32 [ %iNewChunks.2, %for.inc23 ], [ undef, %for.body9 ] %arrayidx16 = getelementptr inbounds [768 x i32], [768 x i32]* undef, i64 0, i64 %indvars.iv3 - %tmp = load i32* %arrayidx16, align 4 + %tmp = load i32, i32* %arrayidx16, align 4 br i1 undef, label %if.then18, label %for.inc23 if.then18: diff --git a/llvm/test/Transforms/LoopVectorize/incorrect-dom-info.ll b/llvm/test/Transforms/LoopVectorize/incorrect-dom-info.ll index 12fc13a5265..798793a641b 100644 --- a/llvm/test/Transforms/LoopVectorize/incorrect-dom-info.ll +++ b/llvm/test/Transforms/LoopVectorize/incorrect-dom-info.ll @@ -59,7 +59,7 @@ thread-pre-split.loopexit: ; preds = %11, %.thread-pre-sp ; <label>:11 ; preds = %.lr.ph21 %12 = getelementptr inbounds [0 x i8], [0 x i8]* @PL_utf8skip, i64 0, i64 undef - %13 = load i8* %12, align 1 + %13 = load i8, i8* %12, align 1 %14 = zext i8 %13 to i64 %15 = icmp ugt i64 %14, %10 %. = select i1 %15, i64 %10, i64 %14 @@ -91,7 +91,7 @@ thread-pre-split.loopexit: ; preds = %11, %.thread-pre-sp br label %26 ; <label>:26 ; preds = %25, %24, %23, %22 - %27 = load i64* %len, align 8 + %27 = load i64, i64* %len, align 8 %28 = add i64 %27, -1 br i1 undef, label %thread-pre-split._crit_edge, label %.lr.ph21 diff --git a/llvm/test/Transforms/LoopVectorize/increment.ll b/llvm/test/Transforms/LoopVectorize/increment.ll index 369bd38c78f..d0b25096e37 100644 --- a/llvm/test/Transforms/LoopVectorize/increment.ll +++ b/llvm/test/Transforms/LoopVectorize/increment.ll @@ -21,7 +21,7 @@ define void @inc(i32 %n) nounwind uwtable noinline ssp { .lr.ph: ; preds = %0, %.lr.ph %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 0, %0 ] %2 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = trunc i64 %indvars.iv to i32 %5 = add nsw i32 %3, %4 store i32 %5, i32* %2, align 4 @@ -50,10 +50,10 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds i32, i32* %B, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %idxprom1 = sext i32 %0 to i64 %arrayidx2 = getelementptr inbounds i32, i32* %A, i64 %idxprom1 - %1 = load i32* %arrayidx2, align 4 + %1 = load i32, i32* %arrayidx2, align 4 %inc = add nsw i32 %1, 1 store i32 %inc, i32* %arrayidx2, align 4 %indvars.iv.next = add i64 %indvars.iv, 1 diff --git a/llvm/test/Transforms/LoopVectorize/induction.ll b/llvm/test/Transforms/LoopVectorize/induction.ll index 811c4927245..2fbb2de797a 100644 --- a/llvm/test/Transforms/LoopVectorize/induction.ll +++ b/llvm/test/Transforms/LoopVectorize/induction.ll @@ -52,10 +52,10 @@ for.body: %iv = phi i64 [ 0, %entry ], [ %iv.next, %for.body ] %ind.sum = add i64 %iv, %offset %arr.idx = getelementptr inbounds float, float* %a, i64 %ind.sum - %l1 = load float* %arr.idx, align 4 + %l1 = load float, float* %arr.idx, align 4 %ind.sum2 = add i64 %iv, %offset2 %arr.idx2 = getelementptr inbounds float, float* %a, i64 %ind.sum2 - %l2 = load float* %arr.idx2, align 4 + %l2 = load float, float* %arr.idx2, align 4 %m = fmul fast float %b, %l2 %ad = fadd fast float %l1, %m store float %ad, float* %arr.idx, align 4 @@ -153,9 +153,9 @@ define i32 @max_i32_backedgetaken() nounwind readnone ssp uwtable { @c = common global i32 0, align 4 define i32 @testoverflowcheck() { entry: - %.pr.i = load i8* @e, align 1 - %0 = load i32* @d, align 4 - %c.promoted.i = load i32* @c, align 4 + %.pr.i = load i8, i8* @e, align 1 + %0 = load i32, i32* @d, align 4 + %c.promoted.i = load i32, i32* @c, align 4 br label %cond.end.i cond.end.i: diff --git a/llvm/test/Transforms/LoopVectorize/intrinsic.ll b/llvm/test/Transforms/LoopVectorize/intrinsic.ll index 029d8b61a4a..fae6a8cb714 100644 --- a/llvm/test/Transforms/LoopVectorize/intrinsic.ll +++ b/llvm/test/Transforms/LoopVectorize/intrinsic.ll @@ -14,7 +14,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %call = tail call float @llvm.sqrt.f32(float %0) nounwind readnone %arrayidx2 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx2, align 4 @@ -40,7 +40,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %call = tail call double @llvm.sqrt.f64(double %0) nounwind readnone %arrayidx2 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx2, align 8 @@ -66,7 +66,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %call = tail call float @llvm.sin.f32(float %0) nounwind readnone %arrayidx2 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx2, align 4 @@ -92,7 +92,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %call = tail call double @llvm.sin.f64(double %0) nounwind readnone %arrayidx2 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx2, align 8 @@ -118,7 +118,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %call = tail call float @llvm.cos.f32(float %0) nounwind readnone %arrayidx2 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx2, align 4 @@ -144,7 +144,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %call = tail call double @llvm.cos.f64(double %0) nounwind readnone %arrayidx2 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx2, align 8 @@ -170,7 +170,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %call = tail call float @llvm.exp.f32(float %0) nounwind readnone %arrayidx2 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx2, align 4 @@ -196,7 +196,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %call = tail call double @llvm.exp.f64(double %0) nounwind readnone %arrayidx2 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx2, align 8 @@ -222,7 +222,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %call = tail call float @llvm.exp2.f32(float %0) nounwind readnone %arrayidx2 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx2, align 4 @@ -248,7 +248,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %call = tail call double @llvm.exp2.f64(double %0) nounwind readnone %arrayidx2 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx2, align 8 @@ -274,7 +274,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %call = tail call float @llvm.log.f32(float %0) nounwind readnone %arrayidx2 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx2, align 4 @@ -300,7 +300,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %call = tail call double @llvm.log.f64(double %0) nounwind readnone %arrayidx2 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx2, align 8 @@ -326,7 +326,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %call = tail call float @llvm.log10.f32(float %0) nounwind readnone %arrayidx2 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx2, align 4 @@ -352,7 +352,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %call = tail call double @llvm.log10.f64(double %0) nounwind readnone %arrayidx2 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx2, align 8 @@ -378,7 +378,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %call = tail call float @llvm.log2.f32(float %0) nounwind readnone %arrayidx2 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx2, align 4 @@ -404,7 +404,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %call = tail call double @llvm.log2.f64(double %0) nounwind readnone %arrayidx2 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx2, align 8 @@ -430,7 +430,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %call = tail call float @llvm.fabs.f32(float %0) nounwind readnone %arrayidx2 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx2, align 4 @@ -453,7 +453,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %call = tail call double @llvm.fabs(double %0) nounwind readnone %arrayidx2 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx2, align 8 @@ -479,9 +479,9 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %arrayidx1 = getelementptr inbounds float, float* %z, i64 %indvars.iv - %1 = load float* %arrayidx1, align 4 + %1 = load float, float* %arrayidx1, align 4 %call = tail call float @llvm.copysign.f32(float %0, float %1) nounwind readnone %arrayidx2 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx2, align 4 @@ -504,9 +504,9 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %arrayidx1 = getelementptr inbounds double, double* %z, i64 %indvars.iv - %1 = load double* %arrayidx, align 8 + %1 = load double, double* %arrayidx, align 8 %call = tail call double @llvm.copysign(double %0, double %1) nounwind readnone %arrayidx2 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx2, align 8 @@ -532,7 +532,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %call = tail call float @llvm.floor.f32(float %0) nounwind readnone %arrayidx2 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx2, align 4 @@ -558,7 +558,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %call = tail call double @llvm.floor.f64(double %0) nounwind readnone %arrayidx2 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx2, align 8 @@ -584,7 +584,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %call = tail call float @llvm.ceil.f32(float %0) nounwind readnone %arrayidx2 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx2, align 4 @@ -610,7 +610,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %call = tail call double @llvm.ceil.f64(double %0) nounwind readnone %arrayidx2 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx2, align 8 @@ -636,7 +636,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %call = tail call float @llvm.trunc.f32(float %0) nounwind readnone %arrayidx2 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx2, align 4 @@ -662,7 +662,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %call = tail call double @llvm.trunc.f64(double %0) nounwind readnone %arrayidx2 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx2, align 8 @@ -688,7 +688,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %call = tail call float @llvm.rint.f32(float %0) nounwind readnone %arrayidx2 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx2, align 4 @@ -714,7 +714,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %call = tail call double @llvm.rint.f64(double %0) nounwind readnone %arrayidx2 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx2, align 8 @@ -740,7 +740,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %call = tail call float @llvm.nearbyint.f32(float %0) nounwind readnone %arrayidx2 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx2, align 4 @@ -766,7 +766,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %call = tail call double @llvm.nearbyint.f64(double %0) nounwind readnone %arrayidx2 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx2, align 8 @@ -792,7 +792,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %call = tail call float @llvm.round.f32(float %0) nounwind readnone %arrayidx2 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx2, align 4 @@ -818,7 +818,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %call = tail call double @llvm.round.f64(double %0) nounwind readnone %arrayidx2 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx2, align 8 @@ -844,11 +844,11 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds float, float* %w, i64 %indvars.iv - %1 = load float* %arrayidx2, align 4 + %1 = load float, float* %arrayidx2, align 4 %arrayidx4 = getelementptr inbounds float, float* %z, i64 %indvars.iv - %2 = load float* %arrayidx4, align 4 + %2 = load float, float* %arrayidx4, align 4 %3 = tail call float @llvm.fma.f32(float %0, float %2, float %1) %arrayidx6 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %3, float* %arrayidx6, align 4 @@ -874,11 +874,11 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %arrayidx2 = getelementptr inbounds double, double* %w, i64 %indvars.iv - %1 = load double* %arrayidx2, align 8 + %1 = load double, double* %arrayidx2, align 8 %arrayidx4 = getelementptr inbounds double, double* %z, i64 %indvars.iv - %2 = load double* %arrayidx4, align 8 + %2 = load double, double* %arrayidx4, align 8 %3 = tail call double @llvm.fma.f64(double %0, double %2, double %1) %arrayidx6 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %3, double* %arrayidx6, align 8 @@ -904,11 +904,11 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds float, float* %w, i64 %indvars.iv - %1 = load float* %arrayidx2, align 4 + %1 = load float, float* %arrayidx2, align 4 %arrayidx4 = getelementptr inbounds float, float* %z, i64 %indvars.iv - %2 = load float* %arrayidx4, align 4 + %2 = load float, float* %arrayidx4, align 4 %3 = tail call float @llvm.fmuladd.f32(float %0, float %2, float %1) %arrayidx6 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %3, float* %arrayidx6, align 4 @@ -934,11 +934,11 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %arrayidx2 = getelementptr inbounds double, double* %w, i64 %indvars.iv - %1 = load double* %arrayidx2, align 8 + %1 = load double, double* %arrayidx2, align 8 %arrayidx4 = getelementptr inbounds double, double* %z, i64 %indvars.iv - %2 = load double* %arrayidx4, align 8 + %2 = load double, double* %arrayidx4, align 8 %3 = tail call double @llvm.fmuladd.f64(double %0, double %2, double %1) %arrayidx6 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %3, double* %arrayidx6, align 8 @@ -964,9 +964,9 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds float, float* %z, i64 %indvars.iv - %1 = load float* %arrayidx2, align 4 + %1 = load float, float* %arrayidx2, align 4 %call = tail call float @llvm.pow.f32(float %0, float %1) nounwind readnone %arrayidx4 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx4, align 4 @@ -992,9 +992,9 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %arrayidx2 = getelementptr inbounds double, double* %z, i64 %indvars.iv - %1 = load double* %arrayidx2, align 8 + %1 = load double, double* %arrayidx2, align 8 %call = tail call double @llvm.pow.f64(double %0, double %1) nounwind readnone %arrayidx4 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx4, align 8 @@ -1017,7 +1017,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds float, float* %x, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %call = tail call float @fabsf(float %0) nounwind readnone store float %call, float* %arrayidx, align 4 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -1051,7 +1051,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds float, float* %x, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %call = tail call float @roundf(float %0) nounwind readnone store float %call, float* %arrayidx, align 4 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -1078,7 +1078,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds double, double* %x, i64 %indvars.iv - %0 = load double* %arrayidx, align 4 + %0 = load double, double* %arrayidx, align 4 store double %0, double* %arrayidx, align 4 tail call void @round(double %0) nounwind readnone %indvars.iv.next = add i64 %indvars.iv, 1 @@ -1103,7 +1103,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %call = tail call double @llvm.powi.f64(double %0, i32 %P) nounwind readnone %arrayidx4 = getelementptr inbounds double, double* %x, i64 %indvars.iv store double %call, double* %arrayidx4, align 8 @@ -1127,7 +1127,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds double, double* %y, i64 %indvars.iv - %0 = load double* %arrayidx, align 8 + %0 = load double, double* %arrayidx, align 8 %1 = trunc i64 %indvars.iv to i32 %call = tail call double @llvm.powi.f64(double %0, i32 %1) nounwind readnone %arrayidx4 = getelementptr inbounds double, double* %x, i64 %indvars.iv @@ -1154,7 +1154,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds i64, i64* %y, i64 %indvars.iv - %0 = load i64* %arrayidx, align 8 + %0 = load i64, i64* %arrayidx, align 8 %call = tail call i64 @llvm.cttz.i64(i64 %0, i1 true) nounwind readnone %arrayidx4 = getelementptr inbounds i64, i64* %x, i64 %indvars.iv store i64 %call, i64* %arrayidx4, align 8 @@ -1180,7 +1180,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds i64, i64* %y, i64 %indvars.iv - %0 = load i64* %arrayidx, align 8 + %0 = load i64, i64* %arrayidx, align 8 %call = tail call i64 @llvm.ctlz.i64(i64 %0, i1 true) nounwind readnone %arrayidx4 = getelementptr inbounds i64, i64* %x, i64 %indvars.iv store i64 %call, i64* %arrayidx4, align 8 @@ -1206,9 +1206,9 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds float, float* %z, i64 %indvars.iv - %1 = load float* %arrayidx2, align 4 + %1 = load float, float* %arrayidx2, align 4 %call = tail call float @llvm.minnum.f32(float %0, float %1) nounwind readnone %arrayidx4 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx4, align 4 @@ -1234,9 +1234,9 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %y, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds float, float* %z, i64 %indvars.iv - %1 = load float* %arrayidx2, align 4 + %1 = load float, float* %arrayidx2, align 4 %call = tail call float @llvm.maxnum.f32(float %0, float %1) nounwind readnone %arrayidx4 = getelementptr inbounds float, float* %x, i64 %indvars.iv store float %call, float* %arrayidx4, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/lifetime.ll b/llvm/test/Transforms/LoopVectorize/lifetime.ll index 63bde5a4570..6e525ca1d82 100644 --- a/llvm/test/Transforms/LoopVectorize/lifetime.ll +++ b/llvm/test/Transforms/LoopVectorize/lifetime.ll @@ -20,7 +20,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] call void @llvm.lifetime.end(i64 4096, i8* %0) #1 %arrayidx = getelementptr inbounds i32, i32* %d, i64 %indvars.iv - %1 = load i32* %arrayidx, align 8 + %1 = load i32, i32* %arrayidx, align 8 store i32 100, i32* %arrayidx, align 8 call void @llvm.lifetime.start(i64 4096, i8* %0) #1 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -50,7 +50,7 @@ for.body: %1 = bitcast [1024 x i32]* %arr to i8* call void @llvm.lifetime.end(i64 4096, i8* %1) #1 %arrayidx = getelementptr inbounds i32, i32* %d, i64 %indvars.iv - %2 = load i32* %arrayidx, align 8 + %2 = load i32, i32* %arrayidx, align 8 store i32 100, i32* %arrayidx, align 8 call void @llvm.lifetime.start(i64 4096, i8* %1) #1 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -79,7 +79,7 @@ for.body: %1 = bitcast [1024 x i32]* %arr to i8* call void @llvm.lifetime.end(i64 4096, i8* %1) #1 %arrayidx = getelementptr inbounds i32, i32* %d, i64 %indvars.iv - %2 = load i32* %arrayidx, align 8 + %2 = load i32, i32* %arrayidx, align 8 store i32 100, i32* %arrayidx, align 8 call void @llvm.lifetime.start(i64 4096, i8* %1) #1 %indvars.iv.next = add i64 %indvars.iv, 1 diff --git a/llvm/test/Transforms/LoopVectorize/loop-vect-memdep.ll b/llvm/test/Transforms/LoopVectorize/loop-vect-memdep.ll index 326c4d42f75..d9efaa5c085 100644 --- a/llvm/test/Transforms/LoopVectorize/loop-vect-memdep.ll +++ b/llvm/test/Transforms/LoopVectorize/loop-vect-memdep.ll @@ -6,7 +6,7 @@ target datalayout = "e-m:e-i64:64-f80:128-n8:16:32:64-S128" define void @test_loop_novect(double** %arr, i64 %n) { for.body.lr.ph: - %t = load double** %arr, align 8 + %t = load double*, double** %arr, align 8 br label %for.body for.body: ; preds = %for.body, %for.body.lr.ph @@ -14,8 +14,8 @@ for.body: ; preds = %for.body, %for.body.lr %a = getelementptr inbounds double, double* %t, i64 %i %i.next = add nuw nsw i64 %i, 1 %a.next = getelementptr inbounds double, double* %t, i64 %i.next - %t1 = load double* %a, align 8 - %t2 = load double* %a.next, align 8 + %t1 = load double, double* %a, align 8 + %t2 = load double, double* %a.next, align 8 store double %t1, double* %a.next, align 8 store double %t2, double* %a, align 8 %c = icmp eq i64 %i, %n diff --git a/llvm/test/Transforms/LoopVectorize/memdep.ll b/llvm/test/Transforms/LoopVectorize/memdep.ll index 43fccb730dd..fb608836dfe 100644 --- a/llvm/test/Transforms/LoopVectorize/memdep.ll +++ b/llvm/test/Transforms/LoopVectorize/memdep.ll @@ -20,7 +20,7 @@ for.body: %indvars.iv = phi i32 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %indvars.iv.next = add i32 %indvars.iv, 1 %arrayidx = getelementptr inbounds i32, i32* %A, i32 %indvars.iv.next - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %add1 = add nsw i32 %0, 1 %arrayidx3 = getelementptr inbounds i32, i32* %A, i32 %indvars.iv store i32 %add1, i32* %arrayidx3, align 4 @@ -45,7 +45,7 @@ entry: for.body: %indvars.iv = phi i32 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds i32, i32* %A, i32 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %add = add nsw i32 %0, 1 %indvars.iv.next = add i32 %indvars.iv, 1 %arrayidx3 = getelementptr inbounds i32, i32* %A, i32 %indvars.iv.next @@ -75,7 +75,7 @@ for.body: %i.01 = phi i32 [ 0, %entry ], [ %inc, %for.body ] %idxprom = sext i32 %i.01 to i64 %arrayidx = getelementptr inbounds i32, i32* %A, i64 %idxprom - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %add = add nsw i32 %0, 1 %add1 = add nsw i32 %i.01, 2 %idxprom2 = sext i32 %add1 to i64 @@ -106,12 +106,12 @@ entry: for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds i32, i32* %B, i64 %indvars.iv store i32 %0, i32* %arrayidx2, align 4 %indvars.iv.next = add nsw i64 %indvars.iv, 1 %arrayidx4 = getelementptr inbounds i32, i32* %B, i64 %indvars.iv.next - %1 = load i32* %arrayidx4, align 4 + %1 = load i32, i32* %arrayidx4, align 4 store i32 %1, i32* %arrayidx, align 4 %lftr.wideiv = trunc i64 %indvars.iv.next to i32 %exitcond = icmp ne i32 %lftr.wideiv, 1024 @@ -141,7 +141,7 @@ for.body: %arrayidx = getelementptr inbounds i32, i32* %a, i64 %indvars.iv.next store i32 %tmp.addr.08, i32* %arrayidx, align 4 %arrayidx3 = getelementptr inbounds i32, i32* %a, i64 %indvars.iv - %0 = load i32* %arrayidx3, align 4 + %0 = load i32, i32* %arrayidx3, align 4 %lftr.wideiv = trunc i64 %indvars.iv.next to i32 %exitcond = icmp ne i32 %lftr.wideiv, 1024 br i1 %exitcond, label %for.body, label %for.end @@ -170,10 +170,10 @@ for.body: %indvars.iv = phi i64 [ 16, %entry ], [ %indvars.iv.next, %for.body ] %0 = add nsw i64 %indvars.iv, -3 %arrayidx = getelementptr inbounds i32, i32* %A, i64 %0 - %1 = load i32* %arrayidx, align 4 + %1 = load i32, i32* %arrayidx, align 4 %2 = add nsw i64 %indvars.iv, 4 %arrayidx2 = getelementptr inbounds i32, i32* %A, i64 %2 - %3 = load i32* %arrayidx2, align 4 + %3 = load i32, i32* %arrayidx2, align 4 %add3 = add nsw i32 %3, %1 %arrayidx5 = getelementptr inbounds i32, i32* %A, i64 %indvars.iv store i32 %add3, i32* %arrayidx5, align 4 @@ -204,12 +204,12 @@ entry: for.body: %indvars.iv = phi i64 [ 16, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds i32, i32* %B, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds i32, i32* %A, i64 %indvars.iv store i32 %0, i32* %arrayidx2, align 4 %1 = add nsw i64 %indvars.iv, -3 %arrayidx4 = getelementptr inbounds i32, i32* %A, i64 %1 - %2 = load i32* %arrayidx4, align 4 + %2 = load i32, i32* %arrayidx4, align 4 %arrayidx6 = getelementptr inbounds i32, i32* %C, i64 %indvars.iv store i32 %2, i32* %arrayidx6, align 4 %indvars.iv.next = add i64 %indvars.iv, 1 diff --git a/llvm/test/Transforms/LoopVectorize/metadata-unroll.ll b/llvm/test/Transforms/LoopVectorize/metadata-unroll.ll index 38df838de2a..3c80ae0cb54 100644 --- a/llvm/test/Transforms/LoopVectorize/metadata-unroll.ll +++ b/llvm/test/Transforms/LoopVectorize/metadata-unroll.ll @@ -24,7 +24,7 @@ define void @inc(i32 %n) nounwind uwtable noinline ssp { .lr.ph: ; preds = %0, %.lr.ph %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 0, %0 ] %2 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = trunc i64 %indvars.iv to i32 %5 = add nsw i32 %3, %4 store i32 %5, i32* %2, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/metadata.ll b/llvm/test/Transforms/LoopVectorize/metadata.ll index 009463cdc11..9a791aedbbe 100644 --- a/llvm/test/Transforms/LoopVectorize/metadata.ll +++ b/llvm/test/Transforms/LoopVectorize/metadata.ll @@ -10,7 +10,7 @@ entry: for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds float, float* %b, i64 %indvars.iv - %0 = load float* %arrayidx, align 4, !tbaa !0 + %0 = load float, float* %arrayidx, align 4, !tbaa !0 %conv = fptosi float %0 to i32 %arrayidx2 = getelementptr inbounds i32, i32* %a, i64 %indvars.iv store i32 %conv, i32* %arrayidx2, align 4, !tbaa !4 @@ -23,7 +23,7 @@ for.end: ; preds = %for.body } ; CHECK-LABEL: @test1 -; CHECK: load <4 x float>* %{{.*}}, align 4, !tbaa ![[TFLT:[0-9]+]] +; CHECK: load <4 x float>, <4 x float>* %{{.*}}, align 4, !tbaa ![[TFLT:[0-9]+]] ; CHECK: store <4 x i32> %{{.*}}, <4 x i32>* %{{.*}}, align 4, !tbaa ![[TINT:[0-9]+]] ; CHECK: ret i32 0 diff --git a/llvm/test/Transforms/LoopVectorize/minmax_reduction.ll b/llvm/test/Transforms/LoopVectorize/minmax_reduction.ll index 858536fbed9..5a0356fe11a 100644 --- a/llvm/test/Transforms/LoopVectorize/minmax_reduction.ll +++ b/llvm/test/Transforms/LoopVectorize/minmax_reduction.ll @@ -27,7 +27,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi i32 [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 0, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %cmp3 = icmp sgt i32 %0, %max.red.08 %max.red.0 = select i1 %cmp3, i32 %0, i32 %max.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -56,7 +56,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi i32 [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 0, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %cmp3 = icmp slt i32 %max.red.08, %0 %max.red.0 = select i1 %cmp3, i32 %0, i32 %max.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -84,7 +84,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi i32 [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 0, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %cmp3 = icmp slt i32 %0, %max.red.08 %max.red.0 = select i1 %cmp3, i32 %0, i32 %max.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -113,7 +113,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi i32 [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 0, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %cmp3 = icmp sgt i32 %max.red.08, %0 %max.red.0 = select i1 %cmp3, i32 %0, i32 %max.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -143,7 +143,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi i32 [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 0, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %cmp3 = icmp ugt i32 %0, %max.red.08 %max.red.0 = select i1 %cmp3, i32 %0, i32 %max.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -172,7 +172,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi i32 [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 0, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %cmp3 = icmp ult i32 %max.red.08, %0 %max.red.0 = select i1 %cmp3, i32 %0, i32 %max.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -200,7 +200,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi i32 [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 0, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %cmp3 = icmp ult i32 %0, %max.red.08 %max.red.0 = select i1 %cmp3, i32 %0, i32 %max.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -229,7 +229,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi i32 [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 0, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %cmp3 = icmp ugt i32 %max.red.08, %0 %max.red.0 = select i1 %cmp3, i32 %0, i32 %max.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -258,7 +258,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi i32 [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 0, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %cmp3 = icmp sge i32 %0, %max.red.08 %max.red.0 = select i1 %cmp3, i32 %max.red.08, i32 %0 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -287,7 +287,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi i32 [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 0, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %cmp3 = icmp sle i32 %0, %max.red.08 %max.red.0 = select i1 %cmp3, i32 %max.red.08, i32 %0 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -316,7 +316,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi i32 [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 0, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %cmp3 = icmp uge i32 %0, %max.red.08 %max.red.0 = select i1 %cmp3, i32 %max.red.08, i32 %0 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -345,7 +345,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi i32 [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 0, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %cmp3 = icmp ule i32 %0, %max.red.08 %max.red.0 = select i1 %cmp3, i32 %max.red.08, i32 %0 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -369,8 +369,8 @@ for.body: %max.red.08 = phi i32 [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 0, i64 %indvars.iv %arrayidx1 = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 1, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 - %1 = load i32* %arrayidx1, align 4 + %0 = load i32, i32* %arrayidx, align 4 + %1 = load i32, i32* %arrayidx1, align 4 %cmp3 = icmp sgt i32 %0, %1 %max.red.0 = select i1 %cmp3, i32 %0, i32 %max.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -393,8 +393,8 @@ for.body: %max.red.08 = phi i32 [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 0, i64 %indvars.iv %arrayidx1 = getelementptr inbounds [1024 x i32], [1024 x i32]* @A, i64 1, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 - %1 = load i32* %arrayidx1, align 4 + %0 = load i32, i32* %arrayidx, align 4 + %1 = load i32, i32* %arrayidx1, align 4 %cmp3 = icmp sgt i32 %0, %max.red.08 %max.red.0 = select i1 %cmp3, i32 %0, i32 %1 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -426,7 +426,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi float [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @fA, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp3 = fcmp ogt float %0, %max.red.08 %max.red.0 = select i1 %cmp3, float %0, float %max.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -452,7 +452,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi float [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @fA, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp3 = fcmp oge float %0, %max.red.08 %max.red.0 = select i1 %cmp3, float %0, float %max.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -478,7 +478,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi float [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @fA, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp3 = fcmp olt float %0, %max.red.08 %max.red.0 = select i1 %cmp3, float %max.red.08, float %0 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -504,7 +504,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi float [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @fA, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp3 = fcmp ole float %0, %max.red.08 %max.red.0 = select i1 %cmp3, float %max.red.08, float %0 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -530,7 +530,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi float [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @fA, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp3 = fcmp ugt float %0, %max.red.08 %max.red.0 = select i1 %cmp3, float %0, float %max.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -556,7 +556,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi float [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @fA, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp3 = fcmp uge float %0, %max.red.08 %max.red.0 = select i1 %cmp3, float %0, float %max.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -582,7 +582,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi float [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @fA, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp3 = fcmp ult float %0, %max.red.08 %max.red.0 = select i1 %cmp3, float %max.red.08, float %0 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -608,7 +608,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi float [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @fA, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp3 = fcmp ule float %0, %max.red.08 %max.red.0 = select i1 %cmp3, float %max.red.08, float %0 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -637,7 +637,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %min.red.08 = phi float [ %min, %entry ], [ %min.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @fA, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp3 = fcmp olt float %0, %min.red.08 %min.red.0 = select i1 %cmp3, float %0, float %min.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -663,7 +663,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %min.red.08 = phi float [ %min, %entry ], [ %min.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @fA, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp3 = fcmp ole float %0, %min.red.08 %min.red.0 = select i1 %cmp3, float %0, float %min.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -689,7 +689,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %min.red.08 = phi float [ %min, %entry ], [ %min.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @fA, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp3 = fcmp ogt float %0, %min.red.08 %min.red.0 = select i1 %cmp3, float %min.red.08, float %0 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -715,7 +715,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %min.red.08 = phi float [ %min, %entry ], [ %min.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @fA, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp3 = fcmp oge float %0, %min.red.08 %min.red.0 = select i1 %cmp3, float %min.red.08, float %0 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -741,7 +741,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %min.red.08 = phi float [ %min, %entry ], [ %min.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @fA, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp3 = fcmp ult float %0, %min.red.08 %min.red.0 = select i1 %cmp3, float %0, float %min.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -767,7 +767,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %min.red.08 = phi float [ %min, %entry ], [ %min.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @fA, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp3 = fcmp ule float %0, %min.red.08 %min.red.0 = select i1 %cmp3, float %0, float %min.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -793,7 +793,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %min.red.08 = phi float [ %min, %entry ], [ %min.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @fA, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp3 = fcmp ugt float %0, %min.red.08 %min.red.0 = select i1 %cmp3, float %min.red.08, float %0 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -819,7 +819,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %min.red.08 = phi float [ %min, %entry ], [ %min.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @fA, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp3 = fcmp uge float %0, %min.red.08 %min.red.0 = select i1 %cmp3, float %min.red.08, float %0 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -846,7 +846,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %min.red.08 = phi double [ %min, %entry ], [ %min.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x double], [1024 x double]* @dA, i64 0, i64 %indvars.iv - %0 = load double* %arrayidx, align 4 + %0 = load double, double* %arrayidx, align 4 %cmp3 = fcmp olt double %0, %min.red.08 %min.red.0 = select i1 %cmp3, double %0, double %min.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -870,7 +870,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %max.red.08 = phi float [ %max, %entry ], [ %max.red.0, %for.body ] %arrayidx = getelementptr inbounds [1024 x float], [1024 x float]* @fA, i64 0, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %cmp3 = fcmp ogt float %0, %max.red.08 %max.red.0 = select i1 %cmp3, float %0, float %max.red.08 %indvars.iv.next = add i64 %indvars.iv, 1 diff --git a/llvm/test/Transforms/LoopVectorize/multi-use-reduction-bug.ll b/llvm/test/Transforms/LoopVectorize/multi-use-reduction-bug.ll index cd022adc939..9f7fb39dc7c 100644 --- a/llvm/test/Transforms/LoopVectorize/multi-use-reduction-bug.ll +++ b/llvm/test/Transforms/LoopVectorize/multi-use-reduction-bug.ll @@ -16,10 +16,10 @@ entry: %n = alloca i32, align 4 %k7 = alloca i32, align 4 %nf = alloca i32, align 4 - %0 = load i32* %k7, align 4 + %0 = load i32, i32* %k7, align 4 %.neg1 = sub i32 0, %0 - %n.promoted = load i32* %n, align 4 - %nf.promoted = load i32* %nf, align 4 + %n.promoted = load i32, i32* %n, align 4 + %nf.promoted = load i32, i32* %nf, align 4 br label %for.body for.body: diff --git a/llvm/test/Transforms/LoopVectorize/multiple-address-spaces.ll b/llvm/test/Transforms/LoopVectorize/multiple-address-spaces.ll index 1659af684e9..e79c93146b1 100644 --- a/llvm/test/Transforms/LoopVectorize/multiple-address-spaces.ll +++ b/llvm/test/Transforms/LoopVectorize/multiple-address-spaces.ll @@ -28,7 +28,7 @@ entry: for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds [40000 x i8], [40000 x i8] addrspace(1)* @Y, i64 0, i64 %indvars.iv - %0 = load i8 addrspace(1)* %arrayidx, align 1 + %0 = load i8, i8 addrspace(1)* %arrayidx, align 1 %add = add i8 %0, 1 %arrayidx3 = getelementptr inbounds [40000 x i8], [40000 x i8]* @X, i64 0, i64 %indvars.iv store i8 %add, i8* %arrayidx3, align 1 diff --git a/llvm/test/Transforms/LoopVectorize/no_array_bounds.ll b/llvm/test/Transforms/LoopVectorize/no_array_bounds.ll index 2835c6604ec..865c8da4c8d 100644 --- a/llvm/test/Transforms/LoopVectorize/no_array_bounds.ll +++ b/llvm/test/Transforms/LoopVectorize/no_array_bounds.ll @@ -34,10 +34,10 @@ for.body7.preheader: ; preds = %for.cond5.preheader for.body: ; preds = %for.body.preheader, %for.body %indvars.iv27 = phi i64 [ %indvars.iv.next28, %for.body ], [ 0, %for.body.preheader ] %arrayidx = getelementptr inbounds i32, i32* %B, i64 %indvars.iv27, !dbg !14 - %0 = load i32* %arrayidx, align 4, !dbg !14, !tbaa !22 + %0 = load i32, i32* %arrayidx, align 4, !dbg !14, !tbaa !22 %idxprom1 = sext i32 %0 to i64, !dbg !14 %arrayidx2 = getelementptr inbounds i32, i32* %A, i64 %idxprom1, !dbg !14 - %1 = load i32* %arrayidx2, align 4, !dbg !14, !tbaa !22 + %1 = load i32, i32* %arrayidx2, align 4, !dbg !14, !tbaa !22 %inc = add nsw i32 %1, 1, !dbg !14 store i32 %inc, i32* %arrayidx2, align 4, !dbg !14, !tbaa !22 %indvars.iv.next28 = add nuw nsw i64 %indvars.iv27, 1, !dbg !10 @@ -48,10 +48,10 @@ for.body: ; preds = %for.body.preheader, for.body7: ; preds = %for.body7.preheader, %for.body7 %indvars.iv = phi i64 [ %indvars.iv.next, %for.body7 ], [ 0, %for.body7.preheader ] %arrayidx9 = getelementptr inbounds i32, i32* %A, i64 %indvars.iv, !dbg !20 - %2 = load i32* %arrayidx9, align 4, !dbg !20, !tbaa !22 + %2 = load i32, i32* %arrayidx9, align 4, !dbg !20, !tbaa !22 %idxprom10 = sext i32 %2 to i64, !dbg !20 %arrayidx11 = getelementptr inbounds i32, i32* %B, i64 %idxprom10, !dbg !20 - %3 = load i32* %arrayidx11, align 4, !dbg !20, !tbaa !22 + %3 = load i32, i32* %arrayidx11, align 4, !dbg !20, !tbaa !22 %inc12 = add nsw i32 %3, 1, !dbg !20 store i32 %inc12, i32* %arrayidx11, align 4, !dbg !20, !tbaa !22 %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1, !dbg !16 diff --git a/llvm/test/Transforms/LoopVectorize/no_idiv_reduction.ll b/llvm/test/Transforms/LoopVectorize/no_idiv_reduction.ll index 326ffc8828d..bfa48a2529b 100644 --- a/llvm/test/Transforms/LoopVectorize/no_idiv_reduction.ll +++ b/llvm/test/Transforms/LoopVectorize/no_idiv_reduction.ll @@ -12,7 +12,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %r.05 = phi i32 [ 80, %entry ], [ %div, %for.body ] %arrayidx = getelementptr inbounds [128 x i32], [128 x i32]* @a, i64 0, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %div = sdiv i32 %r.05, %0 %indvars.iv.next = add i64 %indvars.iv, 1 %lftr.wideiv = trunc i64 %indvars.iv.next to i32 diff --git a/llvm/test/Transforms/LoopVectorize/no_int_induction.ll b/llvm/test/Transforms/LoopVectorize/no_int_induction.ll index 02848a067a2..7e6b26cd900 100644 --- a/llvm/test/Transforms/LoopVectorize/no_int_induction.ll +++ b/llvm/test/Transforms/LoopVectorize/no_int_induction.ll @@ -21,7 +21,7 @@ define i32 @sum_array(i32* %A, i32 %n) nounwind uwtable readonly noinline ssp { .lr.ph.i: ; preds = %0, %.lr.ph.i %.03.i = phi i32* [ %6, %.lr.ph.i ], [ %A, %0 ] %.012.i = phi i32 [ %5, %.lr.ph.i ], [ 0, %0 ] - %4 = load i32* %.03.i, align 4 + %4 = load i32, i32* %.03.i, align 4 %5 = add nsw i32 %4, %.012.i %6 = getelementptr inbounds i32, i32* %.03.i, i64 1 %7 = icmp eq i32* %6, %2 @@ -48,7 +48,7 @@ define i32 @sum_array_as1(i32 addrspace(1)* %A, i32 %n) nounwind uwtable readonl .lr.ph.i: ; preds = %0, %.lr.ph.i %.03.i = phi i32 addrspace(1)* [ %6, %.lr.ph.i ], [ %A, %0 ] %.012.i = phi i32 [ %5, %.lr.ph.i ], [ 0, %0 ] - %4 = load i32 addrspace(1)* %.03.i, align 4 + %4 = load i32, i32 addrspace(1)* %.03.i, align 4 %5 = add nsw i32 %4, %.012.i %6 = getelementptr inbounds i32, i32 addrspace(1)* %.03.i, i64 1 %7 = icmp eq i32 addrspace(1)* %6, %2 diff --git a/llvm/test/Transforms/LoopVectorize/no_outside_user.ll b/llvm/test/Transforms/LoopVectorize/no_outside_user.ll index bcd29c1a439..7030b6b4df2 100644 --- a/llvm/test/Transforms/LoopVectorize/no_outside_user.ll +++ b/llvm/test/Transforms/LoopVectorize/no_outside_user.ll @@ -20,7 +20,7 @@ target datalayout = "e-p:32:32:32-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-f3 define i32 @main() { bb: - %b.promoted = load i32* @b, align 4 + %b.promoted = load i32, i32* @b, align 4 br label %.lr.ph.i .lr.ph.i: @@ -56,7 +56,7 @@ f1.exit.loopexit: define i32 @test2() { entry: store i32 0, i32* @x1, align 4 - %0 = load i32* @x0, align 4 + %0 = load i32, i32* @x0, align 4 br label %for.cond1.preheader for.cond1.preheader: diff --git a/llvm/test/Transforms/LoopVectorize/no_switch.ll b/llvm/test/Transforms/LoopVectorize/no_switch.ll index e24e91f4ec1..76c1c0c16eb 100644 --- a/llvm/test/Transforms/LoopVectorize/no_switch.ll +++ b/llvm/test/Transforms/LoopVectorize/no_switch.ll @@ -22,7 +22,7 @@ for.body.preheader: ; preds = %entry for.body: ; preds = %for.body.preheader, %for.inc %indvars.iv = phi i64 [ %indvars.iv.next, %for.inc ], [ 0, %for.body.preheader ] %arrayidx = getelementptr inbounds i32, i32* %A, i64 %indvars.iv, !dbg !14 - %0 = load i32* %arrayidx, align 4, !dbg !14, !tbaa !16 + %0 = load i32, i32* %arrayidx, align 4, !dbg !14, !tbaa !16 switch i32 %0, label %for.inc [ i32 0, label %sw.bb i32 1, label %sw.bb3 diff --git a/llvm/test/Transforms/LoopVectorize/non-const-n.ll b/llvm/test/Transforms/LoopVectorize/non-const-n.ll index 4b957758d2f..9007b1b00fd 100644 --- a/llvm/test/Transforms/LoopVectorize/non-const-n.ll +++ b/llvm/test/Transforms/LoopVectorize/non-const-n.ll @@ -21,9 +21,9 @@ define void @example1(i32 %n) nounwind uwtable ssp { ; <label>:1 ; preds = %1, %0 %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %2 = getelementptr inbounds [2048 x i32], [2048 x i32]* @b, i64 0, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = getelementptr inbounds [2048 x i32], [2048 x i32]* @c, i64 0, i64 %indvars.iv - %5 = load i32* %4, align 4 + %5 = load i32, i32* %4, align 4 %6 = add nsw i32 %5, %3 %7 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv store i32 %6, i32* %7, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/opt.ll b/llvm/test/Transforms/LoopVectorize/opt.ll index 90d8b098caf..71b7e1f906b 100644 --- a/llvm/test/Transforms/LoopVectorize/opt.ll +++ b/llvm/test/Transforms/LoopVectorize/opt.ll @@ -17,7 +17,7 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %red.05 = phi i32 [ 0, %entry ], [ %add, %for.body ] %arrayidx = getelementptr inbounds i32, i32* %a, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %add = add nsw i32 %0, %red.05 %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1 %exitcond = icmp eq i64 %indvars.iv.next, 255 diff --git a/llvm/test/Transforms/LoopVectorize/ptr_loops.ll b/llvm/test/Transforms/LoopVectorize/ptr_loops.ll index 4d7e315e2f0..62fc1d9cc69 100644 --- a/llvm/test/Transforms/LoopVectorize/ptr_loops.ll +++ b/llvm/test/Transforms/LoopVectorize/ptr_loops.ll @@ -17,7 +17,7 @@ define i32 @_Z5test1v() nounwind uwtable ssp { ; <label>:1 ; preds = %0, %1 %p.02 = phi i32* [ getelementptr inbounds ([36 x i32]* @A, i64 0, i64 18), %0 ], [ %4, %1 ] %b.01 = phi i32* [ getelementptr inbounds ([36 x i32]* @B, i64 0, i64 0), %0 ], [ %5, %1 ] - %2 = load i32* %b.01, align 4 + %2 = load i32, i32* %b.01, align 4 %3 = shl nsw i32 %2, 1 store i32 %3, i32* %p.02, align 4 %4 = getelementptr inbounds i32, i32* %p.02, i64 -1 @@ -40,7 +40,7 @@ define i32 @_Z5test2v() nounwind uwtable ssp { ; <label>:1 ; preds = %0, %1 %p.02 = phi i32* [ getelementptr inbounds ([36 x i32]* @A, i64 0, i64 25), %0 ], [ %3, %1 ] %b.01 = phi i32* [ getelementptr inbounds ([36 x i32]* @B, i64 0, i64 2), %0 ], [ %4, %1 ] - %2 = load i32* %b.01, align 4 + %2 = load i32, i32* %b.01, align 4 store i32 %2, i32* %p.02, align 4 %3 = getelementptr inbounds i32, i32* %p.02, i64 -1 %4 = getelementptr inbounds i32, i32* %b.01, i64 1 @@ -62,7 +62,7 @@ define i32 @_Z5test3v() nounwind uwtable ssp { ; <label>:1 ; preds = %0, %1 %p.02 = phi i32* [ getelementptr inbounds ([36 x i32]* @A, i64 0, i64 29), %0 ], [ %3, %1 ] %b.01 = phi i32* [ getelementptr inbounds ([36 x i32]* @B, i64 0, i64 5), %0 ], [ %4, %1 ] - %2 = load i32* %b.01, align 4 + %2 = load i32, i32* %b.01, align 4 store i32 %2, i32* %p.02, align 4 %3 = getelementptr inbounds i32, i32* %p.02, i64 -1 %4 = getelementptr inbounds i32, i32* %b.01, i64 1 diff --git a/llvm/test/Transforms/LoopVectorize/read-only.ll b/llvm/test/Transforms/LoopVectorize/read-only.ll index ef12edb059a..f81afd6ad34 100644 --- a/llvm/test/Transforms/LoopVectorize/read-only.ll +++ b/llvm/test/Transforms/LoopVectorize/read-only.ll @@ -14,10 +14,10 @@ define i32 @read_only_func(i32* nocapture %A, i32* nocapture %B, i32 %n) nounwin %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 0, %0 ] %sum.02 = phi i32 [ %9, %.lr.ph ], [ 0, %0 ] %2 = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = add nsw i64 %indvars.iv, 13 %5 = getelementptr inbounds i32, i32* %B, i64 %4 - %6 = load i32* %5, align 4 + %6 = load i32, i32* %5, align 4 %7 = shl i32 %6, 1 %8 = add i32 %3, %sum.02 %9 = add i32 %8, %7 diff --git a/llvm/test/Transforms/LoopVectorize/reduction.ll b/llvm/test/Transforms/LoopVectorize/reduction.ll index 70c63feea40..647e58a7e41 100644 --- a/llvm/test/Transforms/LoopVectorize/reduction.ll +++ b/llvm/test/Transforms/LoopVectorize/reduction.ll @@ -21,9 +21,9 @@ define i32 @reduction_sum(i32 %n, i32* noalias nocapture %A, i32* noalias nocapt %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 0, %0 ] %sum.02 = phi i32 [ %9, %.lr.ph ], [ 0, %0 ] %2 = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = getelementptr inbounds i32, i32* %B, i64 %indvars.iv - %5 = load i32* %4, align 4 + %5 = load i32, i32* %4, align 4 %6 = trunc i64 %indvars.iv to i32 %7 = add i32 %sum.02, %6 %8 = add i32 %7, %3 @@ -56,9 +56,9 @@ define i32 @reduction_prod(i32 %n, i32* noalias nocapture %A, i32* noalias nocap %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 0, %0 ] %prod.02 = phi i32 [ %9, %.lr.ph ], [ 1, %0 ] %2 = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = getelementptr inbounds i32, i32* %B, i64 %indvars.iv - %5 = load i32* %4, align 4 + %5 = load i32, i32* %4, align 4 %6 = trunc i64 %indvars.iv to i32 %7 = mul i32 %prod.02, %6 %8 = mul i32 %7, %3 @@ -91,9 +91,9 @@ define i32 @reduction_mix(i32 %n, i32* noalias nocapture %A, i32* noalias nocapt %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 0, %0 ] %sum.02 = phi i32 [ %9, %.lr.ph ], [ 0, %0 ] %2 = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = getelementptr inbounds i32, i32* %B, i64 %indvars.iv - %5 = load i32* %4, align 4 + %5 = load i32, i32* %4, align 4 %6 = mul nsw i32 %5, %3 %7 = trunc i64 %indvars.iv to i32 %8 = add i32 %sum.02, %7 @@ -124,9 +124,9 @@ define i32 @reduction_mul(i32 %n, i32* noalias nocapture %A, i32* noalias nocapt %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 0, %0 ] %sum.02 = phi i32 [ %9, %.lr.ph ], [ 19, %0 ] %2 = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = getelementptr inbounds i32, i32* %B, i64 %indvars.iv - %5 = load i32* %4, align 4 + %5 = load i32, i32* %4, align 4 %6 = trunc i64 %indvars.iv to i32 %7 = add i32 %3, %6 %8 = add i32 %7, %5 @@ -159,9 +159,9 @@ for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %sum.09 = phi i32 [ %add, %for.body ], [ 120, %entry ] %arrayidx = getelementptr inbounds i32, i32* %in, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds i32, i32* %coeff, i64 %indvars.iv - %1 = load i32* %arrayidx2, align 4 + %1 = load i32, i32* %arrayidx2, align 4 %mul = mul nsw i32 %1, %0 %add = add nsw i32 %mul, %sum.09 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -192,9 +192,9 @@ for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %result.08 = phi i32 [ %and, %for.body ], [ -1, %entry ] %arrayidx = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds i32, i32* %B, i64 %indvars.iv - %1 = load i32* %arrayidx2, align 4 + %1 = load i32, i32* %arrayidx2, align 4 %add = add nsw i32 %1, %0 %and = and i32 %add, %result.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -224,9 +224,9 @@ for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %result.08 = phi i32 [ %or, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds i32, i32* %B, i64 %indvars.iv - %1 = load i32* %arrayidx2, align 4 + %1 = load i32, i32* %arrayidx2, align 4 %add = add nsw i32 %1, %0 %or = or i32 %add, %result.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -256,9 +256,9 @@ for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %result.08 = phi i32 [ %xor, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds i32, i32* %B, i64 %indvars.iv - %1 = load i32* %arrayidx2, align 4 + %1 = load i32, i32* %arrayidx2, align 4 %add = add nsw i32 %1, %0 %xor = xor i32 %add, %result.08 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -285,7 +285,7 @@ for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %x.05 = phi i32 [ %sub, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %sub = sub nsw i32 %0, %x.05 %indvars.iv.next = add i64 %indvars.iv, 1 %lftr.wideiv = trunc i64 %indvars.iv.next to i32 @@ -312,7 +312,7 @@ for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %x.05 = phi i32 [ %sub, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds i32, i32* %A, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %sub = sub nsw i32 %x.05, %0 %indvars.iv.next = add i64 %indvars.iv, 1 %lftr.wideiv = trunc i64 %indvars.iv.next to i32 @@ -336,9 +336,9 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.inc ] %sum.033 = phi float [ %S, %entry ], [ %sum.1, %for.inc ] %arrayidx = getelementptr inbounds float, float* %A, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds float, float* %B, i64 %indvars.iv - %1 = load float* %arrayidx2, align 4 + %1 = load float, float* %arrayidx2, align 4 %cmp3 = fcmp ogt float %0, %1 br i1 %cmp3, label %if.then, label %for.inc @@ -381,9 +381,9 @@ for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.inc ] %sum.033 = phi float [ %S, %entry ], [ %sum.1, %for.inc ] %arrayidx = getelementptr inbounds float, float* %A, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds float, float* %B, i64 %indvars.iv - %1 = load float* %arrayidx2, align 4 + %1 = load float, float* %arrayidx2, align 4 %cmp3 = fcmp ogt float %0, %1 br i1 %cmp3, label %if.then, label %for.inc @@ -428,7 +428,7 @@ for.body: %sum2.09 = phi float [ 0.000000e+00, %entry ], [ %add1, %for.body ] %sum.08 = phi float [ %S, %entry ], [ %add, %for.body ] %arrayidx = getelementptr inbounds float, float* %B, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %add = fadd fast float %sum.08, %0 %add1 = fadd fast float %sum2.09, %add %indvars.iv.next = add i64 %indvars.iv, 1 diff --git a/llvm/test/Transforms/LoopVectorize/reverse_induction.ll b/llvm/test/Transforms/LoopVectorize/reverse_induction.ll index 943fb9e59a3..6b63a0d8db6 100644 --- a/llvm/test/Transforms/LoopVectorize/reverse_induction.ll +++ b/llvm/test/Transforms/LoopVectorize/reverse_induction.ll @@ -19,7 +19,7 @@ for.body: %redux5 = phi i32 [ 0, %entry ], [ %inc.redux, %for.body ] %add.i = add i64 %add.i7, -1 %kind_.i = getelementptr inbounds i32, i32* %ptr, i64 %add.i - %tmp.i1 = load i32* %kind_.i, align 4 + %tmp.i1 = load i32, i32* %kind_.i, align 4 %inc.redux = add i32 %tmp.i1, %redux5 %inc4 = add i32 %i.06, 1 %exitcond = icmp ne i32 %inc4, 1024 @@ -42,7 +42,7 @@ for.body: %redux5 = phi i32 [ 0, %entry ], [ %inc.redux, %for.body ] %add.i = add i128 %add.i7, -1 %kind_.i = getelementptr inbounds i32, i32* %ptr, i128 %add.i - %tmp.i1 = load i32* %kind_.i, align 4 + %tmp.i1 = load i32, i32* %kind_.i, align 4 %inc.redux = add i32 %tmp.i1, %redux5 %inc4 = add i32 %i.06, 1 %exitcond = icmp ne i32 %inc4, 1024 @@ -66,7 +66,7 @@ for.body: %redux5 = phi i32 [ 0, %entry ], [ %inc.redux, %for.body ] %add.i = add i16 %add.i7, -1 %kind_.i = getelementptr inbounds i32, i32* %ptr, i16 %add.i - %tmp.i1 = load i32* %kind_.i, align 4 + %tmp.i1 = load i32, i32* %kind_.i, align 4 %inc.redux = add i32 %tmp.i1, %redux5 %inc4 = add i32 %i.06, 1 %exitcond = icmp ne i32 %inc4, 1024 diff --git a/llvm/test/Transforms/LoopVectorize/reverse_iter.ll b/llvm/test/Transforms/LoopVectorize/reverse_iter.ll index b0ecb2e0716..5bbc769a20a 100644 --- a/llvm/test/Transforms/LoopVectorize/reverse_iter.ll +++ b/llvm/test/Transforms/LoopVectorize/reverse_iter.ll @@ -31,7 +31,7 @@ define i32 @foo(i32 %n, i32* nocapture %A) { %5 = shl nsw i32 %4, 1 %6 = sext i32 %5 to i64 %7 = getelementptr inbounds i32, i32* %A, i64 %6 - %8 = load i32* %7, align 4 + %8 = load i32, i32* %7, align 4 %9 = add nsw i32 %8, %sum.01 %indvars.iv.next = add i64 %indvars.iv, -1 %10 = trunc i64 %indvars.iv.next to i32 diff --git a/llvm/test/Transforms/LoopVectorize/runtime-check-address-space.ll b/llvm/test/Transforms/LoopVectorize/runtime-check-address-space.ll index 1d336477172..8e7ac1f865a 100644 --- a/llvm/test/Transforms/LoopVectorize/runtime-check-address-space.ll +++ b/llvm/test/Transforms/LoopVectorize/runtime-check-address-space.ll @@ -38,7 +38,7 @@ for.body: ; preds = %entry, %for.body %i.02 = phi i32 [ %inc, %for.body ], [ 0, %entry ] %idxprom = sext i32 %i.02 to i64 %arrayidx = getelementptr inbounds i32, i32 addrspace(1)* %b, i64 %idxprom - %0 = load i32 addrspace(1)* %arrayidx, align 4 + %0 = load i32, i32 addrspace(1)* %arrayidx, align 4 %mul = mul nsw i32 %0, 3 %idxprom1 = sext i32 %i.02 to i64 %arrayidx2 = getelementptr inbounds i32, i32 addrspace(1)* %a, i64 %idxprom1 @@ -65,7 +65,7 @@ for.body: ; preds = %entry, %for.body %i.02 = phi i32 [ %inc, %for.body ], [ 0, %entry ] %idxprom = sext i32 %i.02 to i64 %arrayidx = getelementptr inbounds i32, i32 addrspace(1)* %b, i64 %idxprom - %0 = load i32 addrspace(1)* %arrayidx, align 4 + %0 = load i32, i32 addrspace(1)* %arrayidx, align 4 %mul = mul nsw i32 %0, 3 %idxprom1 = sext i32 %i.02 to i64 %arrayidx2 = getelementptr inbounds i32, i32* %a, i64 %idxprom1 @@ -92,7 +92,7 @@ for.body: ; preds = %entry, %for.body %i.02 = phi i32 [ %inc, %for.body ], [ 0, %entry ] %idxprom = sext i32 %i.02 to i64 %arrayidx = getelementptr inbounds i32, i32* %b, i64 %idxprom - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %mul = mul nsw i32 %0, 3 %idxprom1 = sext i32 %i.02 to i64 %arrayidx2 = getelementptr inbounds i32, i32 addrspace(1)* %a, i64 %idxprom1 @@ -120,7 +120,7 @@ for.body: ; preds = %entry, %for.body %i.02 = phi i32 [ %inc, %for.body ], [ 0, %entry ] %idxprom = sext i32 %i.02 to i64 %arrayidx = getelementptr inbounds i32, i32 addrspace(1)* %b, i64 %idxprom - %0 = load i32 addrspace(1)* %arrayidx, align 4 + %0 = load i32, i32 addrspace(1)* %arrayidx, align 4 %mul = mul nsw i32 %0, 3 %idxprom1 = sext i32 %i.02 to i64 %arrayidx2 = getelementptr inbounds i32, i32* %a, i64 %idxprom1 @@ -148,7 +148,7 @@ for.body: ; preds = %entry, %for.body %i.02 = phi i32 [ %inc, %for.body ], [ 0, %entry ] %idxprom = sext i32 %i.02 to i64 %arrayidx = getelementptr inbounds i32, i32* %b, i64 %idxprom - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %mul = mul nsw i32 %0, 3 %idxprom1 = sext i32 %i.02 to i64 %arrayidx2 = getelementptr inbounds [1024 x i32], [1024 x i32] addrspace(1)* @g_as1, i64 0, i64 %idxprom1 @@ -177,7 +177,7 @@ for.body: ; preds = %entry, %for.body %i.02 = phi i32 [ %inc, %for.body ], [ 0, %entry ] %idxprom = sext i32 %i.02 to i64 %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32] addrspace(1)* @g_as1, i64 0, i64 %idxprom - %0 = load i32 addrspace(1)* %arrayidx, align 4 + %0 = load i32, i32 addrspace(1)* %arrayidx, align 4 %mul = mul nsw i32 %0, 3 %idxprom1 = sext i32 %i.02 to i64 %arrayidx2 = getelementptr inbounds i32, i32* %b, i64 %idxprom1 @@ -205,7 +205,7 @@ for.body: ; preds = %entry, %for.body %i.02 = phi i32 [ %inc, %for.body ], [ 0, %entry ] %idxprom = sext i32 %i.02 to i64 %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32] addrspace(2)* @q_as2, i64 0, i64 %idxprom - %0 = load i32 addrspace(2)* %arrayidx, align 4 + %0 = load i32, i32 addrspace(2)* %arrayidx, align 4 %mul = mul nsw i32 %0, 3 %idxprom1 = sext i32 %i.02 to i64 %arrayidx2 = getelementptr inbounds [1024 x i32], [1024 x i32] addrspace(1)* @g_as1, i64 0, i64 %idxprom1 diff --git a/llvm/test/Transforms/LoopVectorize/runtime-check-readonly-address-space.ll b/llvm/test/Transforms/LoopVectorize/runtime-check-readonly-address-space.ll index bc3345bf145..6ee983d22c6 100644 --- a/llvm/test/Transforms/LoopVectorize/runtime-check-readonly-address-space.ll +++ b/llvm/test/Transforms/LoopVectorize/runtime-check-readonly-address-space.ll @@ -15,9 +15,9 @@ entry: for.body: ; preds = %entry, %for.body %i.01 = phi i64 [ 0, %entry ], [ %inc, %for.body ] %arrayidx = getelementptr inbounds i32, i32 addrspace(1)* %b, i64 %i.01 - %0 = load i32 addrspace(1)* %arrayidx, align 4 + %0 = load i32, i32 addrspace(1)* %arrayidx, align 4 %arrayidx1 = getelementptr inbounds i32, i32 addrspace(1)* %c, i64 %i.01 - %1 = load i32 addrspace(1)* %arrayidx1, align 4 + %1 = load i32, i32 addrspace(1)* %arrayidx1, align 4 %add = add nsw i32 %0, %1 %arrayidx2 = getelementptr inbounds i32, i32 addrspace(1)* %a, i64 %i.01 store i32 %add, i32 addrspace(1)* %arrayidx2, align 4 @@ -40,9 +40,9 @@ entry: for.body: ; preds = %entry, %for.body %i.01 = phi i64 [ 0, %entry ], [ %inc, %for.body ] %arrayidx = getelementptr inbounds i32, i32* %b, i64 %i.01 - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %arrayidx1 = getelementptr inbounds i32, i32* %c, i64 %i.01 - %1 = load i32* %arrayidx1, align 4 + %1 = load i32, i32* %arrayidx1, align 4 %add = add nsw i32 %0, %1 %arrayidx2 = getelementptr inbounds i32, i32 addrspace(1)* %a, i64 %i.01 store i32 %add, i32 addrspace(1)* %arrayidx2, align 4 @@ -65,9 +65,9 @@ entry: for.body: ; preds = %entry, %for.body %i.01 = phi i64 [ 0, %entry ], [ %inc, %for.body ] %arrayidx = getelementptr inbounds i32, i32 addrspace(1)* %b, i64 %i.01 - %0 = load i32 addrspace(1)* %arrayidx, align 4 + %0 = load i32, i32 addrspace(1)* %arrayidx, align 4 %arrayidx1 = getelementptr inbounds i32, i32* %c, i64 %i.01 - %1 = load i32* %arrayidx1, align 4 + %1 = load i32, i32* %arrayidx1, align 4 %add = add nsw i32 %0, %1 %arrayidx2 = getelementptr inbounds i32, i32* %a, i64 %i.01 store i32 %add, i32* %arrayidx2, align 4 @@ -90,9 +90,9 @@ entry: for.body: ; preds = %entry, %for.body %i.01 = phi i64 [ 0, %entry ], [ %inc, %for.body ] %arrayidx = getelementptr inbounds i32, i32 addrspace(1)* %b, i64 %i.01 - %0 = load i32 addrspace(1)* %arrayidx, align 4 + %0 = load i32, i32 addrspace(1)* %arrayidx, align 4 %arrayidx1 = getelementptr inbounds i32, i32 addrspace(1)* %c, i64 %i.01 - %1 = load i32 addrspace(1)* %arrayidx1, align 4 + %1 = load i32, i32 addrspace(1)* %arrayidx1, align 4 %add = add nsw i32 %0, %1 %arrayidx2 = getelementptr inbounds i32, i32* %a, i64 %i.01 store i32 %add, i32* %arrayidx2, align 4 @@ -115,9 +115,9 @@ entry: for.body: ; preds = %entry, %for.body %i.01 = phi i64 [ 0, %entry ], [ %inc, %for.body ] %arrayidx = getelementptr inbounds i32, i32 addrspace(1)* %b, i64 %i.01 - %0 = load i32 addrspace(1)* %arrayidx, align 4 + %0 = load i32, i32 addrspace(1)* %arrayidx, align 4 %arrayidx1 = getelementptr inbounds i32, i32 addrspace(2)* %c, i64 %i.01 - %1 = load i32 addrspace(2)* %arrayidx1, align 4 + %1 = load i32, i32 addrspace(2)* %arrayidx1, align 4 %add = add nsw i32 %0, %1 %arrayidx2 = getelementptr inbounds i32, i32* %a, i64 %i.01 store i32 %add, i32* %arrayidx2, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/runtime-check-readonly.ll b/llvm/test/Transforms/LoopVectorize/runtime-check-readonly.ll index 75a0d56dc6f..a3b5a598d22 100644 --- a/llvm/test/Transforms/LoopVectorize/runtime-check-readonly.ll +++ b/llvm/test/Transforms/LoopVectorize/runtime-check-readonly.ll @@ -23,9 +23,9 @@ entry: for.body: %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds i32, i32* %B, i64 %indvars.iv - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %arrayidx2 = getelementptr inbounds i32, i32* %C, i64 %indvars.iv - %1 = load i32* %arrayidx2, align 4 + %1 = load i32, i32* %arrayidx2, align 4 %add = add nsw i32 %1, %0 %arrayidx4 = getelementptr inbounds i32, i32* %A, i64 %indvars.iv store i32 %add, i32* %arrayidx4, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/runtime-check.ll b/llvm/test/Transforms/LoopVectorize/runtime-check.ll index 483d35a31dc..b1c5d402c45 100644 --- a/llvm/test/Transforms/LoopVectorize/runtime-check.ll +++ b/llvm/test/Transforms/LoopVectorize/runtime-check.ll @@ -22,7 +22,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %b, i64 %indvars.iv - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %mul = fmul float %0, 3.000000e+00 %arrayidx2 = getelementptr inbounds float, float* %a, i64 %indvars.iv store float %mul, float* %arrayidx2, align 4 @@ -48,10 +48,10 @@ for.body: %iv = phi i64 [ 0, %entry ], [ %iv.next, %for.body ] %ind.sum = add i64 %iv, %offset %arr.idx = getelementptr inbounds float, float* %a, i64 %ind.sum - %l1 = load float* %arr.idx, align 4 + %l1 = load float, float* %arr.idx, align 4 %ind.sum2 = add i64 %iv, %offset2 %arr.idx2 = getelementptr inbounds float, float* %a, i64 %ind.sum2 - %l2 = load float* %arr.idx2, align 4 + %l2 = load float, float* %arr.idx2, align 4 %m = fmul fast float %b, %l2 %ad = fadd fast float %l1, %m store float %ad, float* %arr.idx, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/runtime-limit.ll b/llvm/test/Transforms/LoopVectorize/runtime-limit.ll index 04b44cd9f86..6bc71e160cc 100644 --- a/llvm/test/Transforms/LoopVectorize/runtime-limit.ll +++ b/llvm/test/Transforms/LoopVectorize/runtime-limit.ll @@ -14,18 +14,18 @@ entry: for.body: ; preds = %for.body, %entry %i.016 = phi i64 [ 0, %entry ], [ %inc, %for.body ] %arrayidx = getelementptr inbounds i32, i32* %A, i64 %i.016 - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %arrayidx1 = getelementptr inbounds i32, i32* %B, i64 %i.016 - %1 = load i32* %arrayidx1, align 4 + %1 = load i32, i32* %arrayidx1, align 4 %add = add nsw i32 %1, %0 %arrayidx2 = getelementptr inbounds i32, i32* %C, i64 %i.016 - %2 = load i32* %arrayidx2, align 4 + %2 = load i32, i32* %arrayidx2, align 4 %add3 = add nsw i32 %add, %2 %arrayidx4 = getelementptr inbounds i32, i32* %E, i64 %i.016 - %3 = load i32* %arrayidx4, align 4 + %3 = load i32, i32* %arrayidx4, align 4 %add5 = add nsw i32 %add3, %3 %arrayidx6 = getelementptr inbounds i32, i32* %F, i64 %i.016 - %4 = load i32* %arrayidx6, align 4 + %4 = load i32, i32* %arrayidx6, align 4 %add7 = add nsw i32 %add5, %4 %arrayidx8 = getelementptr inbounds i32, i32* %out, i64 %i.016 store i32 %add7, i32* %arrayidx8, align 4 @@ -48,29 +48,29 @@ entry: for.body: ; preds = %for.body, %entry %i.037 = phi i64 [ 0, %entry ], [ %inc, %for.body ] %arrayidx = getelementptr inbounds i32, i32* %A, i64 %i.037 - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %arrayidx1 = getelementptr inbounds i32, i32* %B, i64 %i.037 - %1 = load i32* %arrayidx1, align 4 + %1 = load i32, i32* %arrayidx1, align 4 %add = add nsw i32 %1, %0 %arrayidx2 = getelementptr inbounds i32, i32* %C, i64 %i.037 - %2 = load i32* %arrayidx2, align 4 + %2 = load i32, i32* %arrayidx2, align 4 %add3 = add nsw i32 %add, %2 %arrayidx4 = getelementptr inbounds i32, i32* %E, i64 %i.037 - %3 = load i32* %arrayidx4, align 4 + %3 = load i32, i32* %arrayidx4, align 4 %add5 = add nsw i32 %add3, %3 %arrayidx6 = getelementptr inbounds i32, i32* %F, i64 %i.037 - %4 = load i32* %arrayidx6, align 4 + %4 = load i32, i32* %arrayidx6, align 4 %add7 = add nsw i32 %add5, %4 %arrayidx8 = getelementptr inbounds i32, i32* %out, i64 %i.037 store i32 %add7, i32* %arrayidx8, align 4 - %5 = load i32* %arrayidx, align 4 - %6 = load i32* %arrayidx1, align 4 + %5 = load i32, i32* %arrayidx, align 4 + %6 = load i32, i32* %arrayidx1, align 4 %add11 = add nsw i32 %6, %5 - %7 = load i32* %arrayidx2, align 4 + %7 = load i32, i32* %arrayidx2, align 4 %add13 = add nsw i32 %add11, %7 - %8 = load i32* %arrayidx4, align 4 + %8 = load i32, i32* %arrayidx4, align 4 %add15 = add nsw i32 %add13, %8 - %9 = load i32* %arrayidx6, align 4 + %9 = load i32, i32* %arrayidx6, align 4 %add17 = add nsw i32 %add15, %9 %arrayidx18 = getelementptr inbounds i32, i32* %out2, i64 %i.037 store i32 %add17, i32* %arrayidx18, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/safegep.ll b/llvm/test/Transforms/LoopVectorize/safegep.ll index 1003759541e..ecef8138f0b 100644 --- a/llvm/test/Transforms/LoopVectorize/safegep.ll +++ b/llvm/test/Transforms/LoopVectorize/safegep.ll @@ -18,9 +18,9 @@ entry: "<bb 3>": %i_15 = phi i32 [ 0, %entry ], [ %i_19, %"<bb 3>" ] %pp3 = getelementptr float, float* %A, i32 %i_15 - %D.1396_10 = load float* %pp3, align 4 + %D.1396_10 = load float, float* %pp3, align 4 %pp24 = getelementptr float, float* %B, i32 %i_15 - %D.1398_15 = load float* %pp24, align 4 + %D.1398_15 = load float, float* %pp24, align 4 %D.1399_17 = fadd float %D.1398_15, %K %D.1400_18 = fmul float %D.1396_10, %D.1399_17 store float %D.1400_18, float* %pp3, align 4 @@ -44,9 +44,9 @@ entry: "<bb 3>": %i_15 = phi i32 [ 0, %entry ], [ %i_19, %"<bb 3>" ] %pp3 = getelementptr float, float addrspace(5) * %A, i32 %i_15 - %D.1396_10 = load float addrspace(5) * %pp3, align 4 + %D.1396_10 = load float, float addrspace(5) * %pp3, align 4 %pp24 = getelementptr float, float* %B, i32 %i_15 - %D.1398_15 = load float* %pp24, align 4 + %D.1398_15 = load float, float* %pp24, align 4 %D.1399_17 = fadd float %D.1398_15, %K %D.1400_18 = fmul float %D.1396_10, %D.1399_17 store float %D.1400_18, float addrspace(5) * %pp3, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/same-base-access.ll b/llvm/test/Transforms/LoopVectorize/same-base-access.ll index fe94dd14b68..31cff0ee653 100644 --- a/llvm/test/Transforms/LoopVectorize/same-base-access.ll +++ b/llvm/test/Transforms/LoopVectorize/same-base-access.ll @@ -26,33 +26,33 @@ define i32 @kernel11(double* %x, double* %y, i32 %n) nounwind uwtable ssp { br label %4 ; <label>:4 ; preds = %25, %0 - %5 = load i32* %k, align 4 - %6 = load i32* %3, align 4 + %5 = load i32, i32* %k, align 4 + %6 = load i32, i32* %3, align 4 %7 = icmp slt i32 %5, %6 br i1 %7, label %8, label %28 ; <label>:8 ; preds = %4 - %9 = load i32* %k, align 4 + %9 = load i32, i32* %k, align 4 %10 = sub nsw i32 %9, 1 %11 = sext i32 %10 to i64 - %12 = load double** %1, align 8 + %12 = load double*, double** %1, align 8 %13 = getelementptr inbounds double, double* %12, i64 %11 - %14 = load double* %13, align 8 - %15 = load i32* %k, align 4 + %14 = load double, double* %13, align 8 + %15 = load i32, i32* %k, align 4 %16 = sext i32 %15 to i64 - %17 = load double** %2, align 8 + %17 = load double*, double** %2, align 8 %18 = getelementptr inbounds double, double* %17, i64 %16 - %19 = load double* %18, align 8 + %19 = load double, double* %18, align 8 %20 = fadd double %14, %19 - %21 = load i32* %k, align 4 + %21 = load i32, i32* %k, align 4 %22 = sext i32 %21 to i64 - %23 = load double** %1, align 8 + %23 = load double*, double** %1, align 8 %24 = getelementptr inbounds double, double* %23, i64 %22 store double %20, double* %24, align 8 br label %25 ; <label>:25 ; preds = %8 - %26 = load i32* %k, align 4 + %26 = load i32, i32* %k, align 4 %27 = add nsw i32 %26, 1 store i32 %27, i32* %k, align 4 br label %4 @@ -87,7 +87,7 @@ define i32 @func2(i32* nocapture %a) nounwind uwtable ssp { %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %7 ] %2 = mul nsw i64 %indvars.iv, 7 %3 = getelementptr inbounds i32, i32* %a, i64 %2 - %4 = load i32* %3, align 4 + %4 = load i32, i32* %3, align 4 %5 = icmp sgt i32 %4, 3 br i1 %5, label %6, label %7 diff --git a/llvm/test/Transforms/LoopVectorize/scalar-select.ll b/llvm/test/Transforms/LoopVectorize/scalar-select.ll index b10e80e21ce..b17b2028821 100644 --- a/llvm/test/Transforms/LoopVectorize/scalar-select.ll +++ b/llvm/test/Transforms/LoopVectorize/scalar-select.ll @@ -19,9 +19,9 @@ define void @example1(i1 %cond) nounwind uwtable ssp { ; <label>:1 ; preds = %1, %0 %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %2 = getelementptr inbounds [2048 x i32], [2048 x i32]* @b, i64 0, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = getelementptr inbounds [2048 x i32], [2048 x i32]* @c, i64 0, i64 %indvars.iv - %5 = load i32* %4, align 4 + %5 = load i32, i32* %4, align 4 %6 = add nsw i32 %5, %3 %7 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv %sel = select i1 %cond, i32 %6, i32 zeroinitializer diff --git a/llvm/test/Transforms/LoopVectorize/scev-exitlim-crash.ll b/llvm/test/Transforms/LoopVectorize/scev-exitlim-crash.ll index 58abcb0d350..cfc1d6d40a7 100644 --- a/llvm/test/Transforms/LoopVectorize/scev-exitlim-crash.ll +++ b/llvm/test/Transforms/LoopVectorize/scev-exitlim-crash.ll @@ -29,14 +29,14 @@ for.cond4.preheader: ; preds = %for.cond br i1 %cmp514, label %for.cond7.preheader.lr.ph, label %for.end26 for.cond7.preheader.lr.ph: ; preds = %for.cond4.preheader - %0 = load i32** @e, align 8, !tbaa !4 + %0 = load i32*, i32** @e, align 8, !tbaa !4 br label %for.cond7.preheader for.cond7.preheader: ; preds = %for.cond7.preheader.lr.ph, %for.inc23 %y.017 = phi i32 [ 0, %for.cond7.preheader.lr.ph ], [ %inc24, %for.inc23 ] %i.116 = phi i32 [ 0, %for.cond7.preheader.lr.ph ], [ %i.2.lcssa, %for.inc23 ] %n.015 = phi i32 [ undef, %for.cond7.preheader.lr.ph ], [ %inc25, %for.inc23 ] - %1 = load i32* @b, align 4, !tbaa !5 + %1 = load i32, i32* @b, align 4, !tbaa !5 %tobool11 = icmp eq i32 %1, 0 br i1 %tobool11, label %for.inc23, label %for.body8.lr.ph @@ -63,7 +63,7 @@ for.body13: ; preds = %for.body13.lr.ph, % %indvars.iv = phi i64 [ %3, %for.body13.lr.ph ], [ %indvars.iv.next, %for.body13 ] %add.ptr.sum = add i64 %idx.ext, %indvars.iv %arrayidx = getelementptr inbounds i32, i32* @a, i64 %add.ptr.sum - %4 = load i32* %arrayidx, align 4, !tbaa !5 + %4 = load i32, i32* %arrayidx, align 4, !tbaa !5 %arrayidx15 = getelementptr inbounds i32, i32* %0, i64 %indvars.iv store i32 %4, i32* %arrayidx15, align 4, !tbaa !5 %indvars.iv.next = add i64 %indvars.iv, 1 @@ -75,11 +75,11 @@ for.cond11.for.inc19_crit_edge: ; preds = %for.body13 br label %for.inc19 for.inc19: ; preds = %for.cond11.for.inc19_crit_edge, %for.body8 - %6 = load i32* @c, align 4, !tbaa !5 + %6 = load i32, i32* @c, align 4, !tbaa !5 %inc20 = add nsw i32 %6, 1 store i32 %inc20, i32* @c, align 4, !tbaa !5 %indvars.iv.next20 = add i64 %indvars.iv19, 1 - %7 = load i32* @b, align 4, !tbaa !5 + %7 = load i32, i32* @b, align 4, !tbaa !5 %tobool = icmp eq i32 %7, 0 br i1 %tobool, label %for.cond7.for.inc23_crit_edge, label %for.body8 diff --git a/llvm/test/Transforms/LoopVectorize/simple-unroll.ll b/llvm/test/Transforms/LoopVectorize/simple-unroll.ll index c925a053c13..fc996ed3818 100644 --- a/llvm/test/Transforms/LoopVectorize/simple-unroll.ll +++ b/llvm/test/Transforms/LoopVectorize/simple-unroll.ll @@ -24,7 +24,7 @@ define void @inc(i32 %n) nounwind uwtable noinline ssp { .lr.ph: ; preds = %0, %.lr.ph %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 0, %0 ] %2 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = trunc i64 %indvars.iv to i32 %5 = add nsw i32 %3, %4 store i32 %5, i32* %2, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/small-loop.ll b/llvm/test/Transforms/LoopVectorize/small-loop.ll index bc6e21f8118..ce606d1b360 100644 --- a/llvm/test/Transforms/LoopVectorize/small-loop.ll +++ b/llvm/test/Transforms/LoopVectorize/small-loop.ll @@ -16,9 +16,9 @@ define void @example1() nounwind uwtable ssp { ; <label>:1 ; preds = %1, %0 %indvars.iv = phi i64 [ 0, %0 ], [ %indvars.iv.next, %1 ] %2 = getelementptr inbounds [2048 x i32], [2048 x i32]* @b, i64 0, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = getelementptr inbounds [2048 x i32], [2048 x i32]* @c, i64 0, i64 %indvars.iv - %5 = load i32* %4, align 4 + %5 = load i32, i32* %4, align 4 %6 = add nsw i32 %5, %3 %7 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv store i32 %6, i32* %7, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/start-non-zero.ll b/llvm/test/Transforms/LoopVectorize/start-non-zero.ll index 12c15a9b3d8..b444a692319 100644 --- a/llvm/test/Transforms/LoopVectorize/start-non-zero.ll +++ b/llvm/test/Transforms/LoopVectorize/start-non-zero.ll @@ -18,7 +18,7 @@ for.body.lr.ph: ; preds = %entry for.body: ; preds = %for.body.lr.ph, %for.body %indvars.iv = phi i64 [ %0, %for.body.lr.ph ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds i32, i32* %a, i64 %indvars.iv - %1 = load i32* %arrayidx, align 4 + %1 = load i32, i32* %arrayidx, align 4 %mul = mul nuw i32 %1, 333 store i32 %mul, i32* %arrayidx, align 4 %indvars.iv.next = add i64 %indvars.iv, 1 diff --git a/llvm/test/Transforms/LoopVectorize/store-shuffle-bug.ll b/llvm/test/Transforms/LoopVectorize/store-shuffle-bug.ll index 2512d4f2f41..4d62df15717 100644 --- a/llvm/test/Transforms/LoopVectorize/store-shuffle-bug.ll +++ b/llvm/test/Transforms/LoopVectorize/store-shuffle-bug.ll @@ -32,12 +32,12 @@ for.body: %0 = add i64 %indvars.iv, 1 %arrayidx = getelementptr inbounds [100 x i32], [100 x i32]* @uf, i64 0, i64 %0 %arrayidx3 = getelementptr inbounds [100 x i32], [100 x i32]* @xi, i64 0, i64 %0 - %1 = load i32* %arrayidx3, align 4 - %2 = load i32* %arrayidx, align 4 + %1 = load i32, i32* %arrayidx3, align 4 + %2 = load i32, i32* %arrayidx, align 4 %add4 = add nsw i32 %2, %1 store i32 %add4, i32* %arrayidx, align 4 %arrayidx7 = getelementptr inbounds [100 x i32], [100 x i32]* @q, i64 0, i64 %0 - %3 = load i32* %arrayidx7, align 4 + %3 = load i32, i32* %arrayidx7, align 4 %add8 = add nsw i32 %add4, %3 store i32 %add8, i32* %arrayidx, align 4 %indvars.iv.next = add i64 %indvars.iv, -1 diff --git a/llvm/test/Transforms/LoopVectorize/struct_access.ll b/llvm/test/Transforms/LoopVectorize/struct_access.ll index 0d68e2d59a1..1e4019a7810 100644 --- a/llvm/test/Transforms/LoopVectorize/struct_access.ll +++ b/llvm/test/Transforms/LoopVectorize/struct_access.ll @@ -33,7 +33,7 @@ for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %sum.05 = phi i32 [ %add, %for.body ], [ 0, %entry ] %x = getelementptr inbounds %struct.coordinate, %struct.coordinate* %A, i64 %indvars.iv, i32 0 - %0 = load i32* %x, align 4 + %0 = load i32, i32* %x, align 4 %add = add nsw i32 %0, %sum.05 %indvars.iv.next = add i64 %indvars.iv, 1 %lftr.wideiv = trunc i64 %indvars.iv.next to i32 @@ -75,7 +75,7 @@ for.body: ; preds = %entry, %for.body %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %sum.05 = phi i32 [ %add, %for.body ], [ 0, %entry ] %x = getelementptr inbounds %struct.lit, %struct.lit* %A, i64 %indvars.iv, i32 0 - %0 = load i32* %x, align 4 + %0 = load i32, i32* %x, align 4 %add = add nsw i32 %0, %sum.05 %indvars.iv.next = add i64 %indvars.iv, 1 %lftr.wideiv = trunc i64 %indvars.iv.next to i32 diff --git a/llvm/test/Transforms/LoopVectorize/tbaa-nodep.ll b/llvm/test/Transforms/LoopVectorize/tbaa-nodep.ll index 8469a9d95c6..06d00023002 100644 --- a/llvm/test/Transforms/LoopVectorize/tbaa-nodep.ll +++ b/llvm/test/Transforms/LoopVectorize/tbaa-nodep.ll @@ -11,7 +11,7 @@ entry: for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds float, float* %b, i64 %indvars.iv - %0 = load float* %arrayidx, align 4, !tbaa !0 + %0 = load float, float* %arrayidx, align 4, !tbaa !0 %conv = fptosi float %0 to i32 %arrayidx2 = getelementptr inbounds i32, i32* %a, i64 %indvars.iv store i32 %conv, i32* %arrayidx2, align 4, !tbaa !4 @@ -30,7 +30,7 @@ for.end: ; preds = %for.body ; CHECK-NEXT: br label %vector.body ; CHECK: vector.body: -; CHECK: load <4 x float>* %{{.*}}, align 4, !tbaa +; CHECK: load <4 x float>, <4 x float>* %{{.*}}, align 4, !tbaa ; CHECK: store <4 x i32> %{{.*}}, <4 x i32>* %{{.*}}, align 4, !tbaa ; CHECK: ret i32 0 @@ -38,7 +38,7 @@ for.end: ; preds = %for.body ; CHECK-NOTBAA-LABEL: @test1 ; CHECK-NOTBAA: icmp uge i32* -; CHECK-NOTBAA: load <4 x float>* %{{.*}}, align 4, !tbaa +; CHECK-NOTBAA: load <4 x float>, <4 x float>* %{{.*}}, align 4, !tbaa ; CHECK-NOTBAA: store <4 x i32> %{{.*}}, <4 x i32>* %{{.*}}, align 4, !tbaa ; CHECK-NOTBAA: ret i32 0 @@ -52,9 +52,9 @@ entry: for.body: ; preds = %for.body, %entry %indvars.iv = phi i64 [ 0, %entry ], [ %indvars.iv.next, %for.body ] %arrayidx = getelementptr inbounds float, float* %b, i64 %indvars.iv - %0 = load float* %arrayidx, align 4, !tbaa !0 + %0 = load float, float* %arrayidx, align 4, !tbaa !0 %arrayidx2 = getelementptr inbounds i32, i32* %a, i64 %indvars.iv - %1 = load i32* %arrayidx2, align 4, !tbaa !4 + %1 = load i32, i32* %arrayidx2, align 4, !tbaa !4 %conv = sitofp i32 %1 to float %mul = fmul float %0, %conv %arrayidx4 = getelementptr inbounds float, float* %c, i64 %indvars.iv @@ -74,7 +74,7 @@ for.end: ; preds = %for.body ; CHECK: icmp uge float* ; CHECK-NOT: icmp uge i32* -; CHECK: load <4 x float>* %{{.*}}, align 4, !tbaa +; CHECK: load <4 x float>, <4 x float>* %{{.*}}, align 4, !tbaa ; CHECK: store <4 x float> %{{.*}}, <4 x float>* %{{.*}}, align 4, !tbaa ; CHECK: ret i32 0 @@ -85,7 +85,7 @@ for.end: ; preds = %for.body ; CHECK-NOTBAA-DAG: icmp uge float* ; CHECK-NOTBAA-DAG: icmp uge i32* -; CHECK-NOTBAA: load <4 x float>* %{{.*}}, align 4, !tbaa +; CHECK-NOTBAA: load <4 x float>, <4 x float>* %{{.*}}, align 4, !tbaa ; CHECK-NOTBAA: store <4 x float> %{{.*}}, <4 x float>* %{{.*}}, align 4, !tbaa ; CHECK-NOTBAA: ret i32 0 diff --git a/llvm/test/Transforms/LoopVectorize/unroll_novec.ll b/llvm/test/Transforms/LoopVectorize/unroll_novec.ll index fc23d3d98ef..c23ad77a31f 100644 --- a/llvm/test/Transforms/LoopVectorize/unroll_novec.ll +++ b/llvm/test/Transforms/LoopVectorize/unroll_novec.ll @@ -10,11 +10,11 @@ target triple = "x86_64-apple-macosx10.8.0" ; a[i] += i; ; } ;CHECK-LABEL: @inc( -;CHECK: load i32* -;CHECK: load i32* -;CHECK: load i32* -;CHECK: load i32* -;CHECK-NOT: load i32* +;CHECK: load i32, i32* +;CHECK: load i32, i32* +;CHECK: load i32, i32* +;CHECK: load i32, i32* +;CHECK-NOT: load i32, i32* ;CHECK: add nsw i32 ;CHECK: add nsw i32 ;CHECK: add nsw i32 @@ -34,7 +34,7 @@ define void @inc(i32 %n) nounwind uwtable noinline ssp { .lr.ph: ; preds = %0, %.lr.ph %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 0, %0 ] %2 = getelementptr inbounds [2048 x i32], [2048 x i32]* @a, i64 0, i64 %indvars.iv - %3 = load i32* %2, align 4 + %3 = load i32, i32* %2, align 4 %4 = trunc i64 %indvars.iv to i32 %5 = add nsw i32 %3, %4 store i32 %5, i32* %2, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/value-ptr-bug.ll b/llvm/test/Transforms/LoopVectorize/value-ptr-bug.ll index 4fb0318b3d5..ce4601f7b92 100644 --- a/llvm/test/Transforms/LoopVectorize/value-ptr-bug.ll +++ b/llvm/test/Transforms/LoopVectorize/value-ptr-bug.ll @@ -38,7 +38,7 @@ do.body272: %sp.4 = phi i8* [ %tmp30, %block1 ], [ %incdec.ptr273, %do.body272 ] %dp.addr.4 = phi i8* [ %tmp29, %block1 ], [ %incdec.ptr274, %do.body272 ] %incdec.ptr273 = getelementptr inbounds i8, i8* %sp.4, i64 1 - %tmp31 = load i8* %sp.4, align 1 + %tmp31 = load i8, i8* %sp.4, align 1 %incdec.ptr274 = getelementptr inbounds i8, i8* %dp.addr.4, i64 1 store i8 %tmp31, i8* %dp.addr.4, align 1 %dec = add i32 %row_width.5, -1 diff --git a/llvm/test/Transforms/LoopVectorize/vect.omp.persistence.ll b/llvm/test/Transforms/LoopVectorize/vect.omp.persistence.ll index 4ad34ada207..0d7f8b9398f 100644 --- a/llvm/test/Transforms/LoopVectorize/vect.omp.persistence.ll +++ b/llvm/test/Transforms/LoopVectorize/vect.omp.persistence.ll @@ -46,10 +46,10 @@ for.body: %0 = add nsw i64 %indvars.iv, -5 %arrayidx = getelementptr inbounds float, float* %a, i64 %0 - %1 = load float* %arrayidx, align 4, !llvm.mem.parallel_loop_access !1 + %1 = load float, float* %arrayidx, align 4, !llvm.mem.parallel_loop_access !1 %2 = add nsw i64 %indvars.iv, 2 %arrayidx2 = getelementptr inbounds float, float* %a, i64 %2 - %3 = load float* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !1 + %3 = load float, float* %arrayidx2, align 4, !llvm.mem.parallel_loop_access !1 %mul = fmul float %1, %3 %arrayidx4 = getelementptr inbounds float, float* %a, i64 %indvars.iv store float %mul, float* %arrayidx4, align 4, !llvm.mem.parallel_loop_access !1 diff --git a/llvm/test/Transforms/LoopVectorize/vect.stats.ll b/llvm/test/Transforms/LoopVectorize/vect.stats.ll index 4a748f28b7a..7bf5a608484 100644 --- a/llvm/test/Transforms/LoopVectorize/vect.stats.ll +++ b/llvm/test/Transforms/LoopVectorize/vect.stats.ll @@ -21,7 +21,7 @@ entry: for.body: ; preds = %entry, %for.body %indvars.iv2 = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %arrayidx = getelementptr inbounds float, float* %a, i64 %indvars.iv2 - %0 = load float* %arrayidx, align 4 + %0 = load float, float* %arrayidx, align 4 %mul = fmul float %0, %0 store float %mul, float* %arrayidx, align 4 %indvars.iv.next = add nuw nsw i64 %indvars.iv2, 1 @@ -43,10 +43,10 @@ for.body: ; preds = %entry, %for.body %indvars.iv2 = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %entry ] %0 = add nsw i64 %indvars.iv2, -5 %arrayidx = getelementptr inbounds float, float* %a, i64 %0 - %1 = load float* %arrayidx, align 4 + %1 = load float, float* %arrayidx, align 4 %2 = add nsw i64 %indvars.iv2, 2 %arrayidx2 = getelementptr inbounds float, float* %a, i64 %2 - %3 = load float* %arrayidx2, align 4 + %3 = load float, float* %arrayidx2, align 4 %mul = fmul float %1, %3 %arrayidx4 = getelementptr inbounds float, float* %a, i64 %indvars.iv2 store float %mul, float* %arrayidx4, align 4 diff --git a/llvm/test/Transforms/LoopVectorize/vectorize-once.ll b/llvm/test/Transforms/LoopVectorize/vectorize-once.ll index f5f39cc1f5b..5d0e96b2c5a 100644 --- a/llvm/test/Transforms/LoopVectorize/vectorize-once.ll +++ b/llvm/test/Transforms/LoopVectorize/vectorize-once.ll @@ -29,7 +29,7 @@ entry: for.body.i: ; preds = %entry, %for.body.i %__init.addr.05.i = phi i32 [ %add.i, %for.body.i ], [ 0, %entry ] %__first.addr.04.i = phi i32* [ %incdec.ptr.i, %for.body.i ], [ %A, %entry ] - %0 = load i32* %__first.addr.04.i, align 4 + %0 = load i32, i32* %__first.addr.04.i, align 4 %add.i = add nsw i32 %0, %__init.addr.05.i %incdec.ptr.i = getelementptr inbounds i32, i32* %__first.addr.04.i, i64 1 %cmp.i = icmp eq i32* %incdec.ptr.i, %add.ptr @@ -55,7 +55,7 @@ entry: for.body.i: ; preds = %entry, %for.body.i %__init.addr.05.i = phi i32 [ %add.i, %for.body.i ], [ 0, %entry ] %__first.addr.04.i = phi i32* [ %incdec.ptr.i, %for.body.i ], [ %A, %entry ] - %0 = load i32* %__first.addr.04.i, align 4 + %0 = load i32, i32* %__first.addr.04.i, align 4 %add.i = add nsw i32 %0, %__init.addr.05.i %incdec.ptr.i = getelementptr inbounds i32, i32* %__first.addr.04.i, i64 1 %cmp.i = icmp eq i32* %incdec.ptr.i, %add.ptr diff --git a/llvm/test/Transforms/LoopVectorize/version-mem-access.ll b/llvm/test/Transforms/LoopVectorize/version-mem-access.ll index 1a0b81ce9ea..a9d319e5a2d 100644 --- a/llvm/test/Transforms/LoopVectorize/version-mem-access.ll +++ b/llvm/test/Transforms/LoopVectorize/version-mem-access.ll @@ -35,10 +35,10 @@ for.body: %mul = mul i32 %iv.trunc, %BStride %mul64 = zext i32 %mul to i64 %arrayidx = getelementptr inbounds i32, i32* %B, i64 %mul64 - %0 = load i32* %arrayidx, align 4 + %0 = load i32, i32* %arrayidx, align 4 %mul2 = mul nsw i64 %indvars.iv, %CStride %arrayidx3 = getelementptr inbounds i32, i32* %C, i64 %mul2 - %1 = load i32* %arrayidx3, align 4 + %1 = load i32, i32* %arrayidx3, align 4 %mul4 = mul nsw i32 %1, %0 %mul3 = mul nsw i64 %indvars.iv, %AStride %arrayidx7 = getelementptr inbounds i32, i32* %A, i64 %mul3 @@ -77,7 +77,7 @@ for.body: %mul = mul nsw i32 %0, %conv %idxprom = sext i32 %mul to i64 %arrayidx = getelementptr inbounds double, double* %x, i64 %idxprom - %1 = load double* %arrayidx, align 8 + %1 = load double, double* %arrayidx, align 8 %arrayidx3 = getelementptr inbounds double, double* %c, i64 %indvars.iv store double %1, double* %arrayidx3, align 8 %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1 diff --git a/llvm/test/Transforms/LoopVectorize/write-only.ll b/llvm/test/Transforms/LoopVectorize/write-only.ll index cd1b33017af..b2bc0458a23 100644 --- a/llvm/test/Transforms/LoopVectorize/write-only.ll +++ b/llvm/test/Transforms/LoopVectorize/write-only.ll @@ -13,7 +13,7 @@ define i32 @read_mod_write_single_ptr(float* nocapture %a, i32 %n) nounwind uwta .lr.ph: ; preds = %0, %.lr.ph %indvars.iv = phi i64 [ %indvars.iv.next, %.lr.ph ], [ 0, %0 ] %2 = getelementptr inbounds float, float* %a, i64 %indvars.iv - %3 = load float* %2, align 4 + %3 = load float, float* %2, align 4 %4 = fmul float %3, 3.000000e+00 store float %4, float* %2, align 4 %indvars.iv.next = add i64 %indvars.iv, 1 |