summaryrefslogtreecommitdiffstats
path: root/llvm/test/Transforms/ObjCARC/nested.ll
diff options
context:
space:
mode:
Diffstat (limited to 'llvm/test/Transforms/ObjCARC/nested.ll')
-rw-r--r--llvm/test/Transforms/ObjCARC/nested.ll825
1 files changed, 0 insertions, 825 deletions
diff --git a/llvm/test/Transforms/ObjCARC/nested.ll b/llvm/test/Transforms/ObjCARC/nested.ll
deleted file mode 100644
index 8b7e673e11a..00000000000
--- a/llvm/test/Transforms/ObjCARC/nested.ll
+++ /dev/null
@@ -1,825 +0,0 @@
-; RUN: opt -objc-arc -S < %s | FileCheck %s
-
-%struct.__objcFastEnumerationState = type { i64, i8**, i64*, [5 x i64] }
-
-@"\01L_OBJC_METH_VAR_NAME_" = internal global [43 x i8] c"countByEnumeratingWithState:objects:count:\00", section "__TEXT,__objc_methname,cstring_literals", align 1
-@"\01L_OBJC_SELECTOR_REFERENCES_" = internal global i8* getelementptr inbounds ([43 x i8], [43 x i8]* @"\01L_OBJC_METH_VAR_NAME_", i64 0, i64 0), section "__DATA, __objc_selrefs, literal_pointers, no_dead_strip"
-@g = common global i8* null, align 8
-@"\01L_OBJC_IMAGE_INFO" = internal constant [2 x i32] [i32 0, i32 16], section "__DATA, __objc_imageinfo, regular, no_dead_strip"
-
-declare void @callee()
-declare i8* @returner()
-declare i8* @llvm.objc.retainAutoreleasedReturnValue(i8*)
-declare i8* @llvm.objc.retain(i8*)
-declare void @llvm.objc.enumerationMutation(i8*)
-declare void @llvm.memset.p0i8.i64(i8* nocapture, i8, i64, i1) nounwind
-declare i8* @llvm.objc.msgSend(i8*, i8*, ...) nonlazybind
-declare void @use(i8*)
-declare void @llvm.objc.release(i8*)
-declare i8* @def()
-declare void @__crasher_block_invoke(i8* nocapture)
-declare i8* @llvm.objc.retainBlock(i8*)
-declare void @__crasher_block_invoke1(i8* nocapture)
-
-!0 = !{}
-
-; Delete a nested retain+release pair.
-
-; CHECK-LABEL: define void @test0(
-; CHECK: call i8* @llvm.objc.retain
-; CHECK-NOT: @llvm.objc.retain
-; CHECK: }
-define void @test0(i8* %a) nounwind {
-entry:
- %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
- %items.ptr = alloca [16 x i8*], align 8
- %0 = call i8* @llvm.objc.retain(i8* %a) nounwind
- %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
- call void @llvm.memset.p0i8.i64(i8* align 8 %tmp, i8 0, i64 64, i1 false)
- %1 = call i8* @llvm.objc.retain(i8* %0) nounwind
- %tmp2 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp2, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %iszero = icmp eq i64 %call, 0
- br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
-
-forcoll.loopinit:
- %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
- %mutationsptr = load i64*, i64** %mutationsptr.ptr, align 8
- %forcoll.initial-mutations = load i64, i64* %mutationsptr, align 8
- %stateitems.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 1
- br label %forcoll.loopbody.outer
-
-forcoll.loopbody.outer:
- %forcoll.count.ph = phi i64 [ %call, %forcoll.loopinit ], [ %call6, %forcoll.refetch ]
- %tmp7 = icmp ugt i64 %forcoll.count.ph, 1
- %umax = select i1 %tmp7, i64 %forcoll.count.ph, i64 1
- br label %forcoll.loopbody
-
-forcoll.loopbody:
- %forcoll.index = phi i64 [ 0, %forcoll.loopbody.outer ], [ %4, %forcoll.notmutated ]
- %mutationsptr3 = load i64*, i64** %mutationsptr.ptr, align 8
- %statemutations = load i64, i64* %mutationsptr3, align 8
- %2 = icmp eq i64 %statemutations, %forcoll.initial-mutations
- br i1 %2, label %forcoll.notmutated, label %forcoll.mutated
-
-forcoll.mutated:
- call void @llvm.objc.enumerationMutation(i8* %1)
- br label %forcoll.notmutated
-
-forcoll.notmutated:
- %stateitems = load i8**, i8*** %stateitems.ptr, align 8
- %currentitem.ptr = getelementptr i8*, i8** %stateitems, i64 %forcoll.index
- %3 = load i8*, i8** %currentitem.ptr, align 8
- call void @use(i8* %3)
- %4 = add i64 %forcoll.index, 1
- %exitcond = icmp eq i64 %4, %umax
- br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
-
-forcoll.refetch:
- %tmp5 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call6 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp5, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %5 = icmp eq i64 %call6, 0
- br i1 %5, label %forcoll.empty, label %forcoll.loopbody.outer
-
-forcoll.empty:
- call void @llvm.objc.release(i8* %1) nounwind
- call void @llvm.objc.release(i8* %0) nounwind, !clang.imprecise_release !0
- ret void
-}
-
-; Delete a nested retain+release pair.
-
-; CHECK-LABEL: define void @test2(
-; CHECK: call i8* @llvm.objc.retainAutoreleasedReturnValue
-; CHECK-NOT: @llvm.objc.retain
-; CHECK: }
-define void @test2() nounwind {
-entry:
- %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
- %items.ptr = alloca [16 x i8*], align 8
- %call = call i8* @returner()
- %0 = call i8* @llvm.objc.retainAutoreleasedReturnValue(i8* %call) nounwind
- %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
- call void @llvm.memset.p0i8.i64(i8* align 8 %tmp, i8 0, i64 64, i1 false)
- %1 = call i8* @llvm.objc.retain(i8* %0) nounwind
- %tmp2 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call3 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp2, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %iszero = icmp eq i64 %call3, 0
- br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
-
-forcoll.loopinit:
- %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
- %mutationsptr = load i64*, i64** %mutationsptr.ptr, align 8
- %forcoll.initial-mutations = load i64, i64* %mutationsptr, align 8
- %stateitems.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 1
- br label %forcoll.loopbody.outer
-
-forcoll.loopbody.outer:
- %forcoll.count.ph = phi i64 [ %call3, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
- %tmp8 = icmp ugt i64 %forcoll.count.ph, 1
- %umax = select i1 %tmp8, i64 %forcoll.count.ph, i64 1
- br label %forcoll.loopbody
-
-forcoll.loopbody:
- %forcoll.index = phi i64 [ 0, %forcoll.loopbody.outer ], [ %4, %forcoll.notmutated ]
- %mutationsptr4 = load i64*, i64** %mutationsptr.ptr, align 8
- %statemutations = load i64, i64* %mutationsptr4, align 8
- %2 = icmp eq i64 %statemutations, %forcoll.initial-mutations
- br i1 %2, label %forcoll.notmutated, label %forcoll.mutated
-
-forcoll.mutated:
- call void @llvm.objc.enumerationMutation(i8* %1)
- br label %forcoll.notmutated
-
-forcoll.notmutated:
- %stateitems = load i8**, i8*** %stateitems.ptr, align 8
- %currentitem.ptr = getelementptr i8*, i8** %stateitems, i64 %forcoll.index
- %3 = load i8*, i8** %currentitem.ptr, align 8
- call void @use(i8* %3)
- %4 = add i64 %forcoll.index, 1
- %exitcond = icmp eq i64 %4, %umax
- br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
-
-forcoll.refetch:
- %tmp6 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %5 = icmp eq i64 %call7, 0
- br i1 %5, label %forcoll.empty, label %forcoll.loopbody.outer
-
-forcoll.empty:
- call void @llvm.objc.release(i8* %1) nounwind
- call void @llvm.objc.release(i8* %0) nounwind, !clang.imprecise_release !0
- ret void
-}
-
-; Delete a nested retain+release pair.
-
-; CHECK-LABEL: define void @test4(
-; CHECK: call i8* @llvm.objc.retain
-; CHECK-NOT: @llvm.objc.retain
-; CHECK: }
-define void @test4() nounwind {
-entry:
- %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
- %items.ptr = alloca [16 x i8*], align 8
- %tmp = load i8*, i8** @g, align 8
- %0 = call i8* @llvm.objc.retain(i8* %tmp) nounwind
- %tmp2 = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
- call void @llvm.memset.p0i8.i64(i8* align 8 %tmp2, i8 0, i64 64, i1 false)
- %1 = call i8* @llvm.objc.retain(i8* %0) nounwind
- %tmp4 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp4, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %iszero = icmp eq i64 %call, 0
- br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
-
-forcoll.loopinit:
- %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
- %mutationsptr = load i64*, i64** %mutationsptr.ptr, align 8
- %forcoll.initial-mutations = load i64, i64* %mutationsptr, align 8
- %stateitems.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 1
- br label %forcoll.loopbody.outer
-
-forcoll.loopbody.outer:
- %forcoll.count.ph = phi i64 [ %call, %forcoll.loopinit ], [ %call8, %forcoll.refetch ]
- %tmp9 = icmp ugt i64 %forcoll.count.ph, 1
- %umax = select i1 %tmp9, i64 %forcoll.count.ph, i64 1
- br label %forcoll.loopbody
-
-forcoll.loopbody:
- %forcoll.index = phi i64 [ 0, %forcoll.loopbody.outer ], [ %4, %forcoll.notmutated ]
- %mutationsptr5 = load i64*, i64** %mutationsptr.ptr, align 8
- %statemutations = load i64, i64* %mutationsptr5, align 8
- %2 = icmp eq i64 %statemutations, %forcoll.initial-mutations
- br i1 %2, label %forcoll.notmutated, label %forcoll.mutated
-
-forcoll.mutated:
- call void @llvm.objc.enumerationMutation(i8* %1)
- br label %forcoll.notmutated
-
-forcoll.notmutated:
- %stateitems = load i8**, i8*** %stateitems.ptr, align 8
- %currentitem.ptr = getelementptr i8*, i8** %stateitems, i64 %forcoll.index
- %3 = load i8*, i8** %currentitem.ptr, align 8
- call void @use(i8* %3)
- %4 = add i64 %forcoll.index, 1
- %exitcond = icmp eq i64 %4, %umax
- br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
-
-forcoll.refetch:
- %tmp7 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call8 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp7, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %5 = icmp eq i64 %call8, 0
- br i1 %5, label %forcoll.empty, label %forcoll.loopbody.outer
-
-forcoll.empty:
- call void @llvm.objc.release(i8* %1) nounwind
- call void @llvm.objc.release(i8* %0) nounwind, !clang.imprecise_release !0
- ret void
-}
-
-; Delete a nested retain+release pair.
-
-; CHECK-LABEL: define void @test5(
-; CHECK: call i8* @llvm.objc.retainAutoreleasedReturnValue
-; CHECK-NOT: @llvm.objc.retain
-; CHECK: }
-define void @test5() nounwind {
-entry:
- %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
- %items.ptr = alloca [16 x i8*], align 8
- %call = call i8* @returner()
- %0 = call i8* @llvm.objc.retainAutoreleasedReturnValue(i8* %call) nounwind
- %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
- call void @llvm.memset.p0i8.i64(i8* align 8 %tmp, i8 0, i64 64, i1 false)
- %1 = call i8* @llvm.objc.retain(i8* %0) nounwind
- %tmp2 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call3 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp2, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %iszero = icmp eq i64 %call3, 0
- br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
-
-forcoll.loopinit:
- %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
- %mutationsptr = load i64*, i64** %mutationsptr.ptr, align 8
- %forcoll.initial-mutations = load i64, i64* %mutationsptr, align 8
- %stateitems.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 1
- br label %forcoll.loopbody.outer
-
-forcoll.loopbody.outer:
- %forcoll.count.ph = phi i64 [ %call3, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
- %tmp8 = icmp ugt i64 %forcoll.count.ph, 1
- %umax = select i1 %tmp8, i64 %forcoll.count.ph, i64 1
- br label %forcoll.loopbody
-
-forcoll.loopbody:
- %forcoll.index = phi i64 [ 0, %forcoll.loopbody.outer ], [ %4, %forcoll.notmutated ]
- %mutationsptr4 = load i64*, i64** %mutationsptr.ptr, align 8
- %statemutations = load i64, i64* %mutationsptr4, align 8
- %2 = icmp eq i64 %statemutations, %forcoll.initial-mutations
- br i1 %2, label %forcoll.notmutated, label %forcoll.mutated
-
-forcoll.mutated:
- call void @llvm.objc.enumerationMutation(i8* %1)
- br label %forcoll.notmutated
-
-forcoll.notmutated:
- %stateitems = load i8**, i8*** %stateitems.ptr, align 8
- %currentitem.ptr = getelementptr i8*, i8** %stateitems, i64 %forcoll.index
- %3 = load i8*, i8** %currentitem.ptr, align 8
- call void @use(i8* %3)
- %4 = add i64 %forcoll.index, 1
- %exitcond = icmp eq i64 %4, %umax
- br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
-
-forcoll.refetch:
- %tmp6 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %5 = icmp eq i64 %call7, 0
- br i1 %5, label %forcoll.empty, label %forcoll.loopbody.outer
-
-forcoll.empty:
- call void @llvm.objc.release(i8* %1) nounwind
- call void @llvm.objc.release(i8* %0) nounwind, !clang.imprecise_release !0
- ret void
-}
-
-; We handle this now due to the fact that a release just needs a post dominating
-; use.
-;
-; CHECK-LABEL: define void @test6(
-; CHECK: call i8* @llvm.objc.retainAutoreleasedReturnValue
-; CHECK-NOT: @llvm.objc.retain
-; CHECK: }
-define void @test6() nounwind {
-entry:
- %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
- %items.ptr = alloca [16 x i8*], align 8
- %call = call i8* @returner()
- %0 = call i8* @llvm.objc.retainAutoreleasedReturnValue(i8* %call) nounwind
- %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
- call void @llvm.memset.p0i8.i64(i8* align 8 %tmp, i8 0, i64 64, i1 false)
- %1 = call i8* @llvm.objc.retain(i8* %0) nounwind
- %tmp2 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call3 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp2, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %iszero = icmp eq i64 %call3, 0
- br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
-
-forcoll.loopinit:
- %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
- %mutationsptr = load i64*, i64** %mutationsptr.ptr, align 8
- %forcoll.initial-mutations = load i64, i64* %mutationsptr, align 8
- %stateitems.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 1
- br label %forcoll.loopbody.outer
-
-forcoll.loopbody.outer:
- %forcoll.count.ph = phi i64 [ %call3, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
- %tmp8 = icmp ugt i64 %forcoll.count.ph, 1
- %umax = select i1 %tmp8, i64 %forcoll.count.ph, i64 1
- br label %forcoll.loopbody
-
-forcoll.loopbody:
- %forcoll.index = phi i64 [ 0, %forcoll.loopbody.outer ], [ %4, %forcoll.notmutated ]
- %mutationsptr4 = load i64*, i64** %mutationsptr.ptr, align 8
- %statemutations = load i64, i64* %mutationsptr4, align 8
- %2 = icmp eq i64 %statemutations, %forcoll.initial-mutations
- br i1 %2, label %forcoll.notmutated, label %forcoll.mutated
-
-forcoll.mutated:
- call void @llvm.objc.enumerationMutation(i8* %1)
- br label %forcoll.notmutated
-
-forcoll.notmutated:
- %stateitems = load i8**, i8*** %stateitems.ptr, align 8
- %currentitem.ptr = getelementptr i8*, i8** %stateitems, i64 %forcoll.index
- %3 = load i8*, i8** %currentitem.ptr, align 8
- call void @use(i8* %3)
- %4 = add i64 %forcoll.index, 1
- %exitcond = icmp eq i64 %4, %umax
- br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
-
-forcoll.refetch:
- %tmp6 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %5 = icmp eq i64 %call7, 0
- br i1 %5, label %forcoll.empty, label %forcoll.loopbody.outer
-
-forcoll.empty:
- call void @llvm.objc.release(i8* %1) nounwind
- call void @callee()
- call void @llvm.objc.release(i8* %0) nounwind, !clang.imprecise_release !0
- ret void
-}
-
-; TODO: Delete a nested retain+release pair.
-; The optimizer currently can't do this, because isn't isn't sophisticated enough in
-; reasnoning about nesting.
-
-; CHECK-LABEL: define void @test7(
-; CHECK: call i8* @llvm.objc.retainAutoreleasedReturnValue
-; CHECK: @llvm.objc.retain
-; CHECK: }
-define void @test7() nounwind {
-entry:
- %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
- %items.ptr = alloca [16 x i8*], align 8
- %call = call i8* @returner()
- %0 = call i8* @llvm.objc.retainAutoreleasedReturnValue(i8* %call) nounwind
- call void @callee()
- %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
- call void @llvm.memset.p0i8.i64(i8* align 8 %tmp, i8 0, i64 64, i1 false)
- %1 = call i8* @llvm.objc.retain(i8* %0) nounwind
- %tmp2 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call3 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp2, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %iszero = icmp eq i64 %call3, 0
- br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
-
-forcoll.loopinit:
- %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
- %mutationsptr = load i64*, i64** %mutationsptr.ptr, align 8
- %forcoll.initial-mutations = load i64, i64* %mutationsptr, align 8
- %stateitems.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 1
- br label %forcoll.loopbody.outer
-
-forcoll.loopbody.outer:
- %forcoll.count.ph = phi i64 [ %call3, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
- %tmp8 = icmp ugt i64 %forcoll.count.ph, 1
- %umax = select i1 %tmp8, i64 %forcoll.count.ph, i64 1
- br label %forcoll.loopbody
-
-forcoll.loopbody:
- %forcoll.index = phi i64 [ 0, %forcoll.loopbody.outer ], [ %4, %forcoll.notmutated ]
- %mutationsptr4 = load i64*, i64** %mutationsptr.ptr, align 8
- %statemutations = load i64, i64* %mutationsptr4, align 8
- %2 = icmp eq i64 %statemutations, %forcoll.initial-mutations
- br i1 %2, label %forcoll.notmutated, label %forcoll.mutated
-
-forcoll.mutated:
- call void @llvm.objc.enumerationMutation(i8* %1)
- br label %forcoll.notmutated
-
-forcoll.notmutated:
- %stateitems = load i8**, i8*** %stateitems.ptr, align 8
- %currentitem.ptr = getelementptr i8*, i8** %stateitems, i64 %forcoll.index
- %3 = load i8*, i8** %currentitem.ptr, align 8
- call void @use(i8* %3)
- %4 = add i64 %forcoll.index, 1
- %exitcond = icmp eq i64 %4, %umax
- br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
-
-forcoll.refetch:
- %tmp6 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %5 = icmp eq i64 %call7, 0
- br i1 %5, label %forcoll.empty, label %forcoll.loopbody.outer
-
-forcoll.empty:
- call void @llvm.objc.release(i8* %1) nounwind
- call void @callee()
- call void @llvm.objc.release(i8* %0) nounwind, !clang.imprecise_release !0
- ret void
-}
-
-; Delete a nested retain+release pair.
-
-; CHECK-LABEL: define void @test8(
-; CHECK: call i8* @llvm.objc.retainAutoreleasedReturnValue
-; CHECK-NOT: @llvm.objc.retain
-; CHECK: }
-define void @test8() nounwind {
-entry:
- %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
- %items.ptr = alloca [16 x i8*], align 8
- %call = call i8* @returner()
- %0 = call i8* @llvm.objc.retainAutoreleasedReturnValue(i8* %call) nounwind
- %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
- call void @llvm.memset.p0i8.i64(i8* align 8 %tmp, i8 0, i64 64, i1 false)
- %1 = call i8* @llvm.objc.retain(i8* %0) nounwind
- %tmp2 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call3 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp2, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %iszero = icmp eq i64 %call3, 0
- br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
-
-forcoll.loopinit:
- %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
- %mutationsptr = load i64*, i64** %mutationsptr.ptr, align 8
- %forcoll.initial-mutations = load i64, i64* %mutationsptr, align 8
- %stateitems.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 1
- br label %forcoll.loopbody.outer
-
-forcoll.loopbody.outer:
- %forcoll.count.ph = phi i64 [ %call3, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
- %tmp8 = icmp ugt i64 %forcoll.count.ph, 1
- %umax = select i1 %tmp8, i64 %forcoll.count.ph, i64 1
- br label %forcoll.loopbody
-
-forcoll.loopbody:
- %forcoll.index = phi i64 [ 0, %forcoll.loopbody.outer ], [ %4, %forcoll.next ]
- %mutationsptr4 = load i64*, i64** %mutationsptr.ptr, align 8
- %statemutations = load i64, i64* %mutationsptr4, align 8
- %2 = icmp eq i64 %statemutations, %forcoll.initial-mutations
- br i1 %2, label %forcoll.notmutated, label %forcoll.mutated
-
-forcoll.mutated:
- call void @llvm.objc.enumerationMutation(i8* %1)
- br label %forcoll.notmutated
-
-forcoll.notmutated:
- %stateitems = load i8**, i8*** %stateitems.ptr, align 8
- %currentitem.ptr = getelementptr i8*, i8** %stateitems, i64 %forcoll.index
- %3 = load i8*, i8** %currentitem.ptr, align 8
- %tobool = icmp eq i8* %3, null
- br i1 %tobool, label %forcoll.next, label %if.then
-
-if.then:
- call void @callee()
- br label %forcoll.next
-
-forcoll.next:
- %4 = add i64 %forcoll.index, 1
- %exitcond = icmp eq i64 %4, %umax
- br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
-
-forcoll.refetch:
- %tmp6 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %1, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %5 = icmp eq i64 %call7, 0
- br i1 %5, label %forcoll.empty, label %forcoll.loopbody.outer
-
-forcoll.empty:
- call void @llvm.objc.release(i8* %1) nounwind
- call void @llvm.objc.release(i8* %0) nounwind, !clang.imprecise_release !0
- ret void
-}
-
-; TODO: Delete a nested retain+release pair.
-; The optimizer currently can't do this, because of a split loop backedge.
-; See test9b for the same testcase without a split backedge.
-
-; CHECK-LABEL: define void @test9(
-; CHECK: call i8* @llvm.objc.retainAutoreleasedReturnValue
-; CHECK: call i8* @llvm.objc.retainAutoreleasedReturnValue
-; CHECK: call i8* @llvm.objc.retain
-; CHECK: }
-define void @test9() nounwind {
-entry:
- %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
- %items.ptr = alloca [16 x i8*], align 8
- %call = call i8* @returner()
- %0 = call i8* @llvm.objc.retainAutoreleasedReturnValue(i8* %call) nounwind
- %call1 = call i8* @returner()
- %1 = call i8* @llvm.objc.retainAutoreleasedReturnValue(i8* %call1) nounwind
- %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
- call void @llvm.memset.p0i8.i64(i8* align 8 %tmp, i8 0, i64 64, i1 false)
- %2 = call i8* @llvm.objc.retain(i8* %0) nounwind
- %tmp3 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call4 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %2, i8* %tmp3, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %iszero = icmp eq i64 %call4, 0
- br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
-
-forcoll.loopinit:
- %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
- %mutationsptr = load i64*, i64** %mutationsptr.ptr, align 8
- %forcoll.initial-mutations = load i64, i64* %mutationsptr, align 8
- br label %forcoll.loopbody.outer
-
-forcoll.loopbody.outer:
- %forcoll.count.ph = phi i64 [ %call4, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
- %tmp9 = icmp ugt i64 %forcoll.count.ph, 1
- %umax = select i1 %tmp9, i64 %forcoll.count.ph, i64 1
- br label %forcoll.loopbody
-
-forcoll.loopbody:
- %forcoll.index = phi i64 [ %phitmp, %forcoll.notmutated.forcoll.loopbody_crit_edge ], [ 1, %forcoll.loopbody.outer ]
- %mutationsptr5 = load i64*, i64** %mutationsptr.ptr, align 8
- %statemutations = load i64, i64* %mutationsptr5, align 8
- %3 = icmp eq i64 %statemutations, %forcoll.initial-mutations
- br i1 %3, label %forcoll.notmutated, label %forcoll.mutated
-
-forcoll.mutated:
- call void @llvm.objc.enumerationMutation(i8* %2)
- br label %forcoll.notmutated
-
-forcoll.notmutated:
- %exitcond = icmp eq i64 %forcoll.index, %umax
- br i1 %exitcond, label %forcoll.refetch, label %forcoll.notmutated.forcoll.loopbody_crit_edge
-
-forcoll.notmutated.forcoll.loopbody_crit_edge:
- %phitmp = add i64 %forcoll.index, 1
- br label %forcoll.loopbody
-
-forcoll.refetch:
- %tmp6 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %2, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %4 = icmp eq i64 %call7, 0
- br i1 %4, label %forcoll.empty, label %forcoll.loopbody.outer
-
-forcoll.empty:
- call void @llvm.objc.release(i8* %2) nounwind
- call void @llvm.objc.release(i8* %1) nounwind, !clang.imprecise_release !0
- call void @llvm.objc.release(i8* %0) nounwind, !clang.imprecise_release !0
- ret void
-}
-
-; Like test9, but without a split backedge. TODO: optimize this.
-
-; CHECK-LABEL: define void @test9b(
-; CHECK: call i8* @llvm.objc.retainAutoreleasedReturnValue
-; CHECK: call i8* @llvm.objc.retainAutoreleasedReturnValue
-; CHECK: @llvm.objc.retain
-; CHECK: }
-define void @test9b() nounwind {
-entry:
- %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
- %items.ptr = alloca [16 x i8*], align 8
- %call = call i8* @returner()
- %0 = call i8* @llvm.objc.retainAutoreleasedReturnValue(i8* %call) nounwind
- %call1 = call i8* @returner()
- %1 = call i8* @llvm.objc.retainAutoreleasedReturnValue(i8* %call1) nounwind
- %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
- call void @llvm.memset.p0i8.i64(i8* align 8 %tmp, i8 0, i64 64, i1 false)
- %2 = call i8* @llvm.objc.retain(i8* %0) nounwind
- %tmp3 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call4 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %2, i8* %tmp3, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %iszero = icmp eq i64 %call4, 0
- br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
-
-forcoll.loopinit:
- %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
- %mutationsptr = load i64*, i64** %mutationsptr.ptr, align 8
- %forcoll.initial-mutations = load i64, i64* %mutationsptr, align 8
- br label %forcoll.loopbody.outer
-
-forcoll.loopbody.outer:
- %forcoll.count.ph = phi i64 [ %call4, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
- %tmp9 = icmp ugt i64 %forcoll.count.ph, 1
- %umax = select i1 %tmp9, i64 %forcoll.count.ph, i64 1
- br label %forcoll.loopbody
-
-forcoll.loopbody:
- %forcoll.index = phi i64 [ %phitmp, %forcoll.notmutated ], [ 0, %forcoll.loopbody.outer ]
- %mutationsptr5 = load i64*, i64** %mutationsptr.ptr, align 8
- %statemutations = load i64, i64* %mutationsptr5, align 8
- %3 = icmp eq i64 %statemutations, %forcoll.initial-mutations
- br i1 %3, label %forcoll.notmutated, label %forcoll.mutated
-
-forcoll.mutated:
- call void @llvm.objc.enumerationMutation(i8* %2)
- br label %forcoll.notmutated
-
-forcoll.notmutated:
- %phitmp = add i64 %forcoll.index, 1
- %exitcond = icmp eq i64 %phitmp, %umax
- br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
-
-forcoll.refetch:
- %tmp6 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %2, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %4 = icmp eq i64 %call7, 0
- br i1 %4, label %forcoll.empty, label %forcoll.loopbody.outer
-
-forcoll.empty:
- call void @llvm.objc.release(i8* %2) nounwind
- call void @llvm.objc.release(i8* %1) nounwind, !clang.imprecise_release !0
- call void @llvm.objc.release(i8* %0) nounwind, !clang.imprecise_release !0
- ret void
-}
-
-; TODO: Delete a nested retain+release pair.
-; The optimizer currently can't do this, because of a split loop backedge.
-; See test10b for the same testcase without a split backedge.
-
-; CHECK-LABEL: define void @test10(
-; CHECK: call i8* @llvm.objc.retainAutoreleasedReturnValue
-; CHECK: call i8* @llvm.objc.retainAutoreleasedReturnValue
-; CHECK: call i8* @llvm.objc.retain
-; CHECK: }
-define void @test10() nounwind {
-entry:
- %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
- %items.ptr = alloca [16 x i8*], align 8
- %call = call i8* @returner()
- %0 = call i8* @llvm.objc.retainAutoreleasedReturnValue(i8* %call) nounwind
- %call1 = call i8* @returner()
- %1 = call i8* @llvm.objc.retainAutoreleasedReturnValue(i8* %call1) nounwind
- call void @callee()
- %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
- call void @llvm.memset.p0i8.i64(i8* align 8 %tmp, i8 0, i64 64, i1 false)
- %2 = call i8* @llvm.objc.retain(i8* %0) nounwind
- %tmp3 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call4 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %2, i8* %tmp3, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %iszero = icmp eq i64 %call4, 0
- br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
-
-forcoll.loopinit:
- %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
- %mutationsptr = load i64*, i64** %mutationsptr.ptr, align 8
- %forcoll.initial-mutations = load i64, i64* %mutationsptr, align 8
- br label %forcoll.loopbody.outer
-
-forcoll.loopbody.outer:
- %forcoll.count.ph = phi i64 [ %call4, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
- %tmp9 = icmp ugt i64 %forcoll.count.ph, 1
- %umax = select i1 %tmp9, i64 %forcoll.count.ph, i64 1
- br label %forcoll.loopbody
-
-forcoll.loopbody:
- %forcoll.index = phi i64 [ %phitmp, %forcoll.notmutated.forcoll.loopbody_crit_edge ], [ 1, %forcoll.loopbody.outer ]
- %mutationsptr5 = load i64*, i64** %mutationsptr.ptr, align 8
- %statemutations = load i64, i64* %mutationsptr5, align 8
- %3 = icmp eq i64 %statemutations, %forcoll.initial-mutations
- br i1 %3, label %forcoll.notmutated, label %forcoll.mutated
-
-forcoll.mutated:
- call void @llvm.objc.enumerationMutation(i8* %2)
- br label %forcoll.notmutated
-
-forcoll.notmutated:
- %exitcond = icmp eq i64 %forcoll.index, %umax
- br i1 %exitcond, label %forcoll.refetch, label %forcoll.notmutated.forcoll.loopbody_crit_edge
-
-forcoll.notmutated.forcoll.loopbody_crit_edge:
- %phitmp = add i64 %forcoll.index, 1
- br label %forcoll.loopbody
-
-forcoll.refetch:
- %tmp6 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %2, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %4 = icmp eq i64 %call7, 0
- br i1 %4, label %forcoll.empty, label %forcoll.loopbody.outer
-
-forcoll.empty:
- call void @llvm.objc.release(i8* %2) nounwind
- call void @llvm.objc.release(i8* %1) nounwind, !clang.imprecise_release !0
- call void @llvm.objc.release(i8* %0) nounwind, !clang.imprecise_release !0
- ret void
-}
-
-; Like test10, but without a split backedge. TODO: optimize this.
-
-; CHECK-LABEL: define void @test10b(
-; CHECK: call i8* @llvm.objc.retainAutoreleasedReturnValue
-; CHECK: call i8* @llvm.objc.retainAutoreleasedReturnValue
-; CHECK: @llvm.objc.retain
-; CHECK: }
-define void @test10b() nounwind {
-entry:
- %state.ptr = alloca %struct.__objcFastEnumerationState, align 8
- %items.ptr = alloca [16 x i8*], align 8
- %call = call i8* @returner()
- %0 = call i8* @llvm.objc.retainAutoreleasedReturnValue(i8* %call) nounwind
- %call1 = call i8* @returner()
- %1 = call i8* @llvm.objc.retainAutoreleasedReturnValue(i8* %call1) nounwind
- call void @callee()
- %tmp = bitcast %struct.__objcFastEnumerationState* %state.ptr to i8*
- call void @llvm.memset.p0i8.i64(i8* align 8 %tmp, i8 0, i64 64, i1 false)
- %2 = call i8* @llvm.objc.retain(i8* %0) nounwind
- %tmp3 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call4 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %2, i8* %tmp3, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %iszero = icmp eq i64 %call4, 0
- br i1 %iszero, label %forcoll.empty, label %forcoll.loopinit
-
-forcoll.loopinit:
- %mutationsptr.ptr = getelementptr inbounds %struct.__objcFastEnumerationState, %struct.__objcFastEnumerationState* %state.ptr, i64 0, i32 2
- %mutationsptr = load i64*, i64** %mutationsptr.ptr, align 8
- %forcoll.initial-mutations = load i64, i64* %mutationsptr, align 8
- br label %forcoll.loopbody.outer
-
-forcoll.loopbody.outer:
- %forcoll.count.ph = phi i64 [ %call4, %forcoll.loopinit ], [ %call7, %forcoll.refetch ]
- %tmp9 = icmp ugt i64 %forcoll.count.ph, 1
- %umax = select i1 %tmp9, i64 %forcoll.count.ph, i64 1
- br label %forcoll.loopbody
-
-forcoll.loopbody:
- %forcoll.index = phi i64 [ %phitmp, %forcoll.notmutated ], [ 0, %forcoll.loopbody.outer ]
- %mutationsptr5 = load i64*, i64** %mutationsptr.ptr, align 8
- %statemutations = load i64, i64* %mutationsptr5, align 8
- %3 = icmp eq i64 %statemutations, %forcoll.initial-mutations
- br i1 %3, label %forcoll.notmutated, label %forcoll.mutated
-
-forcoll.mutated:
- call void @llvm.objc.enumerationMutation(i8* %2)
- br label %forcoll.notmutated
-
-forcoll.notmutated:
- %phitmp = add i64 %forcoll.index, 1
- %exitcond = icmp eq i64 %phitmp, %umax
- br i1 %exitcond, label %forcoll.refetch, label %forcoll.loopbody
-
-forcoll.refetch:
- %tmp6 = load i8*, i8** @"\01L_OBJC_SELECTOR_REFERENCES_", align 8
- %call7 = call i64 bitcast (i8* (i8*, i8*, ...)* @llvm.objc.msgSend to i64 (i8*, i8*, %struct.__objcFastEnumerationState*, [16 x i8*]*, i64)*)(i8* %2, i8* %tmp6, %struct.__objcFastEnumerationState* %state.ptr, [16 x i8*]* %items.ptr, i64 16)
- %4 = icmp eq i64 %call7, 0
- br i1 %4, label %forcoll.empty, label %forcoll.loopbody.outer
-
-forcoll.empty:
- call void @llvm.objc.release(i8* %2) nounwind
- call void @llvm.objc.release(i8* %1) nounwind, !clang.imprecise_release !0
- call void @llvm.objc.release(i8* %0) nounwind, !clang.imprecise_release !0
- ret void
-}
-
-; Pointers to strong pointers can obscure provenance relationships. Be conservative
-; in the face of escaping pointers. rdar://12150909.
-
-%struct.__block_d = type { i64, i64 }
-
-@_NSConcreteStackBlock = external global i8*
-@__block_d_tmp = external hidden constant { i64, i64, i8*, i8*, i8*, i8* }
-@__block_d_tmp5 = external hidden constant { i64, i64, i8*, i8*, i8*, i8* }
-
-; CHECK-LABEL: define void @test11(
-; CHECK: tail call i8* @llvm.objc.retain(i8* %call) [[NUW:#[0-9]+]]
-; CHECK: tail call i8* @llvm.objc.retain(i8* %call) [[NUW]]
-; CHECK: call void @llvm.objc.release(i8* %call) [[NUW]], !clang.imprecise_release !0
-; CHECK: }
-define void @test11() {
-entry:
- %block = alloca <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>, align 8
- %block9 = alloca <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>, align 8
- %call = call i8* @def(), !clang.arc.no_objc_arc_exceptions !0
- %foo = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>, <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block, i64 0, i32 5
- %block.isa = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>, <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block, i64 0, i32 0
- store i8* bitcast (i8** @_NSConcreteStackBlock to i8*), i8** %block.isa, align 8
- %block.flags = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>, <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block, i64 0, i32 1
- store i32 1107296256, i32* %block.flags, align 8
- %block.reserved = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>, <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block, i64 0, i32 2
- store i32 0, i32* %block.reserved, align 4
- %block.invoke = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>, <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block, i64 0, i32 3
- store i8* bitcast (void (i8*)* @__crasher_block_invoke to i8*), i8** %block.invoke, align 8
- %block.d = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>, <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block, i64 0, i32 4
- store %struct.__block_d* bitcast ({ i64, i64, i8*, i8*, i8*, i8* }* @__block_d_tmp to %struct.__block_d*), %struct.__block_d** %block.d, align 8
- %foo2 = tail call i8* @llvm.objc.retain(i8* %call) nounwind
- store i8* %foo2, i8** %foo, align 8
- %foo4 = bitcast <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block to i8*
- %foo5 = call i8* @llvm.objc.retainBlock(i8* %foo4) nounwind
- call void @use(i8* %foo5), !clang.arc.no_objc_arc_exceptions !0
- call void @llvm.objc.release(i8* %foo5) nounwind
- %strongdestroy = load i8*, i8** %foo, align 8
- call void @llvm.objc.release(i8* %strongdestroy) nounwind, !clang.imprecise_release !0
- %foo10 = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>, <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block9, i64 0, i32 5
- %block.isa11 = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>, <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block9, i64 0, i32 0
- store i8* bitcast (i8** @_NSConcreteStackBlock to i8*), i8** %block.isa11, align 8
- %block.flags12 = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>, <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block9, i64 0, i32 1
- store i32 1107296256, i32* %block.flags12, align 8
- %block.reserved13 = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>, <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block9, i64 0, i32 2
- store i32 0, i32* %block.reserved13, align 4
- %block.invoke14 = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>, <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block9, i64 0, i32 3
- store i8* bitcast (void (i8*)* @__crasher_block_invoke1 to i8*), i8** %block.invoke14, align 8
- %block.d15 = getelementptr inbounds <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>, <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block9, i64 0, i32 4
- store %struct.__block_d* bitcast ({ i64, i64, i8*, i8*, i8*, i8* }* @__block_d_tmp5 to %struct.__block_d*), %struct.__block_d** %block.d15, align 8
- %foo18 = call i8* @llvm.objc.retain(i8* %call) nounwind
- store i8* %call, i8** %foo10, align 8
- %foo20 = bitcast <{ i8*, i32, i32, i8*, %struct.__block_d*, i8* }>* %block9 to i8*
- %foo21 = call i8* @llvm.objc.retainBlock(i8* %foo20) nounwind
- call void @use(i8* %foo21), !clang.arc.no_objc_arc_exceptions !0
- call void @llvm.objc.release(i8* %foo21) nounwind
- %strongdestroy25 = load i8*, i8** %foo10, align 8
- call void @llvm.objc.release(i8* %strongdestroy25) nounwind, !clang.imprecise_release !0
- call void @llvm.objc.release(i8* %call) nounwind, !clang.imprecise_release !0
- ret void
-}
-
-
-; CHECK: attributes [[NUW]] = { nounwind }
-; CHECK: attributes #1 = { argmemonly nounwind }
-; CHECK: attributes #2 = { nonlazybind }
OpenPOWER on IntegriCloud