diff options
Diffstat (limited to 'llvm/test/CodeGen')
| -rw-r--r-- | llvm/test/CodeGen/Hexagon/avoid-predspill-calleesaved.ll | 49 | ||||
| -rw-r--r-- | llvm/test/CodeGen/Hexagon/avoid-predspill.ll | 3 | ||||
| -rw-r--r-- | llvm/test/CodeGen/Hexagon/eliminate-pred-spill.ll | 144 | ||||
| -rw-r--r-- | llvm/test/CodeGen/Hexagon/reg-scavengebug-3.ll | 80 | ||||
| -rw-r--r-- | llvm/test/CodeGen/Hexagon/vec-pred-spill1.ll | 80 |
5 files changed, 353 insertions, 3 deletions
diff --git a/llvm/test/CodeGen/Hexagon/avoid-predspill-calleesaved.ll b/llvm/test/CodeGen/Hexagon/avoid-predspill-calleesaved.ll new file mode 100644 index 00000000000..561013b174d --- /dev/null +++ b/llvm/test/CodeGen/Hexagon/avoid-predspill-calleesaved.ll @@ -0,0 +1,49 @@ +; Check that a callee-saved register will be saved correctly if +; the predicate-to-GPR spilling code uses it. +; +; RUN: llc -march=hexagon < %s | FileCheck %s +; +; We expect to spill p0 into a general-purpose register and keep it there, +; without adding an extra spill of that register. +; +; CHECK: PredSpill: +; CHECK: memd(r29{{.*}}) = r17:16 +; CHECK-DAG: r{{[0-9]+}} = p0 +; CHECK-DAG: p0 = r{{[0-9]+}} +; CHECK-NOT: = memw(r29 +; + +define void @PredSpill() { +entry: + br i1 undef, label %if.then, label %if.else.14 + +if.then: ; preds = %entry + br i1 undef, label %if.end.57, label %if.else + +if.else: ; preds = %if.then + unreachable + +if.else.14: ; preds = %entry + br i1 undef, label %if.then.17, label %if.end.57 + +if.then.17: ; preds = %if.else.14 + br i1 undef, label %if.end.57, label %if.then.20 + +if.then.20: ; preds = %if.then.17 + %call21 = tail call i32 @myfun() + %tobool22 = icmp eq i32 %call21, 0 + %0 = tail call i32 @myfun() + br i1 %tobool22, label %if.else.42, label %if.then.23 + +if.then.23: ; preds = %if.then.20 + unreachable + +if.else.42: ; preds = %if.then.20 + ret void + +if.end.57: ; preds = %if.then.17, %if.else.14, %if.then + ret void +} + +declare i32 @myfun() + diff --git a/llvm/test/CodeGen/Hexagon/avoid-predspill.ll b/llvm/test/CodeGen/Hexagon/avoid-predspill.ll index 883c16a51a3..159c149c442 100644 --- a/llvm/test/CodeGen/Hexagon/avoid-predspill.ll +++ b/llvm/test/CodeGen/Hexagon/avoid-predspill.ll @@ -1,6 +1,3 @@ -; This functionality will be restored shortly. -; XFAIL: * - ; RUN: llc -march=hexagon -O2 < %s | FileCheck %s ; ; This checks that predicate registers are moved to GPRs instead of spilling diff --git a/llvm/test/CodeGen/Hexagon/eliminate-pred-spill.ll b/llvm/test/CodeGen/Hexagon/eliminate-pred-spill.ll new file mode 100644 index 00000000000..6fb0a3e2658 --- /dev/null +++ b/llvm/test/CodeGen/Hexagon/eliminate-pred-spill.ll @@ -0,0 +1,144 @@ +; RUN: llc -march=hexagon -mcpu=hexagonv60 -enable-hexagon-hvx-double \ +; RUN: -hexagon-bit=0 < %s | FileCheck %s + +; This spill should be eliminated. +; CHECK-NOT: vmem(r29+#6) + +define void @test(i8* noalias nocapture %key, i8* noalias nocapture %data1) #0 { +entry: + %0 = bitcast i8* %key to <32 x i32>* + %1 = bitcast i8* %data1 to <32 x i32>* + br label %for.body + +for.body: + %pkey.0542 = phi <32 x i32>* [ %0, %entry ], [ null, %for.body ] + %pdata0.0541 = phi <32 x i32>* [ null, %entry ], [ %add.ptr48, %for.body ] + %pdata1.0540 = phi <32 x i32>* [ %1, %entry ], [ %add.ptr49, %for.body ] + %dAccum0.0539 = phi <64 x i32> [ undef, %entry ], [ %86, %for.body ] + %2 = load <32 x i32>, <32 x i32>* %pkey.0542, align 128 + %3 = load <32 x i32>, <32 x i32>* %pdata0.0541, align 128 + %4 = load <32 x i32>, <32 x i32>* undef, align 128 + %arrayidx4 = getelementptr inbounds <32 x i32>, <32 x i32>* %pdata0.0541, i32 2 + %5 = load <32 x i32>, <32 x i32>* %arrayidx4, align 128 + %arrayidx5 = getelementptr inbounds <32 x i32>, <32 x i32>* %pdata1.0540, i32 2 + %6 = load <32 x i32>, <32 x i32>* %arrayidx5, align 128 + %7 = load <32 x i32>, <32 x i32>* null, align 128 + %8 = load <32 x i32>, <32 x i32>* undef, align 128 + %9 = load <32 x i32>, <32 x i32>* null, align 128 + %arrayidx9 = getelementptr inbounds <32 x i32>, <32 x i32>* %pkey.0542, i32 3 + %arrayidx10 = getelementptr inbounds <32 x i32>, <32 x i32>* %pdata0.0541, i32 6 + %10 = load <32 x i32>, <32 x i32>* %arrayidx10, align 128 + %arrayidx12 = getelementptr inbounds <32 x i32>, <32 x i32>* %pkey.0542, i32 4 + %11 = load <32 x i32>, <32 x i32>* %arrayidx12, align 128 + %arrayidx13 = getelementptr inbounds <32 x i32>, <32 x i32>* %pdata0.0541, i32 8 + %arrayidx14 = getelementptr inbounds <32 x i32>, <32 x i32>* %pdata1.0540, i32 8 + %12 = load <32 x i32>, <32 x i32>* %arrayidx14, align 128 + %arrayidx15 = getelementptr inbounds <32 x i32>, <32 x i32>* %pkey.0542, i32 5 + %13 = load <32 x i32>, <32 x i32>* %arrayidx15, align 128 + %arrayidx16 = getelementptr inbounds <32 x i32>, <32 x i32>* %pdata0.0541, i32 10 + %arrayidx17 = getelementptr inbounds <32 x i32>, <32 x i32>* %pdata1.0540, i32 10 + %14 = load <32 x i32>, <32 x i32>* %arrayidx17, align 128 + %arrayidx18 = getelementptr inbounds <32 x i32>, <32 x i32>* %pkey.0542, i32 6 + %15 = load <32 x i32>, <32 x i32>* %arrayidx18, align 128 + %arrayidx19 = getelementptr inbounds <32 x i32>, <32 x i32>* %pdata0.0541, i32 12 + %16 = load <32 x i32>, <32 x i32>* %arrayidx19, align 128 + %arrayidx20 = getelementptr inbounds <32 x i32>, <32 x i32>* %pdata1.0540, i32 12 + %17 = load <32 x i32>, <32 x i32>* %arrayidx20, align 128 + %arrayidx22 = getelementptr inbounds <32 x i32>, <32 x i32>* %pdata0.0541, i32 14 + %18 = load <32 x i32>, <32 x i32>* %arrayidx22, align 128 + %arrayidx23 = getelementptr inbounds <32 x i32>, <32 x i32>* %pdata1.0540, i32 14 + %19 = load <32 x i32>, <32 x i32>* %arrayidx23, align 128 + %20 = tail call <1024 x i1> @llvm.hexagon.V6.vgtb.128B(<32 x i32> %2, <32 x i32> %11) + %21 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %20, <32 x i32> %11, <32 x i32> %2) + %22 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %20, <32 x i32> %2, <32 x i32> %11) + %23 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %20, <32 x i32> undef, <32 x i32> %3) + %24 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %20, <32 x i32> %12, <32 x i32> undef) + %25 = tail call <1024 x i1> @llvm.hexagon.V6.vgtb.128B(<32 x i32> %7, <32 x i32> %15) + %26 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %25, <32 x i32> %15, <32 x i32> %7) + %27 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %25, <32 x i32> %7, <32 x i32> %15) + %28 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %25, <32 x i32> %16, <32 x i32> %8) + %29 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %25, <32 x i32> %8, <32 x i32> %16) + %30 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %25, <32 x i32> %17, <32 x i32> %9) + %31 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %25, <32 x i32> %9, <32 x i32> %17) + %32 = tail call <1024 x i1> @llvm.hexagon.V6.vgtb.128B(<32 x i32> %4, <32 x i32> %13) + %33 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %32, <32 x i32> %13, <32 x i32> %4) + %34 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %32, <32 x i32> %4, <32 x i32> %13) + %35 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %32, <32 x i32> undef, <32 x i32> %5) + %36 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %32, <32 x i32> %5, <32 x i32> undef) + %37 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %32, <32 x i32> %14, <32 x i32> %6) + %38 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %32, <32 x i32> %6, <32 x i32> %14) + %39 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> zeroinitializer, <32 x i32> zeroinitializer, <32 x i32> undef) + %40 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> zeroinitializer, <32 x i32> undef, <32 x i32> zeroinitializer) + %41 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> zeroinitializer, <32 x i32> %18, <32 x i32> %10) + %42 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> zeroinitializer, <32 x i32> %10, <32 x i32> %18) + %43 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> zeroinitializer, <32 x i32> %19, <32 x i32> undef) + %44 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> zeroinitializer, <32 x i32> undef, <32 x i32> %19) + %45 = tail call <1024 x i1> @llvm.hexagon.V6.vgtb.128B(<32 x i32> %21, <32 x i32> %26) + %46 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %45, <32 x i32> %26, <32 x i32> %21) + %47 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %45, <32 x i32> %21, <32 x i32> %26) + %48 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %45, <32 x i32> %28, <32 x i32> %23) + %49 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %45, <32 x i32> %23, <32 x i32> %28) + %50 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %45, <32 x i32> %30, <32 x i32> %24) + %51 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %45, <32 x i32> %24, <32 x i32> %30) + %52 = tail call <1024 x i1> @llvm.hexagon.V6.vgtb.128B(<32 x i32> %22, <32 x i32> %27) + %53 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %52, <32 x i32> %27, <32 x i32> %22) + %54 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %52, <32 x i32> %22, <32 x i32> %27) + %55 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %52, <32 x i32> %29, <32 x i32> undef) + %56 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %52, <32 x i32> undef, <32 x i32> %31) + %57 = tail call <1024 x i1> @llvm.hexagon.V6.vgtb.128B(<32 x i32> %33, <32 x i32> %39) + %58 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %57, <32 x i32> %39, <32 x i32> %33) + %59 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %57, <32 x i32> %33, <32 x i32> %39) + %60 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %57, <32 x i32> %41, <32 x i32> %35) + %61 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %57, <32 x i32> %43, <32 x i32> %37) + %62 = tail call <1024 x i1> @llvm.hexagon.V6.vgtb.128B(<32 x i32> %34, <32 x i32> %40) + %63 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %62, <32 x i32> %42, <32 x i32> %36) + %64 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %62, <32 x i32> %38, <32 x i32> %44) + %65 = tail call <1024 x i1> @llvm.hexagon.V6.vgtb.128B(<32 x i32> %46, <32 x i32> %58) + %66 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %65, <32 x i32> %58, <32 x i32> %46) + %67 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %65, <32 x i32> %60, <32 x i32> %48) + %68 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %65, <32 x i32> %61, <32 x i32> %50) + %69 = tail call <1024 x i1> @llvm.hexagon.V6.vgtb.128B(<32 x i32> %47, <32 x i32> %59) + %70 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %69, <32 x i32> %51, <32 x i32> zeroinitializer) + %71 = tail call <1024 x i1> @llvm.hexagon.V6.vgtb.128B(<32 x i32> %53, <32 x i32> zeroinitializer) + %72 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %71, <32 x i32> %63, <32 x i32> %55) + %73 = tail call <1024 x i1> @llvm.hexagon.V6.vgtb.128B(<32 x i32> %54, <32 x i32> undef) + %74 = tail call <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1> %73, <32 x i32> %56, <32 x i32> %64) + %75 = tail call <32 x i32> @llvm.hexagon.V6.vshuffeb.128B(<32 x i32> %68, <32 x i32> %67) + %76 = tail call <32 x i32> @llvm.hexagon.V6.vshuffeb.128B(<32 x i32> %70, <32 x i32> undef) + %77 = tail call <32 x i32> @llvm.hexagon.V6.vshuffeb.128B(<32 x i32> zeroinitializer, <32 x i32> %72) + %78 = tail call <32 x i32> @llvm.hexagon.V6.vshuffeb.128B(<32 x i32> %74, <32 x i32> zeroinitializer) + %79 = tail call <64 x i32> @llvm.hexagon.V6.vmpyuh.acc.128B(<64 x i32> %dAccum0.0539, <32 x i32> %75, i32 65537) + %80 = tail call <64 x i32> @llvm.hexagon.V6.vmpyuh.acc.128B(<64 x i32> %79, <32 x i32> zeroinitializer, i32 65537) + %81 = tail call <64 x i32> @llvm.hexagon.V6.vmpyuh.acc.128B(<64 x i32> %80, <32 x i32> zeroinitializer, i32 65537) + %82 = tail call <64 x i32> @llvm.hexagon.V6.vmpyuh.acc.128B(<64 x i32> %81, <32 x i32> %76, i32 65537) + %83 = tail call <64 x i32> @llvm.hexagon.V6.vmpyuh.acc.128B(<64 x i32> %82, <32 x i32> %77, i32 65537) + %84 = tail call <64 x i32> @llvm.hexagon.V6.vmpyuh.acc.128B(<64 x i32> %83, <32 x i32> zeroinitializer, i32 65537) + %85 = tail call <64 x i32> @llvm.hexagon.V6.vmpyuh.acc.128B(<64 x i32> %84, <32 x i32> undef, i32 65537) + %86 = tail call <64 x i32> @llvm.hexagon.V6.vmpyuh.acc.128B(<64 x i32> %85, <32 x i32> %78, i32 65537) + store <32 x i32> %66, <32 x i32>* %pkey.0542, align 128 + store <32 x i32> %75, <32 x i32>* %pdata0.0541, align 128 + store <32 x i32> zeroinitializer, <32 x i32>* %arrayidx4, align 128 + store <32 x i32> zeroinitializer, <32 x i32>* undef, align 128 + store <32 x i32> zeroinitializer, <32 x i32>* %arrayidx20, align 128 + store <32 x i32> zeroinitializer, <32 x i32>* null, align 128 + %add.ptr48 = getelementptr inbounds <32 x i32>, <32 x i32>* %pdata0.0541, i32 16 + %add.ptr49 = getelementptr inbounds <32 x i32>, <32 x i32>* %pdata1.0540, i32 16 + br i1 false, label %for.end, label %for.body + +for.end: + %87 = tail call <32 x i32> @llvm.hexagon.V6.hi.128B(<64 x i32> %86) + ret void +} + +declare <1024 x i1> @llvm.hexagon.V6.vgtb.128B(<32 x i32>, <32 x i32>) #1 + +declare <32 x i32> @llvm.hexagon.V6.vmux.128B(<1024 x i1>, <32 x i32>, <32 x i32>) #1 + +declare <32 x i32> @llvm.hexagon.V6.vshuffeb.128B(<32 x i32>, <32 x i32>) #1 + +declare <64 x i32> @llvm.hexagon.V6.vmpyuh.acc.128B(<64 x i32>, <32 x i32>, i32) #1 + +declare <32 x i32> @llvm.hexagon.V6.hi.128B(<64 x i32>) #1 + +attributes #0 = { nounwind "less-precise-fpmad"="false" "no-frame-pointer-elim"="true" "no-frame-pointer-elim-non-leaf" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" } +attributes #1 = { nounwind readnone } diff --git a/llvm/test/CodeGen/Hexagon/reg-scavengebug-3.ll b/llvm/test/CodeGen/Hexagon/reg-scavengebug-3.ll new file mode 100644 index 00000000000..db9ed55d2da --- /dev/null +++ b/llvm/test/CodeGen/Hexagon/reg-scavengebug-3.ll @@ -0,0 +1,80 @@ +; RUN: llc -O0 -march=hexagon -mcpu=hexagonv60 < %s | FileCheck %s + +; CHECK: vmem + +target triple = "hexagon" + +@vecpreds = external global [15 x <16 x i32>], align 64 +@vectors = external global [15 x <16 x i32>], align 64 +@vector_pairs = external global [15 x <32 x i32>], align 128 +@.str1 = external hidden unnamed_addr constant [20 x i8], align 1 +@.str2 = external hidden unnamed_addr constant [43 x i8], align 1 +@Q6VecPredResult = external global <16 x i32>, align 64 +@.str52 = external hidden unnamed_addr constant [57 x i8], align 1 +@.str54 = external hidden unnamed_addr constant [59 x i8], align 1 +@VectorResult = external global <16 x i32>, align 64 +@.str243 = external hidden unnamed_addr constant [60 x i8], align 1 +@.str251 = external hidden unnamed_addr constant [77 x i8], align 1 +@.str290 = external hidden unnamed_addr constant [65 x i8], align 1 +@VectorPairResult = external global <32 x i32>, align 128 + +; Function Attrs: nounwind +declare void @print_vector(i32, i8*) #0 + +; Function Attrs: nounwind +declare i32 @printf(i8*, ...) #0 + +; Function Attrs: nounwind +declare void @print_vecpred(i32, i8*) #0 + +; Function Attrs: nounwind readnone +declare <16 x i32> @llvm.hexagon.V6.vandqrt(<512 x i1>, i32) #1 + +; Function Attrs: nounwind +declare void @init_vectors() #0 + +; Function Attrs: nounwind readnone +declare <512 x i1> @llvm.hexagon.V6.vandvrt(<16 x i32>, i32) #1 + +; Function Attrs: nounwind readnone +declare <16 x i32> @llvm.hexagon.V6.lvsplatw(i32) #1 + +; Function Attrs: nounwind +declare void @init_addresses() #0 + +; Function Attrs: nounwind +declare <16 x i32> @llvm.hexagon.V6.vsubhnq(<512 x i1>, <16 x i32>, <16 x i32>) #1 + +; Function Attrs: nounwind +define i32 @main() #0 { +entry: + %0 = load <16 x i32>, <16 x i32>* getelementptr inbounds ([15 x <16 x i32>], [15 x <16 x i32>]* @vecpreds, i32 0, i32 0), align 64 + %1 = load <16 x i32>, <16 x i32>* getelementptr inbounds ([15 x <16 x i32>], [15 x <16 x i32>]* @vectors, i32 0, i32 1), align 64 + call void @print_vecpred(i32 64, i8* bitcast (<16 x i32>* @Q6VecPredResult to i8*)) + %2 = load <16 x i32>, <16 x i32>* getelementptr inbounds ([15 x <16 x i32>], [15 x <16 x i32>]* @vectors, i32 0, i32 1), align 64 + %call50 = call i32 (i8*, ...) @printf(i8* getelementptr inbounds ([57 x i8], [57 x i8]* @.str52, i32 0, i32 0)) #3 + %3 = load <16 x i32>, <16 x i32>* getelementptr inbounds ([15 x <16 x i32>], [15 x <16 x i32>]* @vectors, i32 0, i32 1), align 64 + %call52 = call i32 (i8*, ...) @printf(i8* getelementptr inbounds ([59 x i8], [59 x i8]* @.str54, i32 0, i32 0)) #3 + %4 = load <16 x i32>, <16 x i32>* getelementptr inbounds ([15 x <16 x i32>], [15 x <16 x i32>]* @vectors, i32 0, i32 1), align 64 + %call300 = call i32 (i8*, ...) @printf(i8* getelementptr inbounds ([65 x i8], [65 x i8]* @.str290, i32 0, i32 0)) #3 + %5 = load <16 x i32>, <16 x i32>* getelementptr inbounds ([15 x <16 x i32>], [15 x <16 x i32>]* @vectors, i32 0, i32 0), align 64 + %6 = load <16 x i32>, <16 x i32>* getelementptr inbounds ([15 x <16 x i32>], [15 x <16 x i32>]* @vectors, i32 0, i32 1), align 64 + %call1373 = call i32 (i8*, ...) @printf(i8* getelementptr inbounds ([20 x i8], [20 x i8]* @.str1, i32 0, i32 0), i8* getelementptr inbounds ([43 x i8], [43 x i8]* @.str2, i32 0, i32 0), i8* getelementptr inbounds ([60 x i8], [60 x i8]* @.str243, i32 0, i32 0)) #3 + %7 = call <16 x i32> @llvm.hexagon.V6.lvsplatw(i32 1) + %call1381 = call i32 (i8*, ...) @printf(i8* getelementptr inbounds ([20 x i8], [20 x i8]* @.str1, i32 0, i32 0), i8* getelementptr inbounds ([43 x i8], [43 x i8]* @.str2, i32 0, i32 0), i8* getelementptr inbounds ([77 x i8], [77 x i8]* @.str251, i32 0, i32 0)) #3 + %8 = call <16 x i32> @llvm.hexagon.V6.lvsplatw(i32 1) + %9 = call <512 x i1> @llvm.hexagon.V6.vandvrt(<16 x i32> %8, i32 16843009) + call void @print_vector(i32 64, i8* bitcast (<16 x i32>* @VectorResult to i8*)) + %10 = call <16 x i32> @llvm.hexagon.V6.lvsplatw(i32 1) + %11 = call <512 x i1> @llvm.hexagon.V6.vandvrt(<16 x i32> %10, i32 16843009) + %12 = bitcast <512 x i1> %11 to <16 x i32> + %13 = bitcast <16 x i32> %12 to <512 x i1> + %14 = call <16 x i32> @llvm.hexagon.V6.vsubhnq(<512 x i1> %13, <16 x i32> undef, <16 x i32> undef) + store <16 x i32> %14, <16 x i32>* @VectorResult, align 64 + ret i32 0 +} + +attributes #0 = { nounwind "less-precise-fpmad"="false" "no-frame-pointer-elim"="true" "no-frame-pointer-elim-non-leaf" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" } +attributes #1 = { nounwind readnone } +attributes #2 = { "less-precise-fpmad"="false" "no-frame-pointer-elim"="true" "no-frame-pointer-elim-non-leaf" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" } +attributes #3 = { nounwind } diff --git a/llvm/test/CodeGen/Hexagon/vec-pred-spill1.ll b/llvm/test/CodeGen/Hexagon/vec-pred-spill1.ll new file mode 100644 index 00000000000..d120295fa52 --- /dev/null +++ b/llvm/test/CodeGen/Hexagon/vec-pred-spill1.ll @@ -0,0 +1,80 @@ +; RUN: llc -march=hexagon -mcpu=hexagonv60 -O2 -enable-hexagon-hvx < %s | FileCheck %s + +; CHECK: vmem(r{{[0-9]+}}+#3) = v{{[0-9]+}} +; CHECK: call puts +; CHECK: call print_vecpred +; CHECK: v{{[0-9]+}}{{ *}}={{ *}}vmem(r{{[0-9]+}}+#3) + +target triple = "hexagon" + +@K = global i64 0, align 8 +@src = global i32 -1, align 4 +@Q6VecPredResult = common global <16 x i32> zeroinitializer, align 64 +@dst_addresses = common global [15 x i64] zeroinitializer, align 8 +@ptr_addresses = common global [15 x i8*] zeroinitializer, align 8 +@src_addresses = common global [15 x i8*] zeroinitializer, align 8 +@ptr = common global [32768 x i32] zeroinitializer, align 8 +@vecpreds = common global [15 x <16 x i32>] zeroinitializer, align 64 +@VectorResult = common global <16 x i32> zeroinitializer, align 64 +@vectors = common global [15 x <16 x i32>] zeroinitializer, align 64 +@VectorPairResult = common global <32 x i32> zeroinitializer, align 128 +@vector_pairs = common global [15 x <32 x i32>] zeroinitializer, align 128 +@str = private unnamed_addr constant [106 x i8] c"Q6VecPred4 : Q6_Q_vandor_QVR(Q6_Q_vand_VR(Q6_V_vsplat_R(1+1),(0x01010101)),Q6_V_vsplat_R(0+1),INT32_MIN)\00" +@str3 = private unnamed_addr constant [99 x i8] c"Q6VecPred4 : Q6_Q_vandor_QVR(Q6_Q_vand_VR(Q6_V_vsplat_R(1+1),(0x01010101)),Q6_V_vsplat_R(0+1),-1)\00" +@str4 = private unnamed_addr constant [98 x i8] c"Q6VecPred4 : Q6_Q_vandor_QVR(Q6_Q_vand_VR(Q6_V_vsplat_R(1+1),(0x01010101)),Q6_V_vsplat_R(0+1),0)\00" + +; Function Attrs: nounwind +define i32 @main() #0 { +entry: + %call = tail call i32 bitcast (i32 (...)* @init_addresses to i32 ()*)() #3 + %call1 = tail call i32 @acquire_vector_unit(i8 zeroext 0) #3 + tail call void @init_vectors() #3 + %0 = tail call <16 x i32> @llvm.hexagon.V6.lvsplatw(i32 2) + %1 = tail call <512 x i1> @llvm.hexagon.V6.vandvrt(<16 x i32> %0, i32 16843009) + %2 = tail call <16 x i32> @llvm.hexagon.V6.lvsplatw(i32 1) + %3 = tail call <512 x i1> @llvm.hexagon.V6.vandvrt.acc(<512 x i1> %1, <16 x i32> %2, i32 -2147483648) + %4 = bitcast <512 x i1> %3 to <16 x i32> + store <16 x i32> %4, <16 x i32>* @Q6VecPredResult, align 64, !tbaa !1 + %puts = tail call i32 @puts(i8* getelementptr inbounds ([106 x i8], [106 x i8]* @str, i32 0, i32 0)) + tail call void @print_vecpred(i32 512, i8* bitcast (<16 x i32>* @Q6VecPredResult to i8*)) #3 + %5 = tail call <512 x i1> @llvm.hexagon.V6.vandvrt.acc(<512 x i1> %1, <16 x i32> %2, i32 -1) + %6 = bitcast <512 x i1> %5 to <16 x i32> + store <16 x i32> %6, <16 x i32>* @Q6VecPredResult, align 64, !tbaa !1 + %puts5 = tail call i32 @puts(i8* getelementptr inbounds ([99 x i8], [99 x i8]* @str3, i32 0, i32 0)) + tail call void @print_vecpred(i32 512, i8* bitcast (<16 x i32>* @Q6VecPredResult to i8*)) #3 + %7 = tail call <512 x i1> @llvm.hexagon.V6.vandvrt.acc(<512 x i1> %1, <16 x i32> %2, i32 0) + %8 = bitcast <512 x i1> %7 to <16 x i32> + store <16 x i32> %8, <16 x i32>* @Q6VecPredResult, align 64, !tbaa !1 + %puts6 = tail call i32 @puts(i8* getelementptr inbounds ([98 x i8], [98 x i8]* @str4, i32 0, i32 0)) + tail call void @print_vecpred(i32 512, i8* bitcast (<16 x i32>* @Q6VecPredResult to i8*)) #3 + ret i32 0 +} + +declare i32 @init_addresses(...) #1 + +declare i32 @acquire_vector_unit(i8 zeroext) #1 + +declare void @init_vectors() #1 + +; Function Attrs: nounwind readnone +declare <512 x i1> @llvm.hexagon.V6.vandvrt.acc(<512 x i1>, <16 x i32>, i32) #2 + +; Function Attrs: nounwind readnone +declare <512 x i1> @llvm.hexagon.V6.vandvrt(<16 x i32>, i32) #2 + +; Function Attrs: nounwind readnone +declare <16 x i32> @llvm.hexagon.V6.lvsplatw(i32) #2 + +declare void @print_vecpred(i32, i8*) #1 + +; Function Attrs: nounwind +declare i32 @puts(i8* nocapture readonly) #3 + +attributes #0 = { nounwind "less-precise-fpmad"="false" "no-frame-pointer-elim"="true" "no-frame-pointer-elim-non-leaf" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" } +attributes #1 = { "less-precise-fpmad"="false" "no-frame-pointer-elim"="true" "no-frame-pointer-elim-non-leaf" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" } +attributes #2 = { nounwind readnone } +attributes #3 = { nounwind } + +!1 = !{!2, !2, i64 0} +!2 = !{!"omnipotent char", !3, i64 0} +!3 = !{!"Simple C/C++ TBAA"} |

