diff options
| -rw-r--r-- | llvm/test/Regression/CodeGen/PowerPC/buildvec_canonicalize.ll | 14 | 
1 files changed, 12 insertions, 2 deletions
| diff --git a/llvm/test/Regression/CodeGen/PowerPC/buildvec_canonicalize.ll b/llvm/test/Regression/CodeGen/PowerPC/buildvec_canonicalize.ll index 345697476a0..c642d42fc43 100644 --- a/llvm/test/Regression/CodeGen/PowerPC/buildvec_canonicalize.ll +++ b/llvm/test/Regression/CodeGen/PowerPC/buildvec_canonicalize.ll @@ -1,7 +1,11 @@ -; RUN: llvm-as < %s | llc -march=ppc32 -mcpu=g5 --enable-unsafe-fp-math | grep vxor | wc -l | grep 1  ; There should be exactly one vxor here. +; RUN: llvm-as < %s | llc -march=ppc32 -mcpu=g5 --enable-unsafe-fp-math | grep vxor | wc -l | grep 1 && -void %test(<4 x float>* %P1, <4 x int>* %P2, <4 x float>* %P3) { +; There should be exactly one vsplti here. +; RUN: llvm-as < %s | llc -march=ppc32 -mcpu=g5 --enable-unsafe-fp-math | grep vsplti | wc -l | grep 1 + + +void %VXOR(<4 x float>* %P1, <4 x int>* %P2, <4 x float>* %P3) {          %tmp = load <4 x float>* %P3          %tmp3 = load <4 x float>* %P1          %tmp4 = mul <4 x float> %tmp, %tmp3 @@ -11,3 +15,9 @@ void %test(<4 x float>* %P1, <4 x int>* %P2, <4 x float>* %P3) {          ret void  } +void %VSPLTI(<4 x int>* %P2, <8 x short>* %P3) { +        store <4 x int> cast (<16 x sbyte> < sbyte -1, sbyte -1, sbyte -1, sbyte -1, sbyte -1, sbyte -1, sbyte -1, sbyte -1, sbyte -1, sbyte -1, sbyte -1, sbyte -1, sbyte -1, sbyte -1, sbyte -1, sbyte -1 > to <4 x int>), <4 x int>* %P2 +        store <8 x short> < short -1, short -1, short -1, short -1, short -1, short -1, short -1, short -1 >, <8 x short>* %P3 +        ret void +} + | 

