diff options
Diffstat (limited to 'llvm/test/CodeGen/X86/avx-cvt-3.ll')
-rw-r--r-- | llvm/test/CodeGen/X86/avx-cvt-3.ll | 24 |
1 files changed, 12 insertions, 12 deletions
diff --git a/llvm/test/CodeGen/X86/avx-cvt-3.ll b/llvm/test/CodeGen/X86/avx-cvt-3.ll index e4f29b5d448..ac99684ab3a 100644 --- a/llvm/test/CodeGen/X86/avx-cvt-3.ll +++ b/llvm/test/CodeGen/X86/avx-cvt-3.ll @@ -6,14 +6,14 @@ define <8 x float> @sitofp_insert_zero_v8i32(<8 x i32> %a0) { ; X86-LABEL: sitofp_insert_zero_v8i32: -; X86: # BB#0: +; X86: # %bb.0: ; X86-NEXT: vxorps %xmm1, %xmm1, %xmm1 ; X86-NEXT: vblendps {{.*#+}} ymm0 = ymm1[0],ymm0[1],ymm1[2],ymm0[3],ymm1[4,5],ymm0[6,7] ; X86-NEXT: vcvtdq2ps %ymm0, %ymm0 ; X86-NEXT: retl ; ; X64-LABEL: sitofp_insert_zero_v8i32: -; X64: # BB#0: +; X64: # %bb.0: ; X64-NEXT: vxorps %xmm1, %xmm1, %xmm1 ; X64-NEXT: vblendps {{.*#+}} ymm0 = ymm1[0],ymm0[1],ymm1[2],ymm0[3],ymm1[4,5],ymm0[6,7] ; X64-NEXT: vcvtdq2ps %ymm0, %ymm0 @@ -28,14 +28,14 @@ define <8 x float> @sitofp_insert_zero_v8i32(<8 x i32> %a0) { define <8 x float> @sitofp_shuffle_zero_v8i32(<8 x i32> %a0) { ; X86-LABEL: sitofp_shuffle_zero_v8i32: -; X86: # BB#0: +; X86: # %bb.0: ; X86-NEXT: vxorps %xmm1, %xmm1, %xmm1 ; X86-NEXT: vblendps {{.*#+}} ymm0 = ymm1[0],ymm0[1],ymm1[2],ymm0[3],ymm1[4],ymm0[5],ymm1[6],ymm0[7] ; X86-NEXT: vcvtdq2ps %ymm0, %ymm0 ; X86-NEXT: retl ; ; X64-LABEL: sitofp_shuffle_zero_v8i32: -; X64: # BB#0: +; X64: # %bb.0: ; X64-NEXT: vxorps %xmm1, %xmm1, %xmm1 ; X64-NEXT: vblendps {{.*#+}} ymm0 = ymm1[0],ymm0[1],ymm1[2],ymm0[3],ymm1[4],ymm0[5],ymm1[6],ymm0[7] ; X64-NEXT: vcvtdq2ps %ymm0, %ymm0 @@ -47,7 +47,7 @@ define <8 x float> @sitofp_shuffle_zero_v8i32(<8 x i32> %a0) { define <8 x float> @sitofp_insert_allbits_v8i32(<8 x i32> %a0) { ; X86-LABEL: sitofp_insert_allbits_v8i32: -; X86: # BB#0: +; X86: # %bb.0: ; X86-NEXT: vxorps %xmm1, %xmm1, %xmm1 ; X86-NEXT: vcmptrueps %ymm1, %ymm1, %ymm1 ; X86-NEXT: vblendps {{.*#+}} ymm0 = ymm1[0],ymm0[1],ymm1[2],ymm0[3],ymm1[4,5],ymm0[6,7] @@ -55,7 +55,7 @@ define <8 x float> @sitofp_insert_allbits_v8i32(<8 x i32> %a0) { ; X86-NEXT: retl ; ; X64-LABEL: sitofp_insert_allbits_v8i32: -; X64: # BB#0: +; X64: # %bb.0: ; X64-NEXT: vxorps %xmm1, %xmm1, %xmm1 ; X64-NEXT: vcmptrueps %ymm1, %ymm1, %ymm1 ; X64-NEXT: vblendps {{.*#+}} ymm0 = ymm1[0],ymm0[1],ymm1[2],ymm0[3],ymm1[4,5],ymm0[6,7] @@ -71,7 +71,7 @@ define <8 x float> @sitofp_insert_allbits_v8i32(<8 x i32> %a0) { define <8 x float> @sitofp_shuffle_allbits_v8i32(<8 x i32> %a0) { ; X86-LABEL: sitofp_shuffle_allbits_v8i32: -; X86: # BB#0: +; X86: # %bb.0: ; X86-NEXT: vxorps %xmm1, %xmm1, %xmm1 ; X86-NEXT: vcmptrueps %ymm1, %ymm1, %ymm1 ; X86-NEXT: vblendps {{.*#+}} ymm0 = ymm1[0],ymm0[1],ymm1[2],ymm0[3],ymm1[4],ymm0[5],ymm1[6],ymm0[7] @@ -79,7 +79,7 @@ define <8 x float> @sitofp_shuffle_allbits_v8i32(<8 x i32> %a0) { ; X86-NEXT: retl ; ; X64-LABEL: sitofp_shuffle_allbits_v8i32: -; X64: # BB#0: +; X64: # %bb.0: ; X64-NEXT: vxorps %xmm1, %xmm1, %xmm1 ; X64-NEXT: vcmptrueps %ymm1, %ymm1, %ymm1 ; X64-NEXT: vblendps {{.*#+}} ymm0 = ymm1[0],ymm0[1],ymm1[2],ymm0[3],ymm1[4],ymm0[5],ymm1[6],ymm0[7] @@ -92,7 +92,7 @@ define <8 x float> @sitofp_shuffle_allbits_v8i32(<8 x i32> %a0) { define <8 x float> @sitofp_insert_constants_v8i32(<8 x i32> %a0) { ; X86-LABEL: sitofp_insert_constants_v8i32: -; X86: # BB#0: +; X86: # %bb.0: ; X86-NEXT: vxorps %xmm1, %xmm1, %xmm1 ; X86-NEXT: vblendps {{.*#+}} ymm0 = ymm1[0],ymm0[1,2,3,4,5,6,7] ; X86-NEXT: vcmptrueps %ymm1, %ymm1, %ymm1 @@ -107,7 +107,7 @@ define <8 x float> @sitofp_insert_constants_v8i32(<8 x i32> %a0) { ; X86-NEXT: retl ; ; X64-LABEL: sitofp_insert_constants_v8i32: -; X64: # BB#0: +; X64: # %bb.0: ; X64-NEXT: vxorps %xmm1, %xmm1, %xmm1 ; X64-NEXT: vblendps {{.*#+}} ymm0 = ymm1[0],ymm0[1,2,3,4,5,6,7] ; X64-NEXT: vcmptrueps %ymm1, %ymm1, %ymm1 @@ -130,13 +130,13 @@ define <8 x float> @sitofp_insert_constants_v8i32(<8 x i32> %a0) { define <8 x float> @sitofp_shuffle_constants_v8i32(<8 x i32> %a0) { ; X86-LABEL: sitofp_shuffle_constants_v8i32: -; X86: # BB#0: +; X86: # %bb.0: ; X86-NEXT: vblendps {{.*#+}} ymm0 = mem[0],ymm0[1],mem[2],ymm0[3],mem[4],ymm0[5],mem[6],ymm0[7] ; X86-NEXT: vcvtdq2ps %ymm0, %ymm0 ; X86-NEXT: retl ; ; X64-LABEL: sitofp_shuffle_constants_v8i32: -; X64: # BB#0: +; X64: # %bb.0: ; X64-NEXT: vblendps {{.*#+}} ymm0 = mem[0],ymm0[1],mem[2],ymm0[3],mem[4],ymm0[5],mem[6],ymm0[7] ; X64-NEXT: vcvtdq2ps %ymm0, %ymm0 ; X64-NEXT: retq |