diff options
Diffstat (limited to 'llvm/test/CodeGen/X86/complex-fastmath.ll')
-rw-r--r-- | llvm/test/CodeGen/X86/complex-fastmath.ll | 24 |
1 files changed, 12 insertions, 12 deletions
diff --git a/llvm/test/CodeGen/X86/complex-fastmath.ll b/llvm/test/CodeGen/X86/complex-fastmath.ll index d31707260a0..9c02ac6667f 100644 --- a/llvm/test/CodeGen/X86/complex-fastmath.ll +++ b/llvm/test/CodeGen/X86/complex-fastmath.ll @@ -11,7 +11,7 @@ define <2 x float> @complex_square_f32(<2 x float>) #0 { ; SSE-LABEL: complex_square_f32: -; SSE: # BB#0: +; SSE: # %bb.0: ; SSE-NEXT: movshdup {{.*#+}} xmm1 = xmm0[1,1,3,3] ; SSE-NEXT: movaps %xmm0, %xmm2 ; SSE-NEXT: addss %xmm2, %xmm2 @@ -23,7 +23,7 @@ define <2 x float> @complex_square_f32(<2 x float>) #0 { ; SSE-NEXT: retq ; ; AVX1-LABEL: complex_square_f32: -; AVX1: # BB#0: +; AVX1: # %bb.0: ; AVX1-NEXT: vmovshdup {{.*#+}} xmm1 = xmm0[1,1,3,3] ; AVX1-NEXT: vaddss %xmm0, %xmm0, %xmm2 ; AVX1-NEXT: vmulss %xmm2, %xmm1, %xmm2 @@ -34,7 +34,7 @@ define <2 x float> @complex_square_f32(<2 x float>) #0 { ; AVX1-NEXT: retq ; ; FMA-LABEL: complex_square_f32: -; FMA: # BB#0: +; FMA: # %bb.0: ; FMA-NEXT: vmovshdup {{.*#+}} xmm1 = xmm0[1,1,3,3] ; FMA-NEXT: vaddss %xmm0, %xmm0, %xmm2 ; FMA-NEXT: vmulss %xmm2, %xmm1, %xmm2 @@ -56,7 +56,7 @@ define <2 x float> @complex_square_f32(<2 x float>) #0 { define <2 x double> @complex_square_f64(<2 x double>) #0 { ; SSE-LABEL: complex_square_f64: -; SSE: # BB#0: +; SSE: # %bb.0: ; SSE-NEXT: movaps %xmm0, %xmm1 ; SSE-NEXT: movhlps {{.*#+}} xmm1 = xmm1[1,1] ; SSE-NEXT: movaps %xmm0, %xmm2 @@ -69,7 +69,7 @@ define <2 x double> @complex_square_f64(<2 x double>) #0 { ; SSE-NEXT: retq ; ; AVX1-LABEL: complex_square_f64: -; AVX1: # BB#0: +; AVX1: # %bb.0: ; AVX1-NEXT: vpermilpd {{.*#+}} xmm1 = xmm0[1,0] ; AVX1-NEXT: vaddsd %xmm0, %xmm0, %xmm2 ; AVX1-NEXT: vmulsd %xmm2, %xmm1, %xmm2 @@ -80,7 +80,7 @@ define <2 x double> @complex_square_f64(<2 x double>) #0 { ; AVX1-NEXT: retq ; ; FMA-LABEL: complex_square_f64: -; FMA: # BB#0: +; FMA: # %bb.0: ; FMA-NEXT: vpermilpd {{.*#+}} xmm1 = xmm0[1,0] ; FMA-NEXT: vaddsd %xmm0, %xmm0, %xmm2 ; FMA-NEXT: vmulsd %xmm2, %xmm1, %xmm2 @@ -106,7 +106,7 @@ define <2 x double> @complex_square_f64(<2 x double>) #0 { define <2 x float> @complex_mul_f32(<2 x float>, <2 x float>) #0 { ; SSE-LABEL: complex_mul_f32: -; SSE: # BB#0: +; SSE: # %bb.0: ; SSE-NEXT: movshdup {{.*#+}} xmm2 = xmm0[1,1,3,3] ; SSE-NEXT: movshdup {{.*#+}} xmm3 = xmm1[1,1,3,3] ; SSE-NEXT: movaps %xmm3, %xmm4 @@ -120,7 +120,7 @@ define <2 x float> @complex_mul_f32(<2 x float>, <2 x float>) #0 { ; SSE-NEXT: retq ; ; AVX1-LABEL: complex_mul_f32: -; AVX1: # BB#0: +; AVX1: # %bb.0: ; AVX1-NEXT: vmovshdup {{.*#+}} xmm2 = xmm0[1,1,3,3] ; AVX1-NEXT: vmovshdup {{.*#+}} xmm3 = xmm1[1,1,3,3] ; AVX1-NEXT: vmulss %xmm0, %xmm3, %xmm4 @@ -133,7 +133,7 @@ define <2 x float> @complex_mul_f32(<2 x float>, <2 x float>) #0 { ; AVX1-NEXT: retq ; ; FMA-LABEL: complex_mul_f32: -; FMA: # BB#0: +; FMA: # %bb.0: ; FMA-NEXT: vmovshdup {{.*#+}} xmm2 = xmm0[1,1,3,3] ; FMA-NEXT: vmovshdup {{.*#+}} xmm3 = xmm1[1,1,3,3] ; FMA-NEXT: vmulss %xmm2, %xmm1, %xmm4 @@ -159,7 +159,7 @@ define <2 x float> @complex_mul_f32(<2 x float>, <2 x float>) #0 { define <2 x double> @complex_mul_f64(<2 x double>, <2 x double>) #0 { ; SSE-LABEL: complex_mul_f64: -; SSE: # BB#0: +; SSE: # %bb.0: ; SSE-NEXT: movaps %xmm0, %xmm2 ; SSE-NEXT: movhlps {{.*#+}} xmm2 = xmm2[1,1] ; SSE-NEXT: movaps %xmm1, %xmm3 @@ -175,7 +175,7 @@ define <2 x double> @complex_mul_f64(<2 x double>, <2 x double>) #0 { ; SSE-NEXT: retq ; ; AVX1-LABEL: complex_mul_f64: -; AVX1: # BB#0: +; AVX1: # %bb.0: ; AVX1-NEXT: vpermilpd {{.*#+}} xmm2 = xmm0[1,0] ; AVX1-NEXT: vpermilpd {{.*#+}} xmm3 = xmm1[1,0] ; AVX1-NEXT: vmulsd %xmm0, %xmm3, %xmm4 @@ -188,7 +188,7 @@ define <2 x double> @complex_mul_f64(<2 x double>, <2 x double>) #0 { ; AVX1-NEXT: retq ; ; FMA-LABEL: complex_mul_f64: -; FMA: # BB#0: +; FMA: # %bb.0: ; FMA-NEXT: vpermilpd {{.*#+}} xmm2 = xmm0[1,0] ; FMA-NEXT: vpermilpd {{.*#+}} xmm3 = xmm1[1,0] ; FMA-NEXT: vmulsd %xmm2, %xmm1, %xmm4 |