diff options
author | Sanjay Patel <spatel@rotateright.com> | 2016-04-01 18:11:30 +0000 |
---|---|---|
committer | Sanjay Patel <spatel@rotateright.com> | 2016-04-01 18:11:30 +0000 |
commit | d3e3d48cb9d430f46406403767cad4aa0f490ec4 (patch) | |
tree | 3bc339854117b9d040601d0f4acbc9985f38d088 /llvm/test/CodeGen/X86/memset-nonzero.ll | |
parent | 36404d00306d81ff21e5fa967653ea63ed7bead2 (diff) | |
download | bcm5719-llvm-d3e3d48cb9d430f46406403767cad4aa0f490ec4.tar.gz bcm5719-llvm-d3e3d48cb9d430f46406403767cad4aa0f490ec4.zip |
[x86] add an SSE1 run for these tests
Note however that this is identical to the existing SSE2 run.
What we really want is yet another run for an SSE2 machine that
also has fast unaligned 16-byte accesses.
llvm-svn: 265167
Diffstat (limited to 'llvm/test/CodeGen/X86/memset-nonzero.ll')
-rw-r--r-- | llvm/test/CodeGen/X86/memset-nonzero.ll | 211 |
1 files changed, 106 insertions, 105 deletions
diff --git a/llvm/test/CodeGen/X86/memset-nonzero.ll b/llvm/test/CodeGen/X86/memset-nonzero.ll index 61d126e2547..6fe720a4b94 100644 --- a/llvm/test/CodeGen/X86/memset-nonzero.ll +++ b/llvm/test/CodeGen/X86/memset-nonzero.ll @@ -1,16 +1,17 @@ ; NOTE: Assertions have been autogenerated by update_test_checks.py -; RUN: llc -mtriple=x86_64-unknown-unknown < %s -mattr=sse2 | FileCheck %s --check-prefix=ANY --check-prefix=SSE2 +; RUN: llc -mtriple=x86_64-unknown-unknown < %s -mattr=sse | FileCheck %s --check-prefix=ANY --check-prefix=SSE --check-prefix=SSE1 +; RUN: llc -mtriple=x86_64-unknown-unknown < %s -mattr=sse2 | FileCheck %s --check-prefix=ANY --check-prefix=SSE --check-prefix=SSE2 ; RUN: llc -mtriple=x86_64-unknown-unknown < %s -mattr=avx | FileCheck %s --check-prefix=ANY --check-prefix=AVX --check-prefix=AVX1 ; RUN: llc -mtriple=x86_64-unknown-unknown < %s -mattr=avx2 | FileCheck %s --check-prefix=ANY --check-prefix=AVX --check-prefix=AVX2 ; https://llvm.org/bugs/show_bug.cgi?id=27100 define void @memset_16_nonzero_bytes(i8* %x) { -; SSE2-LABEL: memset_16_nonzero_bytes: -; SSE2: movabsq $3038287259199220266, %rax # imm = 0x2A2A2A2A2A2A2A2A -; SSE2-NEXT: movq %rax, 8(%rdi) -; SSE2-NEXT: movq %rax, (%rdi) -; SSE2-NEXT: retq +; SSE-LABEL: memset_16_nonzero_bytes: +; SSE: movabsq $3038287259199220266, %rax # imm = 0x2A2A2A2A2A2A2A2A +; SSE-NEXT: movq %rax, 8(%rdi) +; SSE-NEXT: movq %rax, (%rdi) +; SSE-NEXT: retq ; ; AVX-LABEL: memset_16_nonzero_bytes: ; AVX: vmovaps {{.*#+}} xmm0 = [42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42] @@ -22,13 +23,13 @@ define void @memset_16_nonzero_bytes(i8* %x) { } define void @memset_32_nonzero_bytes(i8* %x) { -; SSE2-LABEL: memset_32_nonzero_bytes: -; SSE2: movabsq $3038287259199220266, %rax # imm = 0x2A2A2A2A2A2A2A2A -; SSE2-NEXT: movq %rax, 24(%rdi) -; SSE2-NEXT: movq %rax, 16(%rdi) -; SSE2-NEXT: movq %rax, 8(%rdi) -; SSE2-NEXT: movq %rax, (%rdi) -; SSE2-NEXT: retq +; SSE-LABEL: memset_32_nonzero_bytes: +; SSE: movabsq $3038287259199220266, %rax # imm = 0x2A2A2A2A2A2A2A2A +; SSE-NEXT: movq %rax, 24(%rdi) +; SSE-NEXT: movq %rax, 16(%rdi) +; SSE-NEXT: movq %rax, 8(%rdi) +; SSE-NEXT: movq %rax, (%rdi) +; SSE-NEXT: retq ; ; AVX-LABEL: memset_32_nonzero_bytes: ; AVX: vmovaps {{.*#+}} ymm0 = [42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42] @@ -41,17 +42,17 @@ define void @memset_32_nonzero_bytes(i8* %x) { } define void @memset_64_nonzero_bytes(i8* %x) { -; SSE2-LABEL: memset_64_nonzero_bytes: -; SSE2: movabsq $3038287259199220266, %rax # imm = 0x2A2A2A2A2A2A2A2A -; SSE2-NEXT: movq %rax, 56(%rdi) -; SSE2-NEXT: movq %rax, 48(%rdi) -; SSE2-NEXT: movq %rax, 40(%rdi) -; SSE2-NEXT: movq %rax, 32(%rdi) -; SSE2-NEXT: movq %rax, 24(%rdi) -; SSE2-NEXT: movq %rax, 16(%rdi) -; SSE2-NEXT: movq %rax, 8(%rdi) -; SSE2-NEXT: movq %rax, (%rdi) -; SSE2-NEXT: retq +; SSE-LABEL: memset_64_nonzero_bytes: +; SSE: movabsq $3038287259199220266, %rax # imm = 0x2A2A2A2A2A2A2A2A +; SSE-NEXT: movq %rax, 56(%rdi) +; SSE-NEXT: movq %rax, 48(%rdi) +; SSE-NEXT: movq %rax, 40(%rdi) +; SSE-NEXT: movq %rax, 32(%rdi) +; SSE-NEXT: movq %rax, 24(%rdi) +; SSE-NEXT: movq %rax, 16(%rdi) +; SSE-NEXT: movq %rax, 8(%rdi) +; SSE-NEXT: movq %rax, (%rdi) +; SSE-NEXT: retq ; ; AVX-LABEL: memset_64_nonzero_bytes: ; AVX: vmovaps {{.*#+}} ymm0 = [42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42] @@ -65,25 +66,25 @@ define void @memset_64_nonzero_bytes(i8* %x) { } define void @memset_128_nonzero_bytes(i8* %x) { -; SSE2-LABEL: memset_128_nonzero_bytes: -; SSE2: movabsq $3038287259199220266, %rax # imm = 0x2A2A2A2A2A2A2A2A -; SSE2-NEXT: movq %rax, 120(%rdi) -; SSE2-NEXT: movq %rax, 112(%rdi) -; SSE2-NEXT: movq %rax, 104(%rdi) -; SSE2-NEXT: movq %rax, 96(%rdi) -; SSE2-NEXT: movq %rax, 88(%rdi) -; SSE2-NEXT: movq %rax, 80(%rdi) -; SSE2-NEXT: movq %rax, 72(%rdi) -; SSE2-NEXT: movq %rax, 64(%rdi) -; SSE2-NEXT: movq %rax, 56(%rdi) -; SSE2-NEXT: movq %rax, 48(%rdi) -; SSE2-NEXT: movq %rax, 40(%rdi) -; SSE2-NEXT: movq %rax, 32(%rdi) -; SSE2-NEXT: movq %rax, 24(%rdi) -; SSE2-NEXT: movq %rax, 16(%rdi) -; SSE2-NEXT: movq %rax, 8(%rdi) -; SSE2-NEXT: movq %rax, (%rdi) -; SSE2-NEXT: retq +; SSE-LABEL: memset_128_nonzero_bytes: +; SSE: movabsq $3038287259199220266, %rax # imm = 0x2A2A2A2A2A2A2A2A +; SSE-NEXT: movq %rax, 120(%rdi) +; SSE-NEXT: movq %rax, 112(%rdi) +; SSE-NEXT: movq %rax, 104(%rdi) +; SSE-NEXT: movq %rax, 96(%rdi) +; SSE-NEXT: movq %rax, 88(%rdi) +; SSE-NEXT: movq %rax, 80(%rdi) +; SSE-NEXT: movq %rax, 72(%rdi) +; SSE-NEXT: movq %rax, 64(%rdi) +; SSE-NEXT: movq %rax, 56(%rdi) +; SSE-NEXT: movq %rax, 48(%rdi) +; SSE-NEXT: movq %rax, 40(%rdi) +; SSE-NEXT: movq %rax, 32(%rdi) +; SSE-NEXT: movq %rax, 24(%rdi) +; SSE-NEXT: movq %rax, 16(%rdi) +; SSE-NEXT: movq %rax, 8(%rdi) +; SSE-NEXT: movq %rax, (%rdi) +; SSE-NEXT: retq ; ; AVX-LABEL: memset_128_nonzero_bytes: ; AVX: vmovaps {{.*#+}} ymm0 = [42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42] @@ -99,15 +100,15 @@ define void @memset_128_nonzero_bytes(i8* %x) { } define void @memset_256_nonzero_bytes(i8* %x) { -; SSE2-LABEL: memset_256_nonzero_bytes: -; SSE2: pushq %rax -; SSE2-NEXT: .Ltmp0: -; SSE2-NEXT: .cfi_def_cfa_offset 16 -; SSE2-NEXT: movl $42, %esi -; SSE2-NEXT: movl $256, %edx # imm = 0x100 -; SSE2-NEXT: callq memset -; SSE2-NEXT: popq %rax -; SSE2-NEXT: retq +; SSE-LABEL: memset_256_nonzero_bytes: +; SSE: pushq %rax +; SSE-NEXT: .Ltmp0: +; SSE-NEXT: .cfi_def_cfa_offset 16 +; SSE-NEXT: movl $42, %esi +; SSE-NEXT: movl $256, %edx # imm = 0x100 +; SSE-NEXT: callq memset +; SSE-NEXT: popq %rax +; SSE-NEXT: retq ; ; AVX-LABEL: memset_256_nonzero_bytes: ; AVX: vmovaps {{.*#+}} ymm0 = [42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42,42] @@ -131,13 +132,13 @@ declare i8* @__memset_chk(i8*, i32, i64, i64) ; Repeat with a non-constant value for the stores. define void @memset_16_nonconst_bytes(i8* %x, i8 %c) { -; SSE2-LABEL: memset_16_nonconst_bytes: -; SSE2: movzbl %sil, %eax -; SSE2-NEXT: movabsq $72340172838076673, %rcx # imm = 0x101010101010101 -; SSE2-NEXT: imulq %rax, %rcx -; SSE2-NEXT: movq %rcx, 8(%rdi) -; SSE2-NEXT: movq %rcx, (%rdi) -; SSE2-NEXT: retq +; SSE-LABEL: memset_16_nonconst_bytes: +; SSE: movzbl %sil, %eax +; SSE-NEXT: movabsq $72340172838076673, %rcx # imm = 0x101010101010101 +; SSE-NEXT: imulq %rax, %rcx +; SSE-NEXT: movq %rcx, 8(%rdi) +; SSE-NEXT: movq %rcx, (%rdi) +; SSE-NEXT: retq ; ; AVX1-LABEL: memset_16_nonconst_bytes: ; AVX1: vmovd %esi, %xmm0 @@ -157,15 +158,15 @@ define void @memset_16_nonconst_bytes(i8* %x, i8 %c) { } define void @memset_32_nonconst_bytes(i8* %x, i8 %c) { -; SSE2-LABEL: memset_32_nonconst_bytes: -; SSE2: movzbl %sil, %eax -; SSE2-NEXT: movabsq $72340172838076673, %rcx # imm = 0x101010101010101 -; SSE2-NEXT: imulq %rax, %rcx -; SSE2-NEXT: movq %rcx, 24(%rdi) -; SSE2-NEXT: movq %rcx, 16(%rdi) -; SSE2-NEXT: movq %rcx, 8(%rdi) -; SSE2-NEXT: movq %rcx, (%rdi) -; SSE2-NEXT: retq +; SSE-LABEL: memset_32_nonconst_bytes: +; SSE: movzbl %sil, %eax +; SSE-NEXT: movabsq $72340172838076673, %rcx # imm = 0x101010101010101 +; SSE-NEXT: imulq %rax, %rcx +; SSE-NEXT: movq %rcx, 24(%rdi) +; SSE-NEXT: movq %rcx, 16(%rdi) +; SSE-NEXT: movq %rcx, 8(%rdi) +; SSE-NEXT: movq %rcx, (%rdi) +; SSE-NEXT: retq ; ; AVX1-LABEL: memset_32_nonconst_bytes: ; AVX1: vmovd %esi, %xmm0 @@ -188,19 +189,19 @@ define void @memset_32_nonconst_bytes(i8* %x, i8 %c) { } define void @memset_64_nonconst_bytes(i8* %x, i8 %c) { -; SSE2-LABEL: memset_64_nonconst_bytes: -; SSE2: movzbl %sil, %eax -; SSE2-NEXT: movabsq $72340172838076673, %rcx # imm = 0x101010101010101 -; SSE2-NEXT: imulq %rax, %rcx -; SSE2-NEXT: movq %rcx, 56(%rdi) -; SSE2-NEXT: movq %rcx, 48(%rdi) -; SSE2-NEXT: movq %rcx, 40(%rdi) -; SSE2-NEXT: movq %rcx, 32(%rdi) -; SSE2-NEXT: movq %rcx, 24(%rdi) -; SSE2-NEXT: movq %rcx, 16(%rdi) -; SSE2-NEXT: movq %rcx, 8(%rdi) -; SSE2-NEXT: movq %rcx, (%rdi) -; SSE2-NEXT: retq +; SSE-LABEL: memset_64_nonconst_bytes: +; SSE: movzbl %sil, %eax +; SSE-NEXT: movabsq $72340172838076673, %rcx # imm = 0x101010101010101 +; SSE-NEXT: imulq %rax, %rcx +; SSE-NEXT: movq %rcx, 56(%rdi) +; SSE-NEXT: movq %rcx, 48(%rdi) +; SSE-NEXT: movq %rcx, 40(%rdi) +; SSE-NEXT: movq %rcx, 32(%rdi) +; SSE-NEXT: movq %rcx, 24(%rdi) +; SSE-NEXT: movq %rcx, 16(%rdi) +; SSE-NEXT: movq %rcx, 8(%rdi) +; SSE-NEXT: movq %rcx, (%rdi) +; SSE-NEXT: retq ; ; AVX1-LABEL: memset_64_nonconst_bytes: ; AVX1: vmovd %esi, %xmm0 @@ -225,27 +226,27 @@ define void @memset_64_nonconst_bytes(i8* %x, i8 %c) { } define void @memset_128_nonconst_bytes(i8* %x, i8 %c) { -; SSE2-LABEL: memset_128_nonconst_bytes: -; SSE2: movzbl %sil, %eax -; SSE2-NEXT: movabsq $72340172838076673, %rcx # imm = 0x101010101010101 -; SSE2-NEXT: imulq %rax, %rcx -; SSE2-NEXT: movq %rcx, 120(%rdi) -; SSE2-NEXT: movq %rcx, 112(%rdi) -; SSE2-NEXT: movq %rcx, 104(%rdi) -; SSE2-NEXT: movq %rcx, 96(%rdi) -; SSE2-NEXT: movq %rcx, 88(%rdi) -; SSE2-NEXT: movq %rcx, 80(%rdi) -; SSE2-NEXT: movq %rcx, 72(%rdi) -; SSE2-NEXT: movq %rcx, 64(%rdi) -; SSE2-NEXT: movq %rcx, 56(%rdi) -; SSE2-NEXT: movq %rcx, 48(%rdi) -; SSE2-NEXT: movq %rcx, 40(%rdi) -; SSE2-NEXT: movq %rcx, 32(%rdi) -; SSE2-NEXT: movq %rcx, 24(%rdi) -; SSE2-NEXT: movq %rcx, 16(%rdi) -; SSE2-NEXT: movq %rcx, 8(%rdi) -; SSE2-NEXT: movq %rcx, (%rdi) -; SSE2-NEXT: retq +; SSE-LABEL: memset_128_nonconst_bytes: +; SSE: movzbl %sil, %eax +; SSE-NEXT: movabsq $72340172838076673, %rcx # imm = 0x101010101010101 +; SSE-NEXT: imulq %rax, %rcx +; SSE-NEXT: movq %rcx, 120(%rdi) +; SSE-NEXT: movq %rcx, 112(%rdi) +; SSE-NEXT: movq %rcx, 104(%rdi) +; SSE-NEXT: movq %rcx, 96(%rdi) +; SSE-NEXT: movq %rcx, 88(%rdi) +; SSE-NEXT: movq %rcx, 80(%rdi) +; SSE-NEXT: movq %rcx, 72(%rdi) +; SSE-NEXT: movq %rcx, 64(%rdi) +; SSE-NEXT: movq %rcx, 56(%rdi) +; SSE-NEXT: movq %rcx, 48(%rdi) +; SSE-NEXT: movq %rcx, 40(%rdi) +; SSE-NEXT: movq %rcx, 32(%rdi) +; SSE-NEXT: movq %rcx, 24(%rdi) +; SSE-NEXT: movq %rcx, 16(%rdi) +; SSE-NEXT: movq %rcx, 8(%rdi) +; SSE-NEXT: movq %rcx, (%rdi) +; SSE-NEXT: retq ; ; AVX1-LABEL: memset_128_nonconst_bytes: ; AVX1: vmovd %esi, %xmm0 @@ -274,9 +275,9 @@ define void @memset_128_nonconst_bytes(i8* %x, i8 %c) { } define void @memset_256_nonconst_bytes(i8* %x, i8 %c) { -; SSE2-LABEL: memset_256_nonconst_bytes: -; SSE2: movl $256, %edx # imm = 0x100 -; SSE2-NEXT: jmp memset # TAILCALL +; SSE-LABEL: memset_256_nonconst_bytes: +; SSE: movl $256, %edx # imm = 0x100 +; SSE-NEXT: jmp memset # TAILCALL ; ; AVX1-LABEL: memset_256_nonconst_bytes: ; AVX1: vmovd %esi, %xmm0 |