diff options
author | Alexandros Lamprineas <alexandros.lamprineas@arm.com> | 2016-10-10 16:01:54 +0000 |
---|---|---|
committer | Alexandros Lamprineas <alexandros.lamprineas@arm.com> | 2016-10-10 16:01:54 +0000 |
commit | 20e9ddba7349b556e43daab7433f2a3a7c649417 (patch) | |
tree | 8f7d5c15ad03995244edb98dbbc6226f4caa993f /llvm/test/CodeGen/ARM/load_store_multiple.ll | |
parent | 0c21b40d37d559800490560ad4037f792cf69fab (diff) | |
download | bcm5719-llvm-20e9ddba7349b556e43daab7433f2a3a7c649417.tar.gz bcm5719-llvm-20e9ddba7349b556e43daab7433f2a3a7c649417.zip |
[ARM] Fix invalid VLDM/VSTM access when targeting Big Endian with NEON
The instructions VLDM/VSTM can only access word-aligned memory
locations and produce alignment fault if the condition is not met.
The compiler currently generates VLDM/VSTM for v2f64 load/store
regardless the alignment of the memory access. Instead, if a v2f64
load/store is not word-aligned, the compiler should generate
VLD1/VST1. For each non double-word-aligned VLD1/VST1, a VREV
instruction should be generated when targeting Big Endian.
Differential Revision: https://reviews.llvm.org/D25281
llvm-svn: 283763
Diffstat (limited to 'llvm/test/CodeGen/ARM/load_store_multiple.ll')
-rw-r--r-- | llvm/test/CodeGen/ARM/load_store_multiple.ll | 68 |
1 files changed, 68 insertions, 0 deletions
diff --git a/llvm/test/CodeGen/ARM/load_store_multiple.ll b/llvm/test/CodeGen/ARM/load_store_multiple.ll new file mode 100644 index 00000000000..5ea1f8cf6ff --- /dev/null +++ b/llvm/test/CodeGen/ARM/load_store_multiple.ll @@ -0,0 +1,68 @@ +; RUN: llc -mtriple=armv7-eabi -mattr=+neon %s -o - | FileCheck %s --check-prefix=CHECK-LE +; RUN: llc -mtriple=armv7eb-eabi -mattr=+neon %s -o - | FileCheck %s --check-prefix=CHECK-BE + +define void @ld_st_vec_i8(<16 x i8>* %A, <16 x i8>* %B) nounwind { +;CHECK-LE-LABEL: ld_st_vec_i8: +;CHECK-LE: vld1.8 {[[D1:d[0-9]+]], [[D2:d[0-9]+]]}, [{{r[0-9]+}}] +;CHECK-LE-NOT: vrev +;CHECK-LE: vst1.8 {[[D1]], [[D2]]}, [{{r[0-9]+}}] + +;CHECK-BE-LABEL: ld_st_vec_i8: +;CHECK-BE: vld1.8 {[[D1:d[0-9]+]], [[D2:d[0-9]+]]}, [{{r[0-9]+}}] +;CHECK-BE: vrev64.8 [[Q1:q[0-9]+]], [[Q2:q[0-9]+]] +;CHECK-BE: vrev64.8 [[Q1]], [[Q2]] +;CHECK-BE: vst1.8 {[[D1]], [[D2]]}, [{{r[0-9]+}}] + +%load = load <16 x i8>, <16 x i8>* %A, align 1 +store <16 x i8> %load, <16 x i8>* %B, align 1 +ret void +} + +define void @ld_st_vec_i16(<8 x i16>* %A, <8 x i16>* %B) nounwind { +;CHECK-LE-LABEL: ld_st_vec_i16: +;CHECK-LE: vld1.16 {[[D1:d[0-9]+]], [[D2:d[0-9]+]]}, [{{r[0-9]+}}] +;CHECK-LE-NOT: vrev +;CHECK-LE: vst1.16 {[[D1]], [[D2]]}, [{{r[0-9]+}}] + +;CHECK-BE-LABEL: ld_st_vec_i16: +;CHECK-BE: vld1.16 {[[D1:d[0-9]+]], [[D2:d[0-9]+]]}, [{{r[0-9]+}}] +;CHECK-BE: vrev64.16 [[Q1:q[0-9]+]], [[Q2:q[0-9]+]] +;CHECK-BE: vrev64.16 [[Q1]], [[Q2]] +;CHECK-BE: vst1.16 {[[D1]], [[D2]]}, [{{r[0-9]+}}] + +%load = load <8 x i16>, <8 x i16>* %A, align 2 +store <8 x i16> %load, <8 x i16>* %B, align 2 +ret void +} + +define void @ld_st_vec_i32(<4 x i32>* %A, <4 x i32>* %B) nounwind { +;CHECK-LE-LABEL: ld_st_vec_i32: +;CHECK-LE: vld1.32 {[[D1:d[0-9]+]], [[D2:d[0-9]+]]}, [{{r[0-9]+}}] +;CHECK-LE-NOT: vrev +;CHECK-LE: vst1.32 {[[D1]], [[D2]]}, [{{r[0-9]+}}] + +;CHECK-BE-LABEL: ld_st_vec_i32: +;CHECK-BE: vldmia {{r[0-9]+}}, {[[D1:d[0-9]+]], [[D2:d[0-9]+]]} +;CHECK-BE-NOT: vrev +;CHECK-BE: vstmia {{r[0-9]+}}, {[[D1]], [[D2]]} + +%load = load <4 x i32>, <4 x i32>* %A, align 4 +store <4 x i32> %load, <4 x i32>* %B, align 4 +ret void +} + +define void @ld_st_vec_double(<2 x double>* %A, <2 x double>* %B) nounwind { +;CHECK-LE-LABEL: ld_st_vec_double: +;CHECK-LE: vld1.64 {[[D1:d[0-9]+]], [[D2:d[0-9]+]]}, [{{r[0-9]+}}] +;CHECK-LE-NOT: vrev +;CHECK-LE: vst1.64 {[[D1]], [[D2]]}, [{{r[0-9]+}}] + +;CHECK-BE-LABEL: ld_st_vec_double: +;CHECK-BE: vld1.64 {[[D1:d[0-9]+]], [[D2:d[0-9]+]]}, [{{r[0-9]+}}] +;CHECK-BE-NOT: vrev +;CHECK-BE: vst1.64 {[[D1]], [[D2]]}, [{{r[0-9]+}}] + +%load = load <2 x double>, <2 x double>* %A, align 8 +store <2 x double> %load, <2 x double>* %B, align 8 +ret void +} |