diff options
author | Sjoerd Meijer <sjoerd.meijer@arm.com> | 2018-02-27 19:26:02 +0000 |
---|---|---|
committer | Sjoerd Meijer <sjoerd.meijer@arm.com> | 2018-02-27 19:26:02 +0000 |
commit | fc0d02cbbfaaa91bf09666bffb2c2f64342bdd07 (patch) | |
tree | 78e630f29b1e2ab2e24fc486d052b08fd44de984 | |
parent | de979bfe58a27d7d228b87ac32aeb17dd502c7bc (diff) | |
download | bcm5719-llvm-fc0d02cbbfaaa91bf09666bffb2c2f64342bdd07.tar.gz bcm5719-llvm-fc0d02cbbfaaa91bf09666bffb2c2f64342bdd07.zip |
[ARM] Another f16 litpool fix
We were always setting the block alignment to 2 bytes in Thumb mode
and 4-bytes in ARM mode (r325754, and r325012), but this could cause
reducing the block alignment when it already had been aligned (e.g.
in Thumb mode when the block is a CPE that was already 4-byte aligned).
Patch by Momchil Velikov, I've only added a test.
Differential Revision: https://reviews.llvm.org/D43777
llvm-svn: 326232
-rw-r--r-- | llvm/lib/Target/ARM/ARMConstantIslandPass.cpp | 8 | ||||
-rw-r--r-- | llvm/test/CodeGen/ARM/fp16-litpool3-arm.mir | 113 |
2 files changed, 119 insertions, 2 deletions
diff --git a/llvm/lib/Target/ARM/ARMConstantIslandPass.cpp b/llvm/lib/Target/ARM/ARMConstantIslandPass.cpp index 46e804ded28..fc32fe86829 100644 --- a/llvm/lib/Target/ARM/ARMConstantIslandPass.cpp +++ b/llvm/lib/Target/ARM/ARMConstantIslandPass.cpp @@ -1485,8 +1485,12 @@ bool ARMConstantIslands::handleConstantPoolUser(unsigned CPUserIndex, // We are adding new water. Update NewWaterList. NewWaterList.insert(NewIsland); } - // Always align the new block because CP entries can be smaller than 4 bytes. - NewMBB->setAlignment(isThumb ? 1 : 2); + // Always align the new block because CP entries can be smaller than 4 + // bytes. Be careful not to decrease the existing alignment, e.g. NewMBB may + // be an already aligned constant pool block. + const unsigned Align = isThumb ? 1 : 2; + if (NewMBB->getAlignment() < Align) + NewMBB->setAlignment(Align); // Remove the original WaterList entry; we want subsequent insertions in // this vicinity to go after the one we're about to insert. This diff --git a/llvm/test/CodeGen/ARM/fp16-litpool3-arm.mir b/llvm/test/CodeGen/ARM/fp16-litpool3-arm.mir new file mode 100644 index 00000000000..6385d793d75 --- /dev/null +++ b/llvm/test/CodeGen/ARM/fp16-litpool3-arm.mir @@ -0,0 +1,113 @@ +# RUN: llc -mtriple=arm-none-eabi -run-pass=arm-cp-islands %s -o - | FileCheck %s +# +# This checks alignment of a block when a CPE is placed before/after a +# block (as e.g. opposed to splitting up a block), and also make sure +# we don't decrease alignment. +# +--- | + ; ModuleID = '<stdin>' + source_filename = "<stdin>" + target datalayout = "e-m:e-p:32:32-i64:64-v128:64:128-a:0:32-n32-S64" + target triple = "arm-arm--eabi" + + declare i32 @llvm.arm.space(i32, i32) #0 + + define dso_local i32 @CP() #1 { + entry: + %res = alloca half, align 2 + store half 0xH706B, half* %res, align 2 + %0 = load half, half* %res, align 2 + %tobool = fcmp une half %0, 0xH0000 + br i1 %tobool, label %LA, label %END + + LA: ; preds = %entry + %1 = call i32 @llvm.arm.space(i32 1000, i32 undef) + br label %END + + END: ; preds = %LA, %entry + %2 = call i32 @llvm.arm.space(i32 100, i32 undef) + ret i32 42 + } + + ; Function Attrs: nounwind + declare void @llvm.stackprotector(i8*, i8**) #2 + + attributes #0 = { nounwind "target-features"="+v8.2a,+fullfp16" } + attributes #1 = { "target-features"="+v8.2a,+fullfp16" } + attributes #2 = { nounwind } + +... +--- +name: CP +alignment: 2 +exposesReturnsTwice: false +legalized: false +regBankSelected: false +selected: false +tracksRegLiveness: true +registers: +liveins: +frameInfo: + isFrameAddressTaken: false + isReturnAddressTaken: false + hasStackMap: false + hasPatchPoint: false + stackSize: 4 + offsetAdjustment: 0 + maxAlignment: 2 + adjustsStack: false + hasCalls: false + stackProtector: '' + maxCallFrameSize: 0 + hasOpaqueSPAdjustment: false + hasVAStart: false + hasMustTailInVarArgFunc: false + savePoint: '' + restorePoint: '' +fixedStack: +stack: + - { id: 0, name: res, type: default, offset: -2, size: 2, alignment: 2, + stack-id: 0, callee-saved-register: '', callee-saved-restored: true, + local-offset: -2, di-variable: '', di-expression: '', di-location: '' } +constants: + - id: 0 + value: half 0xH706B + alignment: 2 + isTargetSpecific: false + + +#CHECK: bb.{{.*}} (align 1): +#CHECK: successors: +#CHECK: CONSTPOOL_ENTRY 1, %const{{.*}}, 2 +# +# We don't want to decrease alignment if the block already has been +# aligned; this can e.g. be an existing CPE that has been carefully +# aligned. Here BB.1.LA has already an 8-byte alignment, and we are +# checking we don't set it to 4: +# +#CHECK: bb.{{.*}}.LA (align 3): + +body: | + bb.0.entry: + successors: %bb.1(0x50000000), %bb.2(0x30000000) + + $sp = frame-setup SUBri $sp, 4, 14, $noreg, $noreg + frame-setup CFI_INSTRUCTION def_cfa_offset 4 + renamable $s0 = VLDRH %const.0, 0, 14, $noreg :: (load 2 from constant-pool) + VCMPZH renamable $s0, 14, $noreg, implicit-def $fpscr_nzcv + VSTRH killed renamable $s0, $sp, 1, 14, $noreg :: (store 2 into %ir.res) + FMSTAT 14, $noreg, implicit-def $cpsr, implicit killed $fpscr_nzcv + Bcc %bb.2, 0, killed $cpsr + + bb.1.LA (align 3): + successors: %bb.2(0x80000000) + + dead renamable $r0 = SPACE 1000, undef renamable $r0 + + bb.2.END: + dead renamable $r0 = SPACE 100, undef renamable $r0 + $r0 = MOVi 42, 14, $noreg, $noreg + $sp = ADDri $sp, 4, 14, $noreg, $noreg + BX_RET 14, $noreg, implicit killed $r0 + +... |