diff options
| author | Matt Arsenault <Matthew.Arsenault@amd.com> | 2018-08-16 17:07:52 +0000 |
|---|---|---|
| committer | Matt Arsenault <Matthew.Arsenault@amd.com> | 2018-08-16 17:07:52 +0000 |
| commit | 7121bed210d9f568485c9b5b6854b629353fa0ea (patch) | |
| tree | 89e0486ac8436096c1d1c843adb8f2b61e871e87 /llvm/test | |
| parent | 8b9e545477c64209aeeaa352d485cb9f317f1adb (diff) | |
| download | bcm5719-llvm-7121bed210d9f568485c9b5b6854b629353fa0ea.tar.gz bcm5719-llvm-7121bed210d9f568485c9b5b6854b629353fa0ea.zip | |
AMDGPU: Custom lower fexp
This will allow the library to just use __builtin_expf directly
without expanding this itself. Note f64 still won't work because
there is no exp instruction for it.
llvm-svn: 339902
Diffstat (limited to 'llvm/test')
| -rw-r--r-- | llvm/test/CodeGen/AMDGPU/fexp.ll | 307 |
1 files changed, 307 insertions, 0 deletions
diff --git a/llvm/test/CodeGen/AMDGPU/fexp.ll b/llvm/test/CodeGen/AMDGPU/fexp.ll new file mode 100644 index 00000000000..dd69335fa38 --- /dev/null +++ b/llvm/test/CodeGen/AMDGPU/fexp.ll @@ -0,0 +1,307 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +;RUN: llc -mtriple=amdgcn-- < %s | FileCheck -enable-var-scope -check-prefix=SI %s +;RUN: llc -mtriple=amdgcn-- -mcpu=fiji < %s | FileCheck -enable-var-scope -check-prefix=VI %s +;RUN: llc -mtriple=amdgcn-- -mcpu=gfx900 < %s | FileCheck -enable-var-scope -check-prefix=GFX9 %s + +define float @v_exp_f32(float %arg0) { +; SI-LABEL: v_exp_f32: +; SI: ; %bb.0: +; SI-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; SI-NEXT: v_mul_f32_e32 v0, 0x3fb8aa3b, v0 +; SI-NEXT: v_exp_f32_e32 v0, v0 +; SI-NEXT: s_setpc_b64 s[30:31] +; +; VI-LABEL: v_exp_f32: +; VI: ; %bb.0: +; VI-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; VI-NEXT: v_mul_f32_e32 v0, 0x3fb8aa3b, v0 +; VI-NEXT: v_exp_f32_e32 v0, v0 +; VI-NEXT: s_setpc_b64 s[30:31] +; +; GFX9-LABEL: v_exp_f32: +; GFX9: ; %bb.0: +; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX9-NEXT: v_mul_f32_e32 v0, 0x3fb8aa3b, v0 +; GFX9-NEXT: v_exp_f32_e32 v0, v0 +; GFX9-NEXT: s_setpc_b64 s[30:31] + %result = call float @llvm.exp.f32(float %arg0) + ret float %result +} + +define <2 x float> @v_exp_v2f32(<2 x float> %arg0) { +; SI-LABEL: v_exp_v2f32: +; SI: ; %bb.0: +; SI-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; SI-NEXT: v_mov_b32_e32 v2, 0x3fb8aa3b +; SI-NEXT: v_mul_f32_e32 v0, v0, v2 +; SI-NEXT: v_mul_f32_e32 v1, v1, v2 +; SI-NEXT: v_exp_f32_e32 v0, v0 +; SI-NEXT: v_exp_f32_e32 v1, v1 +; SI-NEXT: s_setpc_b64 s[30:31] +; +; VI-LABEL: v_exp_v2f32: +; VI: ; %bb.0: +; VI-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; VI-NEXT: v_mov_b32_e32 v2, 0x3fb8aa3b +; VI-NEXT: v_mul_f32_e32 v0, v0, v2 +; VI-NEXT: v_mul_f32_e32 v1, v1, v2 +; VI-NEXT: v_exp_f32_e32 v0, v0 +; VI-NEXT: v_exp_f32_e32 v1, v1 +; VI-NEXT: s_setpc_b64 s[30:31] +; +; GFX9-LABEL: v_exp_v2f32: +; GFX9: ; %bb.0: +; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX9-NEXT: v_mov_b32_e32 v2, 0x3fb8aa3b +; GFX9-NEXT: v_mul_f32_e32 v0, v0, v2 +; GFX9-NEXT: v_mul_f32_e32 v1, v1, v2 +; GFX9-NEXT: v_exp_f32_e32 v0, v0 +; GFX9-NEXT: v_exp_f32_e32 v1, v1 +; GFX9-NEXT: s_setpc_b64 s[30:31] + %result = call <2 x float> @llvm.exp.v2f32(<2 x float> %arg0) + ret <2 x float> %result +} + +define <3 x float> @v_exp_v3f32(<3 x float> %arg0) { +; SI-LABEL: v_exp_v3f32: +; SI: ; %bb.0: +; SI-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; SI-NEXT: v_mov_b32_e32 v3, 0x3fb8aa3b +; SI-NEXT: v_mul_f32_e32 v0, v0, v3 +; SI-NEXT: v_mul_f32_e32 v1, v1, v3 +; SI-NEXT: v_mul_f32_e32 v2, v2, v3 +; SI-NEXT: v_exp_f32_e32 v0, v0 +; SI-NEXT: v_exp_f32_e32 v1, v1 +; SI-NEXT: v_exp_f32_e32 v2, v2 +; SI-NEXT: s_setpc_b64 s[30:31] +; +; VI-LABEL: v_exp_v3f32: +; VI: ; %bb.0: +; VI-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; VI-NEXT: v_mov_b32_e32 v3, 0x3fb8aa3b +; VI-NEXT: v_mul_f32_e32 v0, v0, v3 +; VI-NEXT: v_mul_f32_e32 v1, v1, v3 +; VI-NEXT: v_mul_f32_e32 v2, v2, v3 +; VI-NEXT: v_exp_f32_e32 v0, v0 +; VI-NEXT: v_exp_f32_e32 v1, v1 +; VI-NEXT: v_exp_f32_e32 v2, v2 +; VI-NEXT: s_setpc_b64 s[30:31] +; +; GFX9-LABEL: v_exp_v3f32: +; GFX9: ; %bb.0: +; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX9-NEXT: v_mov_b32_e32 v3, 0x3fb8aa3b +; GFX9-NEXT: v_mul_f32_e32 v0, v0, v3 +; GFX9-NEXT: v_mul_f32_e32 v1, v1, v3 +; GFX9-NEXT: v_mul_f32_e32 v2, v2, v3 +; GFX9-NEXT: v_exp_f32_e32 v0, v0 +; GFX9-NEXT: v_exp_f32_e32 v1, v1 +; GFX9-NEXT: v_exp_f32_e32 v2, v2 +; GFX9-NEXT: s_setpc_b64 s[30:31] + %result = call <3 x float> @llvm.exp.v3f32(<3 x float> %arg0) + ret <3 x float> %result +} + +define <4 x float> @v_exp_v4f32(<4 x float> %arg0) { +; SI-LABEL: v_exp_v4f32: +; SI: ; %bb.0: +; SI-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; SI-NEXT: v_mov_b32_e32 v4, 0x3fb8aa3b +; SI-NEXT: v_mul_f32_e32 v0, v0, v4 +; SI-NEXT: v_mul_f32_e32 v1, v1, v4 +; SI-NEXT: v_mul_f32_e32 v2, v2, v4 +; SI-NEXT: v_mul_f32_e32 v3, v3, v4 +; SI-NEXT: v_exp_f32_e32 v0, v0 +; SI-NEXT: v_exp_f32_e32 v1, v1 +; SI-NEXT: v_exp_f32_e32 v2, v2 +; SI-NEXT: v_exp_f32_e32 v3, v3 +; SI-NEXT: s_setpc_b64 s[30:31] +; +; VI-LABEL: v_exp_v4f32: +; VI: ; %bb.0: +; VI-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; VI-NEXT: v_mov_b32_e32 v4, 0x3fb8aa3b +; VI-NEXT: v_mul_f32_e32 v0, v0, v4 +; VI-NEXT: v_mul_f32_e32 v1, v1, v4 +; VI-NEXT: v_mul_f32_e32 v2, v2, v4 +; VI-NEXT: v_mul_f32_e32 v3, v3, v4 +; VI-NEXT: v_exp_f32_e32 v0, v0 +; VI-NEXT: v_exp_f32_e32 v1, v1 +; VI-NEXT: v_exp_f32_e32 v2, v2 +; VI-NEXT: v_exp_f32_e32 v3, v3 +; VI-NEXT: s_setpc_b64 s[30:31] +; +; GFX9-LABEL: v_exp_v4f32: +; GFX9: ; %bb.0: +; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX9-NEXT: v_mov_b32_e32 v4, 0x3fb8aa3b +; GFX9-NEXT: v_mul_f32_e32 v0, v0, v4 +; GFX9-NEXT: v_mul_f32_e32 v1, v1, v4 +; GFX9-NEXT: v_mul_f32_e32 v2, v2, v4 +; GFX9-NEXT: v_mul_f32_e32 v3, v3, v4 +; GFX9-NEXT: v_exp_f32_e32 v0, v0 +; GFX9-NEXT: v_exp_f32_e32 v1, v1 +; GFX9-NEXT: v_exp_f32_e32 v2, v2 +; GFX9-NEXT: v_exp_f32_e32 v3, v3 +; GFX9-NEXT: s_setpc_b64 s[30:31] + %result = call <4 x float> @llvm.exp.v4f32(<4 x float> %arg0) + ret <4 x float> %result +} + +define half @v_exp_f16(half %arg0) { +; SI-LABEL: v_exp_f16: +; SI: ; %bb.0: +; SI-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; SI-NEXT: v_cvt_f16_f32_e32 v0, v0 +; SI-NEXT: v_cvt_f32_f16_e32 v0, v0 +; SI-NEXT: v_mul_f32_e32 v0, 0x3fb8aa3b, v0 +; SI-NEXT: v_exp_f32_e32 v0, v0 +; SI-NEXT: s_setpc_b64 s[30:31] +; +; VI-LABEL: v_exp_f16: +; VI: ; %bb.0: +; VI-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; VI-NEXT: v_mul_f16_e32 v0, 0x3dc5, v0 +; VI-NEXT: v_exp_f16_e32 v0, v0 +; VI-NEXT: s_setpc_b64 s[30:31] +; +; GFX9-LABEL: v_exp_f16: +; GFX9: ; %bb.0: +; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX9-NEXT: v_mul_f16_e32 v0, 0x3dc5, v0 +; GFX9-NEXT: v_exp_f16_e32 v0, v0 +; GFX9-NEXT: s_setpc_b64 s[30:31] + %result = call half @llvm.exp.f16(half %arg0) + ret half %result +} + +define <2 x half> @v_exp_v2f16(<2 x half> %arg0) { +; SI-LABEL: v_exp_v2f16: +; SI: ; %bb.0: +; SI-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; SI-NEXT: v_cvt_f16_f32_e32 v1, v1 +; SI-NEXT: v_cvt_f16_f32_e32 v0, v0 +; SI-NEXT: v_mov_b32_e32 v2, 0x3fb8aa3b +; SI-NEXT: v_cvt_f32_f16_e32 v1, v1 +; SI-NEXT: v_cvt_f32_f16_e32 v0, v0 +; SI-NEXT: v_mul_f32_e32 v0, v0, v2 +; SI-NEXT: v_mul_f32_e32 v1, v1, v2 +; SI-NEXT: v_exp_f32_e32 v0, v0 +; SI-NEXT: v_exp_f32_e32 v1, v1 +; SI-NEXT: s_setpc_b64 s[30:31] +; +; VI-LABEL: v_exp_v2f16: +; VI: ; %bb.0: +; VI-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; VI-NEXT: v_mov_b32_e32 v1, 0x3dc5 +; VI-NEXT: v_mul_f16_sdwa v2, v0, v1 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1 src1_sel:DWORD +; VI-NEXT: v_mul_f16_e32 v0, v0, v1 +; VI-NEXT: v_exp_f16_sdwa v2, v2 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:DWORD +; VI-NEXT: v_exp_f16_e32 v0, v0 +; VI-NEXT: v_or_b32_e32 v0, v0, v2 +; VI-NEXT: s_setpc_b64 s[30:31] +; +; GFX9-LABEL: v_exp_v2f16: +; GFX9: ; %bb.0: +; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX9-NEXT: v_mov_b32_e32 v1, 0x3dc5 +; GFX9-NEXT: v_pk_mul_f16 v0, v0, v1 op_sel_hi:[1,0] +; GFX9-NEXT: v_exp_f16_e32 v1, v0 +; GFX9-NEXT: v_exp_f16_sdwa v0, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1 +; GFX9-NEXT: v_and_b32_e32 v1, 0xffff, v1 +; GFX9-NEXT: v_lshl_or_b32 v0, v0, 16, v1 +; GFX9-NEXT: s_setpc_b64 s[30:31] + %result = call <2 x half> @llvm.exp.v2f16(<2 x half> %arg0) + ret <2 x half> %result +} + +; define <3 x half> @v_exp_v3f16(<3 x half> %arg0) { +; %result = call <3 x half> @llvm.exp.v3f16(<3 x half> %arg0) +; ret <3 x half> %result +; } + +define <4 x half> @v_exp_v4f16(<4 x half> %arg0) { +; SI-LABEL: v_exp_v4f16: +; SI: ; %bb.0: +; SI-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; SI-NEXT: v_cvt_f16_f32_e32 v0, v0 +; SI-NEXT: v_cvt_f16_f32_e32 v1, v1 +; SI-NEXT: v_cvt_f16_f32_e32 v2, v2 +; SI-NEXT: v_cvt_f16_f32_e32 v3, v3 +; SI-NEXT: v_mov_b32_e32 v4, 0x3fb8aa3b +; SI-NEXT: v_cvt_f32_f16_e32 v0, v0 +; SI-NEXT: v_cvt_f32_f16_e32 v1, v1 +; SI-NEXT: v_cvt_f32_f16_e32 v2, v2 +; SI-NEXT: v_cvt_f32_f16_e32 v3, v3 +; SI-NEXT: v_mul_f32_e32 v3, v3, v4 +; SI-NEXT: v_mul_f32_e32 v2, v2, v4 +; SI-NEXT: v_mul_f32_e32 v1, v1, v4 +; SI-NEXT: v_mul_f32_e32 v0, v0, v4 +; SI-NEXT: v_exp_f32_e32 v3, v3 +; SI-NEXT: v_exp_f32_e32 v2, v2 +; SI-NEXT: v_exp_f32_e32 v1, v1 +; SI-NEXT: v_exp_f32_e32 v0, v0 +; SI-NEXT: v_cvt_f16_f32_e32 v3, v3 +; SI-NEXT: v_cvt_f16_f32_e32 v2, v2 +; SI-NEXT: v_cvt_f16_f32_e32 v1, v1 +; SI-NEXT: v_cvt_f16_f32_e32 v0, v0 +; SI-NEXT: v_lshlrev_b32_e32 v3, 16, v3 +; SI-NEXT: v_lshlrev_b32_e32 v1, 16, v1 +; SI-NEXT: v_or_b32_e32 v3, v2, v3 +; SI-NEXT: v_or_b32_e32 v1, v0, v1 +; SI-NEXT: v_cvt_f32_f16_e32 v0, v1 +; SI-NEXT: v_cvt_f32_f16_e32 v2, v3 +; SI-NEXT: v_lshrrev_b32_e32 v1, 16, v1 +; SI-NEXT: v_lshrrev_b32_e32 v3, 16, v3 +; SI-NEXT: v_cvt_f32_f16_e32 v1, v1 +; SI-NEXT: v_cvt_f32_f16_e32 v3, v3 +; SI-NEXT: s_setpc_b64 s[30:31] +; +; VI-LABEL: v_exp_v4f16: +; VI: ; %bb.0: +; VI-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; VI-NEXT: v_mov_b32_e32 v2, 0x3dc5 +; VI-NEXT: v_mul_f16_e32 v3, v1, v2 +; VI-NEXT: v_mul_f16_e32 v4, v0, v2 +; VI-NEXT: v_mul_f16_sdwa v1, v1, v2 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1 src1_sel:DWORD +; VI-NEXT: v_mul_f16_sdwa v0, v0, v2 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1 src1_sel:DWORD +; VI-NEXT: v_exp_f16_e32 v3, v3 +; VI-NEXT: v_exp_f16_sdwa v1, v1 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:DWORD +; VI-NEXT: v_exp_f16_e32 v4, v4 +; VI-NEXT: v_exp_f16_sdwa v0, v0 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:DWORD +; VI-NEXT: v_or_b32_e32 v1, v3, v1 +; VI-NEXT: v_or_b32_e32 v0, v4, v0 +; VI-NEXT: s_setpc_b64 s[30:31] +; +; GFX9-LABEL: v_exp_v4f16: +; GFX9: ; %bb.0: +; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX9-NEXT: v_mov_b32_e32 v2, 0x3dc5 +; GFX9-NEXT: v_mul_f16_e32 v3, v1, v2 +; GFX9-NEXT: v_mul_f16_e32 v4, v0, v2 +; GFX9-NEXT: v_mul_f16_sdwa v1, v1, v2 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1 src1_sel:DWORD +; GFX9-NEXT: v_mul_f16_sdwa v0, v0, v2 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1 src1_sel:DWORD +; GFX9-NEXT: v_exp_f16_e32 v3, v3 +; GFX9-NEXT: v_exp_f16_e32 v4, v4 +; GFX9-NEXT: v_exp_f16_e32 v0, v0 +; GFX9-NEXT: v_exp_f16_e32 v1, v1 +; GFX9-NEXT: v_mov_b32_e32 v2, 0xffff +; GFX9-NEXT: v_and_b32_e32 v4, v2, v4 +; GFX9-NEXT: v_and_b32_e32 v2, v2, v3 +; GFX9-NEXT: v_lshl_or_b32 v0, v0, 16, v4 +; GFX9-NEXT: v_lshl_or_b32 v1, v1, 16, v2 +; GFX9-NEXT: s_setpc_b64 s[30:31] + %result = call <4 x half> @llvm.exp.v4f16(<4 x half> %arg0) + ret <4 x half> %result +} + +declare float @llvm.exp.f32(float) +declare <2 x float> @llvm.exp.v2f32(<2 x float>) +declare <3 x float> @llvm.exp.v3f32(<3 x float>) +declare <4 x float> @llvm.exp.v4f32(<4 x float>) + +declare half @llvm.exp.f16(half) +declare <2 x half> @llvm.exp.v2f16(<2 x half>) +declare <3 x half> @llvm.exp.v3f16(<3 x half>) +declare <4 x half> @llvm.exp.v4f16(<4 x half>) + |

