summaryrefslogtreecommitdiffstats
path: root/llvm/test/CodeGen/AMDGPU/clamp-modifier.ll
Commit message (Collapse)AuthorAgeFilesLines
* AMDGPU: Improve hack for packing conversion opsMatt Arsenault2018-08-011-0/+23
| | | | | | | | | | | Mutate the node type during selection when it doesn't matter. This avoids an intermediate bitcast node on targets with legal i16/f16. Also fixes missing output modifiers on v_cvt_pkrtz_f32_f16, which I assume are OK. llvm-svn: 338619
* [AMDGPU] Enabled v2.16 literals for VOP3PStanislav Mekhanoshin2018-04-171-21/+14
| | | | | | | | Literal encoding needs op_sel_hi to select low 16 bit in this case. Differential Revision: https://reviews.llvm.org/D45745 llvm-svn: 330230
* AMDGPU: Fold clamp modifier for packed instructionsMatt Arsenault2017-08-311-15/+187
| | | | llvm-svn: 312297
* AMDGPU: Keep track of modifiers when converting v_mac to v_madMatt Arsenault2017-03-111-0/+17
| | | | | | | | | | | | | | | | Since v_max_f32_e64/v_max_f16_e64 can be folded if the target instruction supports the clamp bit, we also need to maintain modifiers when converting v_mac to v_mad. This fixes a rendering issue with Dirt Rally because a v_mac instruction with the clamp bit set was converted to a v_mad but that bit was lost during the conversion. Fixes: e184e01dd79 ("AMDGPU: Fold FP clamp as modifier bit") Patch by Samuel Pitoiset <samuel.pitoiset@gmail.com> llvm-svn: 297556
* AMDGPU: Use clamp with f64Matt Arsenault2017-02-221-0/+15
| | | | llvm-svn: 295908
* AMDGPU: Fold FP clamp as modifier bitMatt Arsenault2017-02-221-0/+190
The manual is unclear on the details of this. It's not clear to me if denormals are not allowed with clamp, or if that is only omod. Not allowing denorms for fp16 or fp64 isn't useful so I also question if that is really a restriction. Same with whether this is valid without IEEE mode enabled. llvm-svn: 295905
OpenPOWER on IntegriCloud