; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc -mtriple=thumbv8.1m.main-arm-none-eabi -mattr=+mve.fp -verify-machineinstrs %s -o - | FileCheck %s define void @vldst4(half* nocapture readonly %pIn, half* nocapture %pOut, i32 %numRows, i32 %numCols, i32 %scale.coerce) #0 { ; CHECK-LABEL: vldst4: ; CHECK: @ %bb.0: @ %entry ; CHECK-NEXT: .save {r7, lr} ; CHECK-NEXT: push {r7, lr} ; CHECK-NEXT: .vsave {d8, d9, d10, d11, d12, d13, d14, d15} ; CHECK-NEXT: vpush {d8, d9, d10, d11, d12, d13, d14, d15} ; CHECK-NEXT: .pad #104 ; CHECK-NEXT: sub sp, #104 ; CHECK-NEXT: muls r2, r3, r2 ; CHECK-NEXT: movs r3, #0 ; CHECK-NEXT: cmp.w r3, r2, lsr #2 ; CHECK-NEXT: beq.w .LBB0_3 ; CHECK-NEXT: @ %bb.1: @ %vector.ph ; CHECK-NEXT: mvn r3, #7 ; CHECK-NEXT: and.w r2, r3, r2, lsr #2 ; CHECK-NEXT: vldr.16 s0, [sp, #176] ; CHECK-NEXT: subs r2, #8 ; CHECK-NEXT: movs r3, #1 ; CHECK-NEXT: sub.w r12, r0, #64 ; CHECK-NEXT: add.w lr, r3, r2, lsr #3 ; CHECK-NEXT: vmov r2, s0 ; CHECK-NEXT: vdup.16 q0, r2 ; CHECK-NEXT: subs r1, #64 ; CHECK-NEXT: vstrw.32 q0, [sp, #80] @ 16-byte Spill ; CHECK-NEXT: dls lr, lr ; CHECK-NEXT: .LBB0_2: @ %vector.body ; CHECK-NEXT: @ =>This Inner Loop Header: Depth=1 ; CHECK-NEXT: vldrh.u16 q0, [r12, #64]! ; CHECK-NEXT: vldrh.u16 q6, [r12, #32] ; CHECK-NEXT: vldrh.u16 q4, [r12, #48] ; CHECK-NEXT: vldrh.u16 q7, [r12, #16] ; CHECK-NEXT: vmov r2, s24 ; CHECK-NEXT: vmovx.f16 s12, s16 ; CHECK-NEXT: vmov.16 q1[4], r2 ; CHECK-NEXT: vmov r3, s26 ; CHECK-NEXT: vmov.16 q1[5], r3 ; CHECK-NEXT: vmov r2, s16 ; CHECK-NEXT: vmov.16 q1[6], r2 ; CHECK-NEXT: vmov r2, s0 ; CHECK-NEXT: vmov.16 q2[0], r2 ; CHECK-NEXT: vmov r3, s2 ; CHECK-NEXT: vmov.16 q2[1], r3 ; CHECK-NEXT: vmov r2, s28 ; CHECK-NEXT: vmov.16 q2[2], r2 ; CHECK-NEXT: vmov r2, s18 ; CHECK-NEXT: vmov.16 q1[7], r2 ; CHECK-NEXT: vmov r2, s30 ; CHECK-NEXT: vmov.16 q2[3], r2 ; CHECK-NEXT: vmov.f32 s10, s6 ; CHECK-NEXT: vmov.f32 s11, s7 ; CHECK-NEXT: vldrw.u32 q1, [sp, #80] @ 16-byte Reload ; CHECK-NEXT: vmul.f16 q5, q2, q1 ; CHECK-NEXT: vmovx.f16 s4, s2 ; CHECK-NEXT: vmov r3, s4 ; CHECK-NEXT: vmovx.f16 s4, s0 ; CHECK-NEXT: vmov r0, s4 ; CHECK-NEXT: vmovx.f16 s8, s28 ; CHECK-NEXT: vmov.16 q1[0], r0 ; CHECK-NEXT: vmov r0, s8 ; CHECK-NEXT: vmov.16 q1[1], r3 ; CHECK-NEXT: vmovx.f16 s8, s26 ; CHECK-NEXT: vmov.16 q1[2], r0 ; CHECK-NEXT: vmov r0, s8 ; CHECK-NEXT: vmovx.f16 s8, s24 ; CHECK-NEXT: vmov r2, s22 ; CHECK-NEXT: vmov r3, s8 ; CHECK-NEXT: vmovx.f16 s0, s1 ; CHECK-NEXT: vmov.16 q2[4], r3 ; CHECK-NEXT: vmov r3, s0 ; CHECK-NEXT: vmov.16 q2[5], r0 ; CHECK-NEXT: vmov r0, s12 ; CHECK-NEXT: vmovx.f16 s12, s18 ; CHECK-NEXT: vmov.16 q2[6], r0 ; CHECK-NEXT: vmov r0, s12 ; CHECK-NEXT: vmovx.f16 s12, s30 ; CHECK-NEXT: vmov.16 q2[7], r0 ; CHECK-NEXT: vmov r0, s12 ; CHECK-NEXT: vmov.16 q1[3], r0 ; CHECK-NEXT: vldrw.u32 q3, [sp, #80] @ 16-byte Reload ; CHECK-NEXT: vmov.f32 s6, s10 ; CHECK-NEXT: vstrw.32 q5, [sp, #16] @ 16-byte Spill ; CHECK-NEXT: vmov.f32 s7, s11 ; CHECK-NEXT: vmov.16 q2[0], r2 ; CHECK-NEXT: vmul.f16 q1, q1, q3 ; CHECK-NEXT: vmov r2, s27 ; CHECK-NEXT: vmov r0, s6 ; CHECK-NEXT: vstrw.32 q1, [sp, #32] @ 16-byte Spill ; CHECK-NEXT: vmovx.f16 s4, s22 ; CHECK-NEXT: vmov.16 q2[1], r0 ; CHECK-NEXT: vmov r0, s4 ; CHECK-NEXT: vmov.16 q2[4], r0 ; CHECK-NEXT: vmov r0, s25 ; CHECK-NEXT: vmov.16 q1[4], r0 ; CHECK-NEXT: vmov r0, s17 ; CHECK-NEXT: vmov.16 q1[5], r2 ; CHECK-NEXT: vmov r2, s1 ; CHECK-NEXT: vstrw.32 q2, [sp, #64] @ 16-byte Spill ; CHECK-NEXT: vmov.16 q1[6], r0 ; CHECK-NEXT: vmov r0, s3 ; CHECK-NEXT: vmov.16 q2[0], r2 ; CHECK-NEXT: vmov.16 q2[1], r0 ; CHECK-NEXT: vmov r0, s29 ; CHECK-NEXT: vmov.16 q2[2], r0 ; CHECK-NEXT: vmov r0, s19 ; CHECK-NEXT: vmov.16 q1[7], r0 ; CHECK-NEXT: vmov r0, s31 ; CHECK-NEXT: vmov.16 q2[3], r0 ; CHECK-NEXT: vmov.f32 s10, s6 ; CHECK-NEXT: vmov.f32 s11, s7 ; CHECK-NEXT: vmovx.f16 s4, s3 ; CHECK-NEXT: vmov r0, s4 ; CHECK-NEXT: vmov.16 q0[0], r3 ; CHECK-NEXT: vmovx.f16 s4, s29 ; CHECK-NEXT: vmov.16 q0[1], r0 ; CHECK-NEXT: vmov r0, s4 ; CHECK-NEXT: vmovx.f16 s4, s27 ; CHECK-NEXT: vmov.16 q0[2], r0 ; CHECK-NEXT: vmov r0, s4 ; CHECK-NEXT: vmovx.f16 s4, s25 ; CHECK-NEXT: vmul.f16 q5, q2, q3 ; CHECK-NEXT: vmov r3, s4 ; CHECK-NEXT: vmovx.f16 s8, s17 ; CHECK-NEXT: vmov.16 q1[4], r3 ; CHECK-NEXT: vmov r2, s23 ; CHECK-NEXT: vmov.16 q1[5], r0 ; CHECK-NEXT: vmov r0, s8 ; CHECK-NEXT: vmovx.f16 s8, s19 ; CHECK-NEXT: vmov.16 q1[6], r0 ; CHECK-NEXT: vmov r0, s8 ; CHECK-NEXT: vmovx.f16 s8, s31 ; CHECK-NEXT: vmov.16 q1[7], r0 ; CHECK-NEXT: vmov r0, s8 ; CHECK-NEXT: vmov.16 q0[3], r0 ; CHECK-NEXT: vldrw.u32 q2, [sp, #32] @ 16-byte Reload ; CHECK-NEXT: vmov.f32 s2, s6 ; CHECK-NEXT: vmov.f32 s3, s7 ; CHECK-NEXT: vmov.16 q1[2], r2 ; CHECK-NEXT: vmul.f16 q6, q0, q3 ; CHECK-NEXT: vmovx.f16 s0, s23 ; CHECK-NEXT: vmov r0, s27 ; CHECK-NEXT: vmov.16 q1[3], r0 ; CHECK-NEXT: vmov r0, s0 ; CHECK-NEXT: vmovx.f16 s0, s27 ; CHECK-NEXT: vmov.16 q1[6], r0 ; CHECK-NEXT: vmov r0, s0 ; CHECK-NEXT: vmov.16 q1[7], r0 ; CHECK-NEXT: vmov r2, s11 ; CHECK-NEXT: vstrw.32 q1, [sp, #48] @ 16-byte Spill ; CHECK-NEXT: vldrw.u32 q1, [sp, #16] @ 16-byte Reload ; CHECK-NEXT: vmov r0, s7 ; CHECK-NEXT: vmovx.f16 s0, s7 ; CHECK-NEXT: vmov.16 q4[0], r0 ; CHECK-NEXT: vmov r0, s0 ; CHECK-NEXT: vmov.16 q4[1], r2 ; CHECK-NEXT: vmovx.f16 s0, s11 ; CHECK-NEXT: vmov.16 q4[4], r0 ; CHECK-NEXT: vmov r0, s0 ; CHECK-NEXT: vmov.16 q4[5], r0 ; CHECK-NEXT: vmov r0, s21 ; CHECK-NEXT: vmov r2, s25 ; CHECK-NEXT: vmov.16 q3[2], r0 ; CHECK-NEXT: vmovx.f16 s0, s21 ; CHECK-NEXT: vmov.16 q3[3], r2 ; CHECK-NEXT: vmov r0, s0 ; CHECK-NEXT: vmovx.f16 s0, s25 ; CHECK-NEXT: vmov.16 q3[6], r0 ; CHECK-NEXT: vmov r0, s0 ; CHECK-NEXT: vmov.16 q3[7], r0 ; CHECK-NEXT: vmov r0, s5 ; CHECK-NEXT: vmov r2, s9 ; CHECK-NEXT: vmov.16 q7[0], r0 ; CHECK-NEXT: vmovx.f16 s0, s5 ; CHECK-NEXT: vmov.16 q7[1], r2 ; CHECK-NEXT: vmov r0, s0 ; CHECK-NEXT: vmovx.f16 s0, s9 ; CHECK-NEXT: vmov.16 q7[4], r0 ; CHECK-NEXT: vmov r0, s0 ; CHECK-NEXT: vmov.16 q7[5], r0 ; CHECK-NEXT: vmov r0, s20 ; CHECK-NEXT: vstrw.32 q3, [sp] @ 16-byte Spill ; CHECK-NEXT: vmovx.f16 s0, s20 ; CHECK-NEXT: vmov.16 q3[2], r0 ; CHECK-NEXT: vmov r2, s24 ; CHECK-NEXT: vmov.16 q3[3], r2 ; CHECK-NEXT: vmov r0, s0 ; CHECK-NEXT: vmovx.f16 s0, s24 ; CHECK-NEXT: vmov.16 q3[6], r0 ; CHECK-NEXT: vmov r0, s0 ; CHECK-NEXT: vmovx.f16 s20, s22 ; CHECK-NEXT: vmov.16 q3[7], r0 ; CHECK-NEXT: vmov r0, s4 ; CHECK-NEXT: vmov r2, s8 ; CHECK-NEXT: vmov.16 q0[0], r0 ; CHECK-NEXT: vmovx.f16 s4, s4 ; CHECK-NEXT: vmov.16 q0[1], r2 ; CHECK-NEXT: vmov r0, s4 ; CHECK-NEXT: vmovx.f16 s4, s8 ; CHECK-NEXT: vmov.16 q0[4], r0 ; CHECK-NEXT: vmov r0, s4 ; CHECK-NEXT: vmov.16 q0[5], r0 ; CHECK-NEXT: vmov r0, s22 ; CHECK-NEXT: vmov.16 q1[2], r0 ; CHECK-NEXT: vmov r2, s26 ; CHECK-NEXT: vmov r0, s20 ; CHECK-NEXT: vmov.16 q1[3], r2 ; CHECK-NEXT: vmovx.f16 s20, s26 ; CHECK-NEXT: vmov.16 q1[6], r0 ; CHECK-NEXT: vmov r0, s20 ; CHECK-NEXT: vmovx.f16 s20, s10 ; CHECK-NEXT: vmov.16 q1[7], r0 ; CHECK-NEXT: vmov r0, s20 ; CHECK-NEXT: vldrw.u32 q5, [sp, #64] @ 16-byte Reload ; CHECK-NEXT: vldrw.u32 q2, [sp] @ 16-byte Reload ; CHECK-NEXT: vldrw.u32 q6, [sp, #48] @ 16-byte Reload ; CHECK-NEXT: vmov.f32 s1, s13 ; CHECK-NEXT: vmov.f32 s29, s9 ; CHECK-NEXT: vmov.16 q5[5], r0 ; CHECK-NEXT: vmov.f32 s31, s11 ; CHECK-NEXT: vmov q2, q5 ; CHECK-NEXT: vmov.f32 s17, s25 ; CHECK-NEXT: vmov.f32 s9, s5 ; CHECK-NEXT: vmov.f32 s3, s15 ; CHECK-NEXT: vstrb.8 q0, [r1, #64]! ; CHECK-NEXT: vmov.f32 s19, s27 ; CHECK-NEXT: vmov.f32 s11, s7 ; CHECK-NEXT: vstrh.16 q4, [r1, #48] ; CHECK-NEXT: vstrh.16 q2, [r1, #32] ; CHECK-NEXT: vstrh.16 q7, [r1, #16] ; CHECK-NEXT: le lr, .LBB0_2 ; CHECK-NEXT: .LBB0_3: @ %while.end ; CHECK-NEXT: add sp, #104 ; CHECK-NEXT: vpop {d8, d9, d10, d11, d12, d13, d14, d15} ; CHECK-NEXT: pop {r7, pc} entry: %tmp.0.extract.trunc = trunc i32 %scale.coerce to i16 %l0 = bitcast i16 %tmp.0.extract.trunc to half %mul = mul i32 %numCols, %numRows %shr = lshr i32 %mul, 2 %cmp38 = icmp eq i32 %shr, 0 br i1 %cmp38, label %while.end, label %vector.ph vector.ph: ; preds = %vector.memcheck %n.vec = and i32 %shr, 1073741816 %l2 = shl nuw i32 %n.vec, 2 %ind.end = getelementptr half, half* %pIn, i32 %l2 %l3 = shl nuw i32 %n.vec, 2 %ind.end48 = getelementptr half, half* %pOut, i32 %l3 %ind.end50 = sub nsw i32 %shr, %n.vec %broadcast.splatinsert55 = insertelement <8 x half> undef, half %l0, i32 0 %broadcast.splat56 = shufflevector <8 x half> %broadcast.splatinsert55, <8 x half> undef, <8 x i32> zeroinitializer br label %vector.body vector.body: ; preds = %vector.body, %vector.ph %index = phi i32 [ 0, %vector.ph ], [ %index.next, %vector.body ] %l4 = shl i32 %index, 2 %next.gep = getelementptr half, half* %pIn, i32 %l4 %l5 = shl i32 %index, 2 %l6 = bitcast half* %next.gep to <32 x half>* %wide.vec = load <32 x half>, <32 x half>* %l6, align 2 %strided.vec = shufflevector <32 x half> %wide.vec, <32 x half> undef, <8 x i32> %strided.vec52 = shufflevector <32 x half> %wide.vec, <32 x half> undef, <8 x i32> %strided.vec53 = shufflevector <32 x half> %wide.vec, <32 x half> undef, <8 x i32> %strided.vec54 = shufflevector <32 x half> %wide.vec, <32 x half> undef, <8 x i32> %l7 = fmul <8 x half> %strided.vec, %broadcast.splat56 %l8 = fmul <8 x half> %strided.vec52, %broadcast.splat56 %l9 = fmul <8 x half> %strided.vec53, %broadcast.splat56 %l10 = fmul <8 x half> %strided.vec54, %broadcast.splat56 %l11 = getelementptr inbounds half, half* %pOut, i32 %l5 %l12 = bitcast half* %l11 to <32 x half>* %l13 = shufflevector <8 x half> %l7, <8 x half> %l8, <16 x i32> %l14 = shufflevector <8 x half> %l9, <8 x half> %l10, <16 x i32> %interleaved.vec = shufflevector <16 x half> %l13, <16 x half> %l14, <32 x i32> store <32 x half> %interleaved.vec, <32 x half>* %l12, align 2 %index.next = add i32 %index, 8 %l15 = icmp eq i32 %index.next, %n.vec br i1 %l15, label %while.end, label %vector.body while.end: ; preds = %while.body, %middle.block, %entry ret void }