xref: /llvm-project/llvm/test/CodeGen/RISCV/rvv/vfma-vp-combine.ll (revision 1cb599835ccf7ee8b2d1d5a7f3107e19a26fc6f5)
1; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
2; RUN: llc -mtriple=riscv32 -mattr=+d,+zvfh,+v,+m -target-abi=ilp32d \
3; RUN:     -verify-machineinstrs < %s | FileCheck %s
4; RUN: llc -mtriple=riscv64 -mattr=+d,+zvfh,+v,+m -target-abi=lp64d \
5; RUN:     -verify-machineinstrs < %s | FileCheck %s
6
7declare <vscale x 1 x double> @llvm.vp.fma.nxv1f64(<vscale x 1 x double>, <vscale x 1 x double>, <vscale x 1 x double>, <vscale x 1 x i1>, i32)
8declare <vscale x 1 x double> @llvm.vp.fneg.nxv1f64(<vscale x 1 x double>, <vscale x 1 x i1>, i32)
9declare <vscale x 1 x double> @llvm.vp.fmul.nxv1f64(<vscale x 1 x double>, <vscale x 1 x double>, <vscale x 1 x i1>, i32)
10
11; (-N0 * -N1) + N2 --> (N0 * N1) + N2
12define <vscale x 1 x double> @test1(<vscale x 1 x double> %a, <vscale x 1 x double> %b, <vscale x 1 x double> %c, <vscale x 1 x i1> %m, i32 zeroext %evl) {
13; CHECK-LABEL: test1:
14; CHECK:       # %bb.0:
15; CHECK-NEXT:    vsetvli zero, a0, e64, m1, ta, ma
16; CHECK-NEXT:    vfmadd.vv v9, v8, v10, v0.t
17; CHECK-NEXT:    vmv.v.v v8, v9
18; CHECK-NEXT:    ret
19  %nega = call <vscale x 1 x double> @llvm.vp.fneg.nxv1f64(<vscale x 1 x double> %a, <vscale x 1 x i1> %m, i32 %evl)
20  %negb = call <vscale x 1 x double> @llvm.vp.fneg.nxv1f64(<vscale x 1 x double> %b, <vscale x 1 x i1> %m, i32 %evl)
21  %v = call <vscale x 1 x double> @llvm.vp.fma.nxv1f64(<vscale x 1 x double> %nega, <vscale x 1 x double> %negb, <vscale x 1 x double> %c, <vscale x 1 x i1> %m, i32 %evl)
22  ret <vscale x 1 x double> %v
23}
24
25; (fma x, c1, (fmul x, c2)) -> (fmul x, c1+c2)
26define <vscale x 1 x double> @test2(<vscale x 1 x double> %a, <vscale x 1 x i1> %m, i32 zeroext %evl) {
27; CHECK-LABEL: test2:
28; CHECK:       # %bb.0:
29; CHECK-NEXT:    lui a1, %hi(.LCPI1_0)
30; CHECK-NEXT:    fld fa5, %lo(.LCPI1_0)(a1)
31; CHECK-NEXT:    lui a1, %hi(.LCPI1_1)
32; CHECK-NEXT:    fld fa4, %lo(.LCPI1_1)(a1)
33; CHECK-NEXT:    vsetvli a1, zero, e64, m1, ta, ma
34; CHECK-NEXT:    vfmv.v.f v9, fa5
35; CHECK-NEXT:    vsetvli zero, a0, e64, m1, ta, ma
36; CHECK-NEXT:    vfadd.vf v9, v9, fa4, v0.t
37; CHECK-NEXT:    vfmul.vv v8, v8, v9, v0.t
38; CHECK-NEXT:    ret
39  %t = call <vscale x 1 x double> @llvm.vp.fmul.nxv1f64(<vscale x 1 x double> %a, <vscale x 1 x double> splat (double 2.0), <vscale x 1 x i1> %m, i32 %evl)
40  %v = call fast <vscale x 1 x double> @llvm.vp.fma.nxv1f64(<vscale x 1 x double> %a, <vscale x 1 x double> splat (double 4.0), <vscale x 1 x double> %t, <vscale x 1 x i1> %m, i32 %evl)
41  ret <vscale x 1 x double> %v
42}
43
44; (fma (fmul x, c1), c2, y) -> (fma x, c1*c2, y)
45define <vscale x 1 x double> @test3(<vscale x 1 x double> %a, <vscale x 1 x double> %b, <vscale x 1 x i1> %m, i32 zeroext %evl) {
46; CHECK-LABEL: test3:
47; CHECK:       # %bb.0:
48; CHECK-NEXT:    lui a1, %hi(.LCPI2_0)
49; CHECK-NEXT:    fld fa5, %lo(.LCPI2_0)(a1)
50; CHECK-NEXT:    lui a1, %hi(.LCPI2_1)
51; CHECK-NEXT:    fld fa4, %lo(.LCPI2_1)(a1)
52; CHECK-NEXT:    vsetvli a1, zero, e64, m1, ta, ma
53; CHECK-NEXT:    vfmv.v.f v10, fa5
54; CHECK-NEXT:    vsetvli zero, a0, e64, m1, ta, ma
55; CHECK-NEXT:    vfmul.vf v10, v10, fa4, v0.t
56; CHECK-NEXT:    vfmadd.vv v10, v8, v9, v0.t
57; CHECK-NEXT:    vmv.v.v v8, v10
58; CHECK-NEXT:    ret
59  %t = call <vscale x 1 x double> @llvm.vp.fmul.nxv1f64(<vscale x 1 x double> %a, <vscale x 1 x double> splat (double 2.0), <vscale x 1 x i1> %m, i32 %evl)
60  %v = call fast <vscale x 1 x double> @llvm.vp.fma.nxv1f64(<vscale x 1 x double> %t, <vscale x 1 x double> splat (double 4.0), <vscale x 1 x double> %b, <vscale x 1 x i1> %m, i32 %evl)
61  ret <vscale x 1 x double> %v
62}
63