xref: /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/ARM/ARMInstrMVE.td (revision 82d56013d7b633d116a93943de88e08335357a7c)
1//===-- ARMInstrMVE.td - MVE support for ARM ---------------*- tablegen -*-===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This file describes the ARM MVE instruction set.
10//
11//===----------------------------------------------------------------------===//
12
13// VPT condition mask
14def vpt_mask : Operand<i32> {
15  let PrintMethod = "printVPTMask";
16  let ParserMatchClass = it_mask_asmoperand;
17  let EncoderMethod = "getVPTMaskOpValue";
18  let DecoderMethod = "DecodeVPTMaskOperand";
19}
20
21// VPT/VCMP restricted predicate for sign invariant types
22def pred_restricted_i_asmoperand : AsmOperandClass {
23  let Name = "CondCodeRestrictedI";
24  let RenderMethod = "addITCondCodeOperands";
25  let PredicateMethod = "isITCondCodeRestrictedI";
26  let ParserMethod = "parseITCondCode";
27  let DiagnosticString = "condition code for sign-independent integer "#
28                         "comparison must be EQ or NE";
29}
30
31// VPT/VCMP restricted predicate for signed types
32def pred_restricted_s_asmoperand : AsmOperandClass {
33  let Name = "CondCodeRestrictedS";
34  let RenderMethod = "addITCondCodeOperands";
35  let PredicateMethod = "isITCondCodeRestrictedS";
36  let ParserMethod = "parseITCondCode";
37  let DiagnosticString = "condition code for signed integer "#
38                         "comparison must be EQ, NE, LT, GT, LE or GE";
39}
40
41// VPT/VCMP restricted predicate for unsigned types
42def pred_restricted_u_asmoperand : AsmOperandClass {
43  let Name = "CondCodeRestrictedU";
44  let RenderMethod = "addITCondCodeOperands";
45  let PredicateMethod = "isITCondCodeRestrictedU";
46  let ParserMethod = "parseITCondCode";
47  let DiagnosticString = "condition code for unsigned integer "#
48                         "comparison must be EQ, NE, HS or HI";
49}
50
51// VPT/VCMP restricted predicate for floating point
52def pred_restricted_fp_asmoperand : AsmOperandClass {
53  let Name = "CondCodeRestrictedFP";
54  let RenderMethod = "addITCondCodeOperands";
55  let PredicateMethod = "isITCondCodeRestrictedFP";
56  let ParserMethod = "parseITCondCode";
57  let DiagnosticString = "condition code for floating-point "#
58                         "comparison must be EQ, NE, LT, GT, LE or GE";
59}
60
61class VCMPPredicateOperand : Operand<i32>;
62
63def pred_basic_i : VCMPPredicateOperand {
64  let PrintMethod = "printMandatoryRestrictedPredicateOperand";
65  let ParserMatchClass = pred_restricted_i_asmoperand;
66  let DecoderMethod = "DecodeRestrictedIPredicateOperand";
67  let EncoderMethod = "getRestrictedCondCodeOpValue";
68}
69
70def pred_basic_u : VCMPPredicateOperand {
71  let PrintMethod = "printMandatoryRestrictedPredicateOperand";
72  let ParserMatchClass = pred_restricted_u_asmoperand;
73  let DecoderMethod = "DecodeRestrictedUPredicateOperand";
74  let EncoderMethod = "getRestrictedCondCodeOpValue";
75}
76
77def pred_basic_s : VCMPPredicateOperand {
78  let PrintMethod = "printMandatoryRestrictedPredicateOperand";
79  let ParserMatchClass = pred_restricted_s_asmoperand;
80  let DecoderMethod = "DecodeRestrictedSPredicateOperand";
81  let EncoderMethod = "getRestrictedCondCodeOpValue";
82}
83
84def pred_basic_fp : VCMPPredicateOperand {
85  let PrintMethod = "printMandatoryRestrictedPredicateOperand";
86  let ParserMatchClass = pred_restricted_fp_asmoperand;
87  let DecoderMethod = "DecodeRestrictedFPPredicateOperand";
88  let EncoderMethod = "getRestrictedCondCodeOpValue";
89}
90
91// Register list operands for interleaving load/stores
92def VecList2QAsmOperand : AsmOperandClass {
93  let Name = "VecListTwoMQ";
94  let ParserMethod = "parseVectorList";
95  let RenderMethod = "addMVEVecListOperands";
96  let DiagnosticString = "operand must be a list of two consecutive "#
97                         "q-registers in range [q0,q7]";
98}
99
100def VecList2Q : RegisterOperand<QQPR, "printMVEVectorListTwoQ"> {
101  let ParserMatchClass = VecList2QAsmOperand;
102  let PrintMethod = "printMVEVectorList<2>";
103}
104
105def VecList4QAsmOperand : AsmOperandClass {
106  let Name = "VecListFourMQ";
107  let ParserMethod = "parseVectorList";
108  let RenderMethod = "addMVEVecListOperands";
109  let DiagnosticString = "operand must be a list of four consecutive "#
110                         "q-registers in range [q0,q7]";
111}
112
113def VecList4Q : RegisterOperand<QQQQPR, "printMVEVectorListFourQ"> {
114  let ParserMatchClass = VecList4QAsmOperand;
115  let PrintMethod = "printMVEVectorList<4>";
116}
117
118// taddrmode_imm7  := reg[r0-r7] +/- (imm7 << shift)
119class TMemImm7ShiftOffsetAsmOperand<int shift> : AsmOperandClass {
120  let Name = "TMemImm7Shift"#shift#"Offset";
121  let PredicateMethod = "isMemImm7ShiftedOffset<"#shift#",ARM::tGPRRegClassID>";
122  let RenderMethod = "addMemImmOffsetOperands";
123}
124
125class taddrmode_imm7<int shift> : MemOperand,
126    ComplexPattern<i32, 2, "SelectTAddrModeImm7<"#shift#">", []>  {
127  let ParserMatchClass = TMemImm7ShiftOffsetAsmOperand<shift>;
128  // They are printed the same way as the T2 imm8 version
129  let PrintMethod = "printT2AddrModeImm8Operand<false>";
130  // This can also be the same as the T2 version.
131  let EncoderMethod = "getT2AddrModeImmOpValue<7,"#shift#">";
132  let DecoderMethod = "DecodeTAddrModeImm7<"#shift#">";
133  let MIOperandInfo = (ops tGPR:$base, i32imm:$offsimm);
134}
135
136// t2addrmode_imm7  := reg +/- (imm7)
137class MemImm7ShiftOffsetAsmOperand<int shift> : AsmOperandClass {
138  let Name = "MemImm7Shift"#shift#"Offset";
139  let PredicateMethod = "isMemImm7ShiftedOffset<" # shift #
140                        ",ARM::GPRnopcRegClassID>";
141  let RenderMethod = "addMemImmOffsetOperands";
142}
143
144def MemImm7Shift0OffsetAsmOperand : MemImm7ShiftOffsetAsmOperand<0>;
145def MemImm7Shift1OffsetAsmOperand : MemImm7ShiftOffsetAsmOperand<1>;
146def MemImm7Shift2OffsetAsmOperand : MemImm7ShiftOffsetAsmOperand<2>;
147class T2AddrMode_Imm7<int shift> : MemOperand,
148      ComplexPattern<i32, 2, "SelectT2AddrModeImm7<"#shift#">", []> {
149  let EncoderMethod = "getT2AddrModeImmOpValue<7,"#shift#">";
150  let DecoderMethod = "DecodeT2AddrModeImm7<"#shift#", 0>";
151  let ParserMatchClass =
152    !cast<AsmOperandClass>("MemImm7Shift"#shift#"OffsetAsmOperand");
153  let MIOperandInfo = (ops GPRnopc:$base, i32imm:$offsimm);
154}
155
156class t2addrmode_imm7<int shift> : T2AddrMode_Imm7<shift> {
157  // They are printed the same way as the imm8 version
158  let PrintMethod = "printT2AddrModeImm8Operand<false>";
159}
160
161class MemImm7ShiftOffsetWBAsmOperand<int shift> : AsmOperandClass {
162  let Name = "MemImm7Shift"#shift#"OffsetWB";
163  let PredicateMethod = "isMemImm7ShiftedOffset<" # shift #
164                        ",ARM::rGPRRegClassID>";
165  let RenderMethod = "addMemImmOffsetOperands";
166}
167
168def MemImm7Shift0OffsetWBAsmOperand : MemImm7ShiftOffsetWBAsmOperand<0>;
169def MemImm7Shift1OffsetWBAsmOperand : MemImm7ShiftOffsetWBAsmOperand<1>;
170def MemImm7Shift2OffsetWBAsmOperand : MemImm7ShiftOffsetWBAsmOperand<2>;
171
172class t2addrmode_imm7_pre<int shift> : T2AddrMode_Imm7<shift> {
173  // They are printed the same way as the imm8 version
174  let PrintMethod = "printT2AddrModeImm8Operand<true>";
175  let ParserMatchClass =
176    !cast<AsmOperandClass>("MemImm7Shift"#shift#"OffsetWBAsmOperand");
177  let DecoderMethod = "DecodeT2AddrModeImm7<"#shift#", 1>";
178  let MIOperandInfo = (ops rGPR:$base, i32imm:$offsim);
179}
180
181class t2am_imm7shiftOffsetAsmOperand<int shift>
182  : AsmOperandClass { let Name = "Imm7Shift"#shift; }
183def t2am_imm7shift0OffsetAsmOperand : t2am_imm7shiftOffsetAsmOperand<0>;
184def t2am_imm7shift1OffsetAsmOperand : t2am_imm7shiftOffsetAsmOperand<1>;
185def t2am_imm7shift2OffsetAsmOperand : t2am_imm7shiftOffsetAsmOperand<2>;
186
187class t2am_imm7_offset<int shift> : MemOperand,
188      ComplexPattern<i32, 1, "SelectT2AddrModeImm7Offset<"#shift#">",
189                     [], [SDNPWantRoot]> {
190  // They are printed the same way as the imm8 version
191  let PrintMethod = "printT2AddrModeImm8OffsetOperand";
192  let ParserMatchClass =
193    !cast<AsmOperandClass>("t2am_imm7shift"#shift#"OffsetAsmOperand");
194  let EncoderMethod = "getT2ScaledImmOpValue<7,"#shift#">";
195  let DecoderMethod = "DecodeT2Imm7<"#shift#">";
196}
197
198// Operands for gather/scatter loads of the form [Rbase, Qoffsets]
199class MemRegRQOffsetAsmOperand<int shift> : AsmOperandClass {
200  let Name = "MemRegRQS"#shift#"Offset";
201  let PredicateMethod = "isMemRegRQOffset<"#shift#">";
202  let RenderMethod = "addMemRegRQOffsetOperands";
203}
204
205def MemRegRQS0OffsetAsmOperand : MemRegRQOffsetAsmOperand<0>;
206def MemRegRQS1OffsetAsmOperand : MemRegRQOffsetAsmOperand<1>;
207def MemRegRQS2OffsetAsmOperand : MemRegRQOffsetAsmOperand<2>;
208def MemRegRQS3OffsetAsmOperand : MemRegRQOffsetAsmOperand<3>;
209
210// mve_addr_rq_shift  := reg + vreg{ << UXTW #shift}
211class mve_addr_rq_shift<int shift> : MemOperand {
212  let EncoderMethod = "getMveAddrModeRQOpValue";
213  let PrintMethod = "printMveAddrModeRQOperand<"#shift#">";
214  let ParserMatchClass =
215    !cast<AsmOperandClass>("MemRegRQS"#shift#"OffsetAsmOperand");
216  let DecoderMethod = "DecodeMveAddrModeRQ";
217  let MIOperandInfo = (ops GPRnopc:$base, MQPR:$offsreg);
218}
219
220class MemRegQOffsetAsmOperand<int shift> : AsmOperandClass {
221  let Name = "MemRegQS"#shift#"Offset";
222  let PredicateMethod = "isMemRegQOffset<"#shift#">";
223  let RenderMethod = "addMemImmOffsetOperands";
224}
225
226def MemRegQS2OffsetAsmOperand : MemRegQOffsetAsmOperand<2>;
227def MemRegQS3OffsetAsmOperand : MemRegQOffsetAsmOperand<3>;
228
229// mve_addr_q_shift  := vreg {+ #imm7s2/4}
230class mve_addr_q_shift<int shift> : MemOperand {
231  let EncoderMethod = "getMveAddrModeQOpValue<"#shift#">";
232  // Can be printed same way as other reg + imm operands
233  let PrintMethod = "printT2AddrModeImm8Operand<false>";
234  let ParserMatchClass =
235    !cast<AsmOperandClass>("MemRegQS"#shift#"OffsetAsmOperand");
236  let DecoderMethod = "DecodeMveAddrModeQ<"#shift#">";
237  let MIOperandInfo = (ops MQPR:$base, i32imm:$imm);
238}
239
240// A family of classes wrapping up information about the vector types
241// used by MVE.
242class MVEVectorVTInfo<ValueType vec, ValueType dblvec,
243                      ValueType pred, ValueType dblpred,
244                      bits<2> size, string suffixletter, bit unsigned> {
245  // The LLVM ValueType representing the vector, so we can use it in
246  // ISel patterns.
247  ValueType Vec = vec;
248
249  // The LLVM ValueType representing a vector with elements double the size
250  // of those in Vec, so we can use it in ISel patterns. It is up to the
251  // invoker of this class to ensure that this is a correct choice.
252  ValueType DblVec = dblvec;
253
254  // An LLVM ValueType representing a corresponding vector of
255  // predicate bits, for use in ISel patterns that handle an IR
256  // intrinsic describing the predicated form of the instruction.
257  //
258  // Usually, for a vector of N things, this will be vNi1. But for
259  // vectors of 2 values, we make an exception, and use v4i1 instead
260  // of v2i1. Rationale: MVE codegen doesn't support doing all the
261  // auxiliary operations on v2i1 (vector shuffles etc), and also,
262  // there's no MVE compare instruction that will _generate_ v2i1
263  // directly.
264  ValueType Pred = pred;
265
266  // Same as Pred but for DblVec rather than Vec.
267  ValueType DblPred = dblpred;
268
269  // The most common representation of the vector element size in MVE
270  // instruction encodings: a 2-bit value V representing an (8<<V)-bit
271  // vector element.
272  bits<2> Size = size;
273
274  // For vectors explicitly mentioning a signedness of integers: 0 for
275  // signed and 1 for unsigned. For anything else, undefined.
276  bit Unsigned = unsigned;
277
278  // The number of bits in a vector element, in integer form.
279  int LaneBits = !shl(8, Size);
280
281  // The suffix used in assembly language on an instruction operating
282  // on this lane if it only cares about number of bits.
283  string BitsSuffix = !if(!eq(suffixletter, "p"),
284                          !if(!eq(unsigned, 0b0), "8", "16"),
285                          !cast<string>(LaneBits));
286
287  // The suffix used on an instruction that mentions the whole type.
288  string Suffix = suffixletter # BitsSuffix;
289
290  // The letter part of the suffix only.
291  string SuffixLetter = suffixletter;
292}
293
294// Integer vector types that don't treat signed and unsigned differently.
295def MVE_v16i8 : MVEVectorVTInfo<v16i8, v8i16, v16i1, v8i1, 0b00, "i", ?>;
296def MVE_v8i16 : MVEVectorVTInfo<v8i16, v4i32, v8i1,  v4i1, 0b01, "i", ?>;
297def MVE_v4i32 : MVEVectorVTInfo<v4i32, v2i64, v4i1,  v4i1, 0b10, "i", ?>;
298def MVE_v2i64 : MVEVectorVTInfo<v2i64, ?,     v4i1,  ?,    0b11, "i", ?>;
299
300// Explicitly signed and unsigned integer vectors. They map to the
301// same set of LLVM ValueTypes as above, but are represented
302// differently in assembly and instruction encodings.
303def MVE_v16s8 : MVEVectorVTInfo<v16i8, v8i16, v16i1, v8i1, 0b00, "s", 0b0>;
304def MVE_v8s16 : MVEVectorVTInfo<v8i16, v4i32, v8i1,  v4i1, 0b01, "s", 0b0>;
305def MVE_v4s32 : MVEVectorVTInfo<v4i32, v2i64, v4i1,  v4i1, 0b10, "s", 0b0>;
306def MVE_v2s64 : MVEVectorVTInfo<v2i64, ?,     v4i1,  ?,    0b11, "s", 0b0>;
307def MVE_v16u8 : MVEVectorVTInfo<v16i8, v8i16, v16i1, v8i1, 0b00, "u", 0b1>;
308def MVE_v8u16 : MVEVectorVTInfo<v8i16, v4i32, v8i1,  v4i1, 0b01, "u", 0b1>;
309def MVE_v4u32 : MVEVectorVTInfo<v4i32, v2i64, v4i1,  v4i1, 0b10, "u", 0b1>;
310def MVE_v2u64 : MVEVectorVTInfo<v2i64, ?,     v4i1,  ?,    0b11, "u", 0b1>;
311
312// FP vector types.
313def MVE_v8f16 : MVEVectorVTInfo<v8f16, v4f32, v8i1,  v4i1, 0b01, "f", ?>;
314def MVE_v4f32 : MVEVectorVTInfo<v4f32, v2f64, v4i1,  v4i1, 0b10, "f", ?>;
315def MVE_v2f64 : MVEVectorVTInfo<v2f64, ?,     v4i1,  ?,    0b11, "f", ?>;
316
317// Polynomial vector types.
318def MVE_v16p8 : MVEVectorVTInfo<v16i8, v8i16, v16i1, v8i1, 0b11, "p", 0b0>;
319def MVE_v8p16 : MVEVectorVTInfo<v8i16, v4i32, v8i1,  v4i1, 0b11, "p", 0b1>;
320
321multiclass MVE_TwoOpPattern<MVEVectorVTInfo VTI, SDPatternOperator Op, Intrinsic PredInt,
322                            dag PredOperands, Instruction Inst,
323                            SDPatternOperator IdentityVec = null_frag> {
324  // Unpredicated
325  def : Pat<(VTI.Vec (Op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn))),
326            (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
327
328  // Predicated with select
329  if !ne(VTI.Size, 0b11) then {
330    def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$mask),
331                                (VTI.Vec (Op (VTI.Vec MQPR:$Qm),
332                                             (VTI.Vec MQPR:$Qn))),
333                                (VTI.Vec MQPR:$inactive))),
334              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
335                              ARMVCCThen, (VTI.Pred VCCR:$mask),
336                              (VTI.Vec MQPR:$inactive)))>;
337
338    // Optionally with the select folded through the op
339    def : Pat<(VTI.Vec (Op (VTI.Vec MQPR:$Qm),
340                           (VTI.Vec (vselect (VTI.Pred VCCR:$mask),
341                                             (VTI.Vec MQPR:$Qn),
342                                             (VTI.Vec IdentityVec))))),
343              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
344                              ARMVCCThen, (VTI.Pred VCCR:$mask),
345                              (VTI.Vec MQPR:$Qm)))>;
346  }
347
348  // Predicated with intrinsic
349  def : Pat<(VTI.Vec !con((PredInt (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)),
350                          PredOperands,
351                          (? (VTI.Pred VCCR:$mask), (VTI.Vec MQPR:$inactive)))),
352            (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
353                            ARMVCCThen, (VTI.Pred VCCR:$mask),
354                            (VTI.Vec MQPR:$inactive)))>;
355}
356
357multiclass MVE_TwoOpPatternDup<MVEVectorVTInfo VTI, SDPatternOperator Op, Intrinsic PredInt,
358                               dag PredOperands, Instruction Inst,
359                               SDPatternOperator IdentityVec = null_frag> {
360  // Unpredicated
361  def : Pat<(VTI.Vec (Op (VTI.Vec MQPR:$Qm), (VTI.Vec (ARMvdup rGPR:$Rn)))),
362            (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), rGPR:$Rn))>;
363
364  // Predicated with select
365  if !ne(VTI.Size, 0b11) then {
366    def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$mask),
367                                (VTI.Vec (Op (VTI.Vec MQPR:$Qm),
368                                             (VTI.Vec (ARMvdup rGPR:$Rn)))),
369                                (VTI.Vec MQPR:$inactive))),
370              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), rGPR:$Rn,
371                              ARMVCCThen, (VTI.Pred VCCR:$mask),
372                              (VTI.Vec MQPR:$inactive)))>;
373
374    // Optionally with the select folded through the op
375    def : Pat<(VTI.Vec (Op (VTI.Vec MQPR:$Qm),
376                           (VTI.Vec (vselect (VTI.Pred VCCR:$mask),
377                                             (ARMvdup rGPR:$Rn),
378                                             (VTI.Vec IdentityVec))))),
379              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), rGPR:$Rn,
380                              ARMVCCThen, (VTI.Pred VCCR:$mask),
381                              (VTI.Vec MQPR:$Qm)))>;
382  }
383
384  // Predicated with intrinsic
385  def : Pat<(VTI.Vec !con((PredInt (VTI.Vec MQPR:$Qm), (VTI.Vec (ARMvdup rGPR:$Rn))),
386                          PredOperands,
387                          (? (VTI.Pred VCCR:$mask), (VTI.Vec MQPR:$inactive)))),
388            (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), rGPR:$Rn,
389                            ARMVCCThen, (VTI.Pred VCCR:$mask),
390                            (VTI.Vec MQPR:$inactive)))>;
391}
392
393// --------- Start of base classes for the instructions themselves
394
395class MVE_MI<dag oops, dag iops, InstrItinClass itin, string asm,
396             string ops, string cstr, list<dag> pattern>
397  : Thumb2XI<oops, iops, AddrModeNone, 4, itin, !strconcat(asm, "\t", ops), cstr,
398             pattern>,
399    Requires<[HasMVEInt]> {
400  let D = MVEDomain;
401  let DecoderNamespace = "MVE";
402}
403
404// MVE_p is used for most predicated instructions, to add the cluster
405// of input operands that provides the VPT suffix (none, T or E) and
406// the input predicate register.
407class MVE_p<dag oops, dag iops, InstrItinClass itin, string iname,
408            string suffix, string ops, vpred_ops vpred, string cstr,
409            list<dag> pattern=[]>
410  : MVE_MI<oops, !con(iops, (ins vpred:$vp)), itin,
411           // If the instruction has a suffix, like vadd.f32, then the
412           // VPT predication suffix goes before the dot, so the full
413           // name has to be "vadd${vp}.f32".
414           !strconcat(iname, "${vp}",
415                      !if(!eq(suffix, ""), "", !strconcat(".", suffix))),
416           ops, !strconcat(cstr, vpred.vpred_constraint), pattern> {
417  let Inst{31-29} = 0b111;
418  let Inst{27-26} = 0b11;
419}
420
421class MVE_f<dag oops, dag iops, InstrItinClass itin, string iname,
422            string suffix, string ops, vpred_ops vpred, string cstr,
423            list<dag> pattern=[]>
424  : MVE_p<oops, iops, itin, iname, suffix, ops, vpred, cstr, pattern> {
425  let Predicates = [HasMVEFloat];
426}
427
428class MVE_MI_with_pred<dag oops, dag iops, InstrItinClass itin, string asm,
429                       string ops, string cstr, list<dag> pattern>
430  : Thumb2I<oops, iops, AddrModeNone, 4, itin, asm, !strconcat("\t", ops), cstr,
431             pattern>,
432    Requires<[HasV8_1MMainline, HasMVEInt]> {
433  let D = MVEDomain;
434  let DecoderNamespace = "MVE";
435}
436
437class MVE_VMOV_lane_base<dag oops, dag iops, InstrItinClass itin, string asm,
438                         string suffix, string ops, string cstr,
439                         list<dag> pattern>
440  : Thumb2I<oops, iops, AddrModeNone, 4, itin, asm,
441            !if(!eq(suffix, ""), "", "." # suffix) # "\t" # ops,
442            cstr, pattern>,
443    Requires<[HasV8_1MMainline, HasMVEInt]> {
444  let D = MVEDomain;
445  let DecoderNamespace = "MVE";
446}
447
448class MVE_ScalarShift<string iname, dag oops, dag iops, string asm, string cstr,
449            list<dag> pattern=[]>
450  : MVE_MI_with_pred<oops, iops, NoItinerary, iname, asm, cstr, pattern> {
451  let Inst{31-20} = 0b111010100101;
452  let Inst{8} = 0b1;
453  let validForTailPredication=1;
454}
455
456class MVE_ScalarShiftSingleReg<string iname, dag iops, string asm, string cstr,
457                    list<dag> pattern=[]>
458  : MVE_ScalarShift<iname, (outs rGPR:$RdaDest), iops, asm, cstr, pattern> {
459  bits<4> RdaDest;
460
461  let Inst{19-16} = RdaDest{3-0};
462}
463
464class MVE_ScalarShiftSRegImm<string iname, bits<2> op5_4>
465  : MVE_ScalarShiftSingleReg<iname, (ins rGPR:$RdaSrc, long_shift:$imm),
466                     "$RdaSrc, $imm", "$RdaDest = $RdaSrc",
467                     [(set rGPR:$RdaDest,
468                          (i32 (!cast<Intrinsic>("int_arm_mve_" # iname)
469                                    (i32 rGPR:$RdaSrc), (i32 imm:$imm))))]> {
470  bits<5> imm;
471
472  let Inst{15} = 0b0;
473  let Inst{14-12} = imm{4-2};
474  let Inst{11-8} = 0b1111;
475  let Inst{7-6} = imm{1-0};
476  let Inst{5-4} = op5_4{1-0};
477  let Inst{3-0} = 0b1111;
478}
479
480def MVE_SQSHL : MVE_ScalarShiftSRegImm<"sqshl", 0b11>;
481def MVE_SRSHR : MVE_ScalarShiftSRegImm<"srshr", 0b10>;
482def MVE_UQSHL : MVE_ScalarShiftSRegImm<"uqshl", 0b00>;
483def MVE_URSHR : MVE_ScalarShiftSRegImm<"urshr", 0b01>;
484
485class MVE_ScalarShiftSRegReg<string iname, bits<2> op5_4>
486  : MVE_ScalarShiftSingleReg<iname, (ins rGPR:$RdaSrc, rGPR:$Rm),
487                     "$RdaSrc, $Rm", "$RdaDest = $RdaSrc",
488                     [(set rGPR:$RdaDest,
489                         (i32 (!cast<Intrinsic>("int_arm_mve_" # iname)
490                                   (i32 rGPR:$RdaSrc), (i32 rGPR:$Rm))))]> {
491  bits<4> Rm;
492
493  let Inst{15-12} = Rm{3-0};
494  let Inst{11-8} = 0b1111;
495  let Inst{7-6} = 0b00;
496  let Inst{5-4} = op5_4{1-0};
497  let Inst{3-0} = 0b1101;
498
499  let Unpredictable{8-6} = 0b111;
500}
501
502def MVE_SQRSHR : MVE_ScalarShiftSRegReg<"sqrshr", 0b10>;
503def MVE_UQRSHL : MVE_ScalarShiftSRegReg<"uqrshl", 0b00>;
504
505class MVE_ScalarShiftDoubleReg<string iname, dag iops, string asm,
506                               string cstr, list<dag> pattern=[]>
507  : MVE_ScalarShift<iname, (outs tGPREven:$RdaLo, tGPROdd:$RdaHi),
508                    iops, asm, cstr, pattern> {
509  bits<4> RdaLo;
510  bits<4> RdaHi;
511
512  let Inst{19-17} = RdaLo{3-1};
513  let Inst{11-9} = RdaHi{3-1};
514
515  let hasSideEffects = 0;
516}
517
518class MVE_ScalarShiftDRegImm<string iname, bits<2> op5_4, bit op16,
519                             list<dag> pattern=[]>
520  : MVE_ScalarShiftDoubleReg<
521      iname, (ins tGPREven:$RdaLo_src, tGPROdd:$RdaHi_src, long_shift:$imm),
522      "$RdaLo, $RdaHi, $imm", "$RdaLo = $RdaLo_src,$RdaHi = $RdaHi_src",
523      pattern> {
524  bits<5> imm;
525
526  let Inst{16} = op16;
527  let Inst{15} = 0b0;
528  let Inst{14-12} = imm{4-2};
529  let Inst{7-6} = imm{1-0};
530  let Inst{5-4} = op5_4{1-0};
531  let Inst{3-0} = 0b1111;
532}
533
534class MVE_ScalarShiftDRegRegBase<string iname, dag iops, string asm,
535                                 bit op5, bit op16, list<dag> pattern=[]>
536  : MVE_ScalarShiftDoubleReg<
537     iname, iops, asm, "@earlyclobber $RdaHi,@earlyclobber $RdaLo,"
538                       "$RdaLo = $RdaLo_src,$RdaHi = $RdaHi_src",
539     pattern> {
540  bits<4> Rm;
541
542  let Inst{16} = op16;
543  let Inst{15-12} = Rm{3-0};
544  let Inst{6} = 0b0;
545  let Inst{5} = op5;
546  let Inst{4} = 0b0;
547  let Inst{3-0} = 0b1101;
548
549  // Custom decoder method because of the following overlapping encodings:
550  // ASRL and SQRSHR
551  // LSLL and UQRSHL
552  // SQRSHRL and SQRSHR
553  // UQRSHLL and UQRSHL
554  let DecoderMethod = "DecodeMVEOverlappingLongShift";
555}
556
557class MVE_ScalarShiftDRegReg<string iname, bit op5, list<dag> pattern=[]>
558  : MVE_ScalarShiftDRegRegBase<
559     iname, (ins tGPREven:$RdaLo_src, tGPROdd:$RdaHi_src, rGPR:$Rm),
560     "$RdaLo, $RdaHi, $Rm", op5, 0b0, pattern> {
561
562  let Inst{7} = 0b0;
563}
564
565class MVE_ScalarShiftDRegRegWithSat<string iname, bit op5, list<dag> pattern=[]>
566  : MVE_ScalarShiftDRegRegBase<
567     iname, (ins tGPREven:$RdaLo_src, tGPROdd:$RdaHi_src, rGPR:$Rm, saturateop:$sat),
568     "$RdaLo, $RdaHi, $sat, $Rm", op5, 0b1, pattern> {
569  bit sat;
570
571  let Inst{7} = sat;
572}
573
574def MVE_ASRLr   : MVE_ScalarShiftDRegReg<"asrl",    0b1,  [(set tGPREven:$RdaLo, tGPROdd:$RdaHi,
575                                        (ARMasrl tGPREven:$RdaLo_src,
576                                        tGPROdd:$RdaHi_src, rGPR:$Rm))]>;
577def MVE_ASRLi   : MVE_ScalarShiftDRegImm<"asrl",    0b10, ?, [(set tGPREven:$RdaLo, tGPROdd:$RdaHi,
578                                        (ARMasrl tGPREven:$RdaLo_src,
579                                        tGPROdd:$RdaHi_src, (i32 long_shift:$imm)))]>;
580def MVE_LSLLr   : MVE_ScalarShiftDRegReg<"lsll",    0b0,  [(set tGPREven:$RdaLo, tGPROdd:$RdaHi,
581                                        (ARMlsll tGPREven:$RdaLo_src,
582                                        tGPROdd:$RdaHi_src, rGPR:$Rm))]>;
583def MVE_LSLLi   : MVE_ScalarShiftDRegImm<"lsll",    0b00, ?, [(set tGPREven:$RdaLo, tGPROdd:$RdaHi,
584                                        (ARMlsll tGPREven:$RdaLo_src,
585                                        tGPROdd:$RdaHi_src, (i32 long_shift:$imm)))]>;
586def MVE_LSRL    : MVE_ScalarShiftDRegImm<"lsrl",    0b01, ?, [(set tGPREven:$RdaLo, tGPROdd:$RdaHi,
587                                        (ARMlsrl tGPREven:$RdaLo_src,
588                                        tGPROdd:$RdaHi_src, (i32 long_shift:$imm)))]>;
589
590def MVE_SQRSHRL : MVE_ScalarShiftDRegRegWithSat<"sqrshrl", 0b1>;
591def MVE_SQSHLL  : MVE_ScalarShiftDRegImm<"sqshll",  0b11, 0b1>;
592def MVE_SRSHRL  : MVE_ScalarShiftDRegImm<"srshrl",  0b10, 0b1>;
593
594def MVE_UQRSHLL : MVE_ScalarShiftDRegRegWithSat<"uqrshll", 0b0>;
595def MVE_UQSHLL  : MVE_ScalarShiftDRegImm<"uqshll",  0b00, 0b1>;
596def MVE_URSHRL  : MVE_ScalarShiftDRegImm<"urshrl",  0b01, 0b1>;
597
598// start of mve_rDest instructions
599
600class MVE_rDest<dag oops, dag iops, InstrItinClass itin,
601                string iname, string suffix,
602                string ops, string cstr, list<dag> pattern=[]>
603// Always use vpred_n and not vpred_r: with the output register being
604// a GPR and not a vector register, there can't be any question of
605// what to put in its inactive lanes.
606  : MVE_p<oops, iops, itin, iname, suffix, ops, vpred_n, cstr, pattern> {
607
608  let Inst{25-23} = 0b101;
609  let Inst{11-9} = 0b111;
610  let Inst{4} = 0b0;
611}
612
613class MVE_VABAV<string suffix, bit U, bits<2> size>
614  : MVE_rDest<(outs rGPR:$Rda), (ins rGPR:$Rda_src, MQPR:$Qn, MQPR:$Qm),
615              NoItinerary, "vabav", suffix, "$Rda, $Qn, $Qm", "$Rda = $Rda_src",
616              []> {
617  bits<4> Qm;
618  bits<4> Qn;
619  bits<4> Rda;
620
621  let Inst{28} = U;
622  let Inst{22} = 0b0;
623  let Inst{21-20} = size{1-0};
624  let Inst{19-17} = Qn{2-0};
625  let Inst{16} = 0b0;
626  let Inst{15-12} = Rda{3-0};
627  let Inst{8} = 0b1;
628  let Inst{7} = Qn{3};
629  let Inst{6} = 0b0;
630  let Inst{5} = Qm{3};
631  let Inst{3-1} = Qm{2-0};
632  let Inst{0} = 0b1;
633  let horizontalReduction = 1;
634}
635
636multiclass MVE_VABAV_m<MVEVectorVTInfo VTI> {
637  def "" : MVE_VABAV<VTI.Suffix, VTI.Unsigned, VTI.Size>;
638  defvar Inst = !cast<Instruction>(NAME);
639
640  let Predicates = [HasMVEInt] in {
641    def : Pat<(i32 (int_arm_mve_vabav
642                         (i32 VTI.Unsigned),
643                         (i32 rGPR:$Rda_src),
644                         (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
645              (i32 (Inst (i32 rGPR:$Rda_src),
646                         (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm)))>;
647
648    def : Pat<(i32 (int_arm_mve_vabav_predicated
649                         (i32 VTI.Unsigned),
650                         (i32 rGPR:$Rda_src),
651                         (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
652                         (VTI.Pred VCCR:$mask))),
653              (i32 (Inst (i32 rGPR:$Rda_src),
654                         (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
655                         ARMVCCThen, (VTI.Pred VCCR:$mask)))>;
656  }
657}
658
659defm MVE_VABAVs8  : MVE_VABAV_m<MVE_v16s8>;
660defm MVE_VABAVs16 : MVE_VABAV_m<MVE_v8s16>;
661defm MVE_VABAVs32 : MVE_VABAV_m<MVE_v4s32>;
662defm MVE_VABAVu8  : MVE_VABAV_m<MVE_v16u8>;
663defm MVE_VABAVu16 : MVE_VABAV_m<MVE_v8u16>;
664defm MVE_VABAVu32 : MVE_VABAV_m<MVE_v4u32>;
665
666class MVE_VADDV<string iname, string suffix, dag iops, string cstr,
667              bit A, bit U, bits<2> size, list<dag> pattern=[]>
668  : MVE_rDest<(outs tGPREven:$Rda), iops, NoItinerary,
669              iname, suffix, "$Rda, $Qm", cstr, pattern> {
670  bits<3> Qm;
671  bits<4> Rda;
672
673  let Inst{28} = U;
674  let Inst{22-20} = 0b111;
675  let Inst{19-18} = size{1-0};
676  let Inst{17-16} = 0b01;
677  let Inst{15-13} = Rda{3-1};
678  let Inst{12} = 0b0;
679  let Inst{8-6} = 0b100;
680  let Inst{5} = A;
681  let Inst{3-1} = Qm{2-0};
682  let Inst{0} = 0b0;
683  let horizontalReduction = 1;
684  let validForTailPredication = 1;
685}
686
687def SDTVecReduceP : SDTypeProfile<1, 2, [    // VADDLVp
688  SDTCisInt<0>, SDTCisVec<1>, SDTCisVec<2>
689]>;
690def ARMVADDVs       : SDNode<"ARMISD::VADDVs", SDTVecReduce>;
691def ARMVADDVu       : SDNode<"ARMISD::VADDVu", SDTVecReduce>;
692def ARMVADDVps      : SDNode<"ARMISD::VADDVps", SDTVecReduceP>;
693def ARMVADDVpu      : SDNode<"ARMISD::VADDVpu", SDTVecReduceP>;
694
695multiclass MVE_VADDV_A<MVEVectorVTInfo VTI> {
696  def acc    : MVE_VADDV<"vaddva", VTI.Suffix,
697                         (ins tGPREven:$Rda_src, MQPR:$Qm), "$Rda = $Rda_src",
698                         0b1, VTI.Unsigned, VTI.Size>;
699  def no_acc : MVE_VADDV<"vaddv", VTI.Suffix,
700                         (ins MQPR:$Qm), "",
701                         0b0, VTI.Unsigned, VTI.Size>;
702
703  defvar InstA = !cast<Instruction>(NAME # "acc");
704  defvar InstN = !cast<Instruction>(NAME # "no_acc");
705
706  let Predicates = [HasMVEInt] in {
707    if VTI.Unsigned then {
708      def : Pat<(i32 (vecreduce_add (VTI.Vec MQPR:$vec))),
709                (i32 (InstN $vec))>;
710      def : Pat<(i32 (vecreduce_add (VTI.Vec (vselect (VTI.Pred VCCR:$pred),
711                                                      (VTI.Vec MQPR:$vec),
712                                                      (VTI.Vec ARMimmAllZerosV))))),
713                (i32 (InstN $vec, ARMVCCThen, $pred))>;
714      def : Pat<(i32 (ARMVADDVu (VTI.Vec MQPR:$vec))),
715                (i32 (InstN $vec))>;
716      def : Pat<(i32 (ARMVADDVpu (VTI.Vec MQPR:$vec), (VTI.Pred VCCR:$pred))),
717                (i32 (InstN $vec, ARMVCCThen, $pred))>;
718      def : Pat<(i32 (add (i32 (vecreduce_add (VTI.Vec MQPR:$vec))),
719                          (i32 tGPREven:$acc))),
720                (i32 (InstA $acc, $vec))>;
721      def : Pat<(i32 (add (i32 (vecreduce_add (VTI.Vec (vselect (VTI.Pred VCCR:$pred),
722                                                                (VTI.Vec MQPR:$vec),
723                                                                (VTI.Vec ARMimmAllZerosV))))),
724                          (i32 tGPREven:$acc))),
725                (i32 (InstA $acc, $vec, ARMVCCThen, $pred))>;
726      def : Pat<(i32 (add (i32 (ARMVADDVu (VTI.Vec MQPR:$vec))),
727                          (i32 tGPREven:$acc))),
728                (i32 (InstA $acc, $vec))>;
729      def : Pat<(i32 (add (i32 (ARMVADDVpu (VTI.Vec MQPR:$vec), (VTI.Pred VCCR:$pred))),
730                          (i32 tGPREven:$acc))),
731                (i32 (InstA $acc, $vec, ARMVCCThen, $pred))>;
732    } else {
733      def : Pat<(i32 (ARMVADDVs (VTI.Vec MQPR:$vec))),
734                (i32 (InstN $vec))>;
735      def : Pat<(i32 (add (i32 (ARMVADDVs (VTI.Vec MQPR:$vec))),
736                          (i32 tGPREven:$acc))),
737                (i32 (InstA $acc, $vec))>;
738      def : Pat<(i32 (ARMVADDVps (VTI.Vec MQPR:$vec), (VTI.Pred VCCR:$pred))),
739                (i32 (InstN $vec, ARMVCCThen, $pred))>;
740      def : Pat<(i32 (add (i32 (ARMVADDVps (VTI.Vec MQPR:$vec), (VTI.Pred VCCR:$pred))),
741                          (i32 tGPREven:$acc))),
742                (i32 (InstA $acc, $vec, ARMVCCThen, $pred))>;
743    }
744
745    def : Pat<(i32 (int_arm_mve_addv_predicated (VTI.Vec MQPR:$vec),
746                                                (i32 VTI.Unsigned),
747                                                (VTI.Pred VCCR:$pred))),
748              (i32 (InstN $vec, ARMVCCThen, $pred))>;
749    def : Pat<(i32 (add (int_arm_mve_addv_predicated (VTI.Vec MQPR:$vec),
750                                                     (i32 VTI.Unsigned),
751                                                     (VTI.Pred VCCR:$pred)),
752                        (i32 tGPREven:$acc))),
753              (i32 (InstA $acc, $vec, ARMVCCThen, $pred))>;
754  }
755}
756
757defm MVE_VADDVs8  : MVE_VADDV_A<MVE_v16s8>;
758defm MVE_VADDVs16 : MVE_VADDV_A<MVE_v8s16>;
759defm MVE_VADDVs32 : MVE_VADDV_A<MVE_v4s32>;
760defm MVE_VADDVu8  : MVE_VADDV_A<MVE_v16u8>;
761defm MVE_VADDVu16 : MVE_VADDV_A<MVE_v8u16>;
762defm MVE_VADDVu32 : MVE_VADDV_A<MVE_v4u32>;
763
764class MVE_VADDLV<string iname, string suffix, dag iops, string cstr,
765               bit A, bit U, list<dag> pattern=[]>
766  : MVE_rDest<(outs tGPREven:$RdaLo, tGPROdd:$RdaHi), iops, NoItinerary, iname,
767              suffix, "$RdaLo, $RdaHi, $Qm", cstr, pattern> {
768  bits<3> Qm;
769  bits<4> RdaLo;
770  bits<4> RdaHi;
771
772  let Inst{28} = U;
773  let Inst{22-20} = RdaHi{3-1};
774  let Inst{19-18} = 0b10;
775  let Inst{17-16} = 0b01;
776  let Inst{15-13} = RdaLo{3-1};
777  let Inst{12} = 0b0;
778  let Inst{8-6} = 0b100;
779  let Inst{5} = A;
780  let Inst{3-1} = Qm{2-0};
781  let Inst{0} = 0b0;
782  let horizontalReduction = 1;
783}
784
785def SDTVecReduceL : SDTypeProfile<2, 1, [    // VADDLV
786  SDTCisInt<0>, SDTCisInt<1>, SDTCisVec<2>
787]>;
788def SDTVecReduceLA : SDTypeProfile<2, 3, [    // VADDLVA
789  SDTCisInt<0>, SDTCisInt<1>, SDTCisInt<2>, SDTCisInt<3>,
790  SDTCisVec<4>
791]>;
792def SDTVecReduceLP : SDTypeProfile<2, 2, [    // VADDLVp
793  SDTCisInt<0>, SDTCisInt<1>, SDTCisVec<2>, SDTCisVec<2>
794]>;
795def SDTVecReduceLPA : SDTypeProfile<2, 4, [    // VADDLVAp
796  SDTCisInt<0>, SDTCisInt<1>, SDTCisInt<2>, SDTCisInt<3>,
797  SDTCisVec<4>, SDTCisVec<5>
798]>;
799
800multiclass MVE_VADDLV_A<MVEVectorVTInfo VTI> {
801  def acc    : MVE_VADDLV<"vaddlva", VTI.Suffix,
802                        (ins tGPREven:$RdaLo_src, tGPROdd:$RdaHi_src, MQPR:$Qm),
803                        "$RdaLo = $RdaLo_src,$RdaHi = $RdaHi_src",
804                        0b1, VTI.Unsigned>;
805  def no_acc : MVE_VADDLV<"vaddlv", VTI.Suffix,
806                        (ins MQPR:$Qm), "",
807                        0b0, VTI.Unsigned>;
808
809  defvar InstA = !cast<Instruction>(NAME # "acc");
810  defvar InstN = !cast<Instruction>(NAME # "no_acc");
811
812  defvar letter = VTI.SuffixLetter;
813  defvar ARMVADDLV = SDNode<"ARMISD::VADDLV" # letter, SDTVecReduceL>;
814  defvar ARMVADDLVA = SDNode<"ARMISD::VADDLVA" # letter, SDTVecReduceLA>;
815  defvar ARMVADDLVp = SDNode<"ARMISD::VADDLVp" # letter, SDTVecReduceLP>;
816  defvar ARMVADDLVAp = SDNode<"ARMISD::VADDLVAp" # letter, SDTVecReduceLPA>;
817
818  let Predicates = [HasMVEInt] in {
819    def : Pat<(ARMVADDLV (v4i32 MQPR:$vec)),
820              (InstN (v4i32 MQPR:$vec))>;
821    def : Pat<(ARMVADDLVA tGPREven:$acclo, tGPROdd:$acchi, (v4i32 MQPR:$vec)),
822              (InstA tGPREven:$acclo, tGPROdd:$acchi, (v4i32 MQPR:$vec))>;
823    def : Pat<(ARMVADDLVp (v4i32 MQPR:$vec), (VTI.Pred VCCR:$pred)),
824              (InstN (v4i32 MQPR:$vec), ARMVCCThen, (VTI.Pred VCCR:$pred))>;
825    def : Pat<(ARMVADDLVAp tGPREven:$acclo, tGPROdd:$acchi, (v4i32 MQPR:$vec),
826                           (VTI.Pred VCCR:$pred)),
827              (InstA tGPREven:$acclo, tGPROdd:$acchi, (v4i32 MQPR:$vec),
828                     ARMVCCThen, (VTI.Pred VCCR:$pred))>;
829  }
830}
831
832defm MVE_VADDLVs32 : MVE_VADDLV_A<MVE_v4s32>;
833defm MVE_VADDLVu32 : MVE_VADDLV_A<MVE_v4u32>;
834
835class MVE_VMINMAXNMV<string iname, string suffix, bit sz,
836                     bit bit_17, bit bit_7, list<dag> pattern=[]>
837  : MVE_rDest<(outs rGPR:$RdaDest), (ins rGPR:$RdaSrc, MQPR:$Qm),
838              NoItinerary, iname, suffix, "$RdaSrc, $Qm",
839              "$RdaDest = $RdaSrc", pattern> {
840  bits<3> Qm;
841  bits<4> RdaDest;
842
843  let Inst{28} = sz;
844  let Inst{22-20} = 0b110;
845  let Inst{19-18} = 0b11;
846  let Inst{17} = bit_17;
847  let Inst{16} = 0b0;
848  let Inst{15-12} = RdaDest{3-0};
849  let Inst{8} = 0b1;
850  let Inst{7} = bit_7;
851  let Inst{6-5} = 0b00;
852  let Inst{3-1} = Qm{2-0};
853  let Inst{0} = 0b0;
854  let horizontalReduction = 1;
855
856  let Predicates = [HasMVEFloat];
857  let hasSideEffects = 0;
858}
859
860multiclass MVE_VMINMAXNMV_p<string iname, bit notAbs, bit isMin,
861                            MVEVectorVTInfo VTI, string intrBaseName,
862                            ValueType Scalar, RegisterClass ScalarReg> {
863  def "": MVE_VMINMAXNMV<iname, VTI.Suffix, VTI.Size{0}, notAbs, isMin>;
864  defvar Inst        = !cast<Instruction>(NAME);
865  defvar unpred_intr = !cast<Intrinsic>(intrBaseName);
866  defvar pred_intr   = !cast<Intrinsic>(intrBaseName#"_predicated");
867
868  let Predicates = [HasMVEFloat] in {
869    def : Pat<(Scalar (unpred_intr (Scalar ScalarReg:$prev),
870                                   (VTI.Vec MQPR:$vec))),
871           (COPY_TO_REGCLASS (Inst (COPY_TO_REGCLASS ScalarReg:$prev, rGPR),
872                                   (VTI.Vec MQPR:$vec)),
873                              ScalarReg)>;
874    def : Pat<(Scalar (pred_intr   (Scalar ScalarReg:$prev),
875                                   (VTI.Vec MQPR:$vec),
876                                   (VTI.Pred VCCR:$pred))),
877           (COPY_TO_REGCLASS (Inst (COPY_TO_REGCLASS ScalarReg:$prev, rGPR),
878                                   (VTI.Vec MQPR:$vec),
879                                   ARMVCCThen, (VTI.Pred VCCR:$pred)),
880                              ScalarReg)>;
881  }
882}
883
884multiclass MVE_VMINMAXNMV_fty<string iname, bit notAbs, bit isMin,
885                              string intrBase> {
886  defm f32 : MVE_VMINMAXNMV_p<iname, notAbs, isMin, MVE_v4f32, intrBase,
887                              f32, SPR>;
888  defm f16 : MVE_VMINMAXNMV_p<iname, notAbs, isMin, MVE_v8f16, intrBase,
889                              f16, HPR>;
890}
891
892defm MVE_VMINNMV : MVE_VMINMAXNMV_fty<"vminnmv",  1, 1, "int_arm_mve_minnmv">;
893defm MVE_VMAXNMV : MVE_VMINMAXNMV_fty<"vmaxnmv",  1, 0, "int_arm_mve_maxnmv">;
894defm MVE_VMINNMAV: MVE_VMINMAXNMV_fty<"vminnmav", 0, 1, "int_arm_mve_minnmav">;
895defm MVE_VMAXNMAV: MVE_VMINMAXNMV_fty<"vmaxnmav", 0, 0, "int_arm_mve_maxnmav">;
896
897class MVE_VMINMAXV<string iname, string suffix, bit U, bits<2> size,
898                 bit bit_17, bit bit_7, list<dag> pattern=[]>
899  : MVE_rDest<(outs rGPR:$RdaDest), (ins rGPR:$RdaSrc, MQPR:$Qm), NoItinerary,
900              iname, suffix, "$RdaSrc, $Qm", "$RdaDest = $RdaSrc", pattern> {
901  bits<3> Qm;
902  bits<4> RdaDest;
903
904  let Inst{28} = U;
905  let Inst{22-20} = 0b110;
906  let Inst{19-18} = size{1-0};
907  let Inst{17} = bit_17;
908  let Inst{16} = 0b0;
909  let Inst{15-12} = RdaDest{3-0};
910  let Inst{8} = 0b1;
911  let Inst{7} = bit_7;
912  let Inst{6-5} = 0b00;
913  let Inst{3-1} = Qm{2-0};
914  let Inst{0} = 0b0;
915  let horizontalReduction = 1;
916}
917
918multiclass MVE_VMINMAXV_p<string iname, bit notAbs, bit isMin,
919                          MVEVectorVTInfo VTI, string intrBaseName> {
920  def "": MVE_VMINMAXV<iname, VTI.Suffix, VTI.Unsigned, VTI.Size,
921                       notAbs, isMin>;
922  defvar Inst        = !cast<Instruction>(NAME);
923  defvar unpred_intr = !cast<Intrinsic>(intrBaseName);
924  defvar pred_intr   = !cast<Intrinsic>(intrBaseName#"_predicated");
925  defvar base_args   = (? (i32 rGPR:$prev), (VTI.Vec MQPR:$vec));
926  defvar args        = !if(notAbs, !con(base_args, (? (i32 VTI.Unsigned))),
927                           base_args);
928
929  let Predicates = [HasMVEInt] in {
930    def : Pat<(i32 !con(args, (unpred_intr))),
931              (i32 (Inst (i32 rGPR:$prev), (VTI.Vec MQPR:$vec)))>;
932    def : Pat<(i32 !con(args, (pred_intr (VTI.Pred VCCR:$pred)))),
933              (i32 (Inst (i32 rGPR:$prev), (VTI.Vec MQPR:$vec),
934                         ARMVCCThen, (VTI.Pred VCCR:$pred)))>;
935  }
936}
937
938multiclass MVE_VMINMAXV_ty<string iname, bit isMin, string intrBaseName> {
939  defm s8 : MVE_VMINMAXV_p<iname, 1, isMin, MVE_v16s8, intrBaseName>;
940  defm s16: MVE_VMINMAXV_p<iname, 1, isMin, MVE_v8s16, intrBaseName>;
941  defm s32: MVE_VMINMAXV_p<iname, 1, isMin, MVE_v4s32, intrBaseName>;
942  defm u8 : MVE_VMINMAXV_p<iname, 1, isMin, MVE_v16u8, intrBaseName>;
943  defm u16: MVE_VMINMAXV_p<iname, 1, isMin, MVE_v8u16, intrBaseName>;
944  defm u32: MVE_VMINMAXV_p<iname, 1, isMin, MVE_v4u32, intrBaseName>;
945}
946
947def SDTVecReduceR : SDTypeProfile<1, 2, [   // Reduction of an integer and vector into an integer
948  SDTCisInt<0>, SDTCisInt<1>, SDTCisVec<2>
949]>;
950def ARMVMINVu       : SDNode<"ARMISD::VMINVu", SDTVecReduceR>;
951def ARMVMINVs       : SDNode<"ARMISD::VMINVs", SDTVecReduceR>;
952def ARMVMAXVu       : SDNode<"ARMISD::VMAXVu", SDTVecReduceR>;
953def ARMVMAXVs       : SDNode<"ARMISD::VMAXVs", SDTVecReduceR>;
954
955defm MVE_VMINV : MVE_VMINMAXV_ty<"vminv", 1, "int_arm_mve_minv">;
956defm MVE_VMAXV : MVE_VMINMAXV_ty<"vmaxv", 0, "int_arm_mve_maxv">;
957
958let Predicates = [HasMVEInt] in {
959  def : Pat<(i32 (vecreduce_smax (v16i8 MQPR:$src))),
960            (i32 (MVE_VMAXVs8 (t2MVNi (i32 127)), $src))>;
961  def : Pat<(i32 (vecreduce_smax (v8i16 MQPR:$src))),
962            (i32 (MVE_VMAXVs16 (t2MOVi32imm (i32 -32768)), $src))>;
963  def : Pat<(i32 (vecreduce_smax (v4i32 MQPR:$src))),
964            (i32 (MVE_VMAXVs32 (t2MOVi (i32 -2147483648)), $src))>;
965  def : Pat<(i32 (vecreduce_umax (v16i8 MQPR:$src))),
966            (i32 (MVE_VMAXVu8 (t2MOVi (i32 0)), $src))>;
967  def : Pat<(i32 (vecreduce_umax (v8i16 MQPR:$src))),
968            (i32 (MVE_VMAXVu16 (t2MOVi (i32 0)), $src))>;
969  def : Pat<(i32 (vecreduce_umax (v4i32 MQPR:$src))),
970            (i32 (MVE_VMAXVu32 (t2MOVi (i32 0)), $src))>;
971
972  def : Pat<(i32 (vecreduce_smin (v16i8 MQPR:$src))),
973            (i32 (MVE_VMINVs8 (t2MOVi (i32 127)), $src))>;
974  def : Pat<(i32 (vecreduce_smin (v8i16 MQPR:$src))),
975            (i32 (MVE_VMINVs16 (t2MOVi16 (i32 32767)), $src))>;
976  def : Pat<(i32 (vecreduce_smin (v4i32 MQPR:$src))),
977            (i32 (MVE_VMINVs32 (t2MVNi (i32 -2147483648)), $src))>;
978  def : Pat<(i32 (vecreduce_umin (v16i8 MQPR:$src))),
979            (i32 (MVE_VMINVu8 (t2MOVi (i32 255)), $src))>;
980  def : Pat<(i32 (vecreduce_umin (v8i16 MQPR:$src))),
981            (i32 (MVE_VMINVu16 (t2MOVi16 (i32 65535)), $src))>;
982  def : Pat<(i32 (vecreduce_umin (v4i32 MQPR:$src))),
983            (i32 (MVE_VMINVu32 (t2MOVi (i32 4294967295)), $src))>;
984
985  def : Pat<(i32 (ARMVMINVu (i32 rGPR:$x), (v16i8 MQPR:$src))),
986            (i32 (MVE_VMINVu8 $x, $src))>;
987  def : Pat<(i32 (ARMVMINVu (i32 rGPR:$x), (v8i16 MQPR:$src))),
988            (i32 (MVE_VMINVu16 $x, $src))>;
989  def : Pat<(i32 (ARMVMINVu (i32 rGPR:$x), (v4i32 MQPR:$src))),
990            (i32 (MVE_VMINVu32 $x, $src))>;
991  def : Pat<(i32 (ARMVMINVs (i32 rGPR:$x), (v16i8 MQPR:$src))),
992            (i32 (MVE_VMINVs8 $x, $src))>;
993  def : Pat<(i32 (ARMVMINVs (i32 rGPR:$x), (v8i16 MQPR:$src))),
994            (i32 (MVE_VMINVs16 $x, $src))>;
995  def : Pat<(i32 (ARMVMINVs (i32 rGPR:$x), (v4i32 MQPR:$src))),
996            (i32 (MVE_VMINVs32 $x, $src))>;
997
998  def : Pat<(i32 (ARMVMAXVu (i32 rGPR:$x), (v16i8 MQPR:$src))),
999            (i32 (MVE_VMAXVu8 $x, $src))>;
1000  def : Pat<(i32 (ARMVMAXVu (i32 rGPR:$x), (v8i16 MQPR:$src))),
1001            (i32 (MVE_VMAXVu16 $x, $src))>;
1002  def : Pat<(i32 (ARMVMAXVu (i32 rGPR:$x), (v4i32 MQPR:$src))),
1003            (i32 (MVE_VMAXVu32 $x, $src))>;
1004  def : Pat<(i32 (ARMVMAXVs (i32 rGPR:$x), (v16i8 MQPR:$src))),
1005            (i32 (MVE_VMAXVs8 $x, $src))>;
1006  def : Pat<(i32 (ARMVMAXVs (i32 rGPR:$x), (v8i16 MQPR:$src))),
1007            (i32 (MVE_VMAXVs16 $x, $src))>;
1008  def : Pat<(i32 (ARMVMAXVs (i32 rGPR:$x), (v4i32 MQPR:$src))),
1009            (i32 (MVE_VMAXVs32 $x, $src))>;
1010
1011}
1012
1013multiclass MVE_VMINMAXAV_ty<string iname, bit isMin, string intrBaseName> {
1014  defm s8 : MVE_VMINMAXV_p<iname, 0, isMin, MVE_v16s8, intrBaseName>;
1015  defm s16: MVE_VMINMAXV_p<iname, 0, isMin, MVE_v8s16, intrBaseName>;
1016  defm s32: MVE_VMINMAXV_p<iname, 0, isMin, MVE_v4s32, intrBaseName>;
1017}
1018
1019defm MVE_VMINAV : MVE_VMINMAXAV_ty<"vminav", 1, "int_arm_mve_minav">;
1020defm MVE_VMAXAV : MVE_VMINMAXAV_ty<"vmaxav", 0, "int_arm_mve_maxav">;
1021
1022class MVE_VMLAMLSDAV<string iname, string suffix, dag iops, string cstr,
1023                   bit sz, bit bit_28, bit A, bit X, bit bit_8, bit bit_0>
1024  : MVE_rDest<(outs tGPREven:$RdaDest), iops, NoItinerary, iname, suffix,
1025              "$RdaDest, $Qn, $Qm", cstr, []> {
1026  bits<4> RdaDest;
1027  bits<3> Qm;
1028  bits<3> Qn;
1029
1030  let Inst{28} = bit_28;
1031  let Inst{22-20} = 0b111;
1032  let Inst{19-17} = Qn{2-0};
1033  let Inst{16} = sz;
1034  let Inst{15-13} = RdaDest{3-1};
1035  let Inst{12} = X;
1036  let Inst{8} = bit_8;
1037  let Inst{7-6} = 0b00;
1038  let Inst{5} = A;
1039  let Inst{3-1} = Qm{2-0};
1040  let Inst{0} = bit_0;
1041  let horizontalReduction = 1;
1042  // Allow tail predication for non-exchanging versions. As this is also a
1043  // horizontalReduction, ARMLowOverheadLoops will also have to check that
1044  // the vector operands contain zeros in their false lanes for the instruction
1045  // to be properly valid.
1046  let validForTailPredication = !eq(X, 0);
1047}
1048
1049multiclass MVE_VMLAMLSDAV_A<string iname, string x, MVEVectorVTInfo VTI,
1050                            bit sz, bit bit_28, bit X, bit bit_8, bit bit_0> {
1051  def ""#x#VTI.Suffix : MVE_VMLAMLSDAV<iname # x, VTI.Suffix,
1052                                   (ins MQPR:$Qn, MQPR:$Qm), "",
1053                                   sz, bit_28, 0b0, X, bit_8, bit_0>;
1054  def "a"#x#VTI.Suffix : MVE_VMLAMLSDAV<iname # "a" # x, VTI.Suffix,
1055                                    (ins tGPREven:$RdaSrc, MQPR:$Qn, MQPR:$Qm),
1056                                    "$RdaDest = $RdaSrc",
1057                                    sz, bit_28, 0b1, X, bit_8, bit_0>;
1058  let Predicates = [HasMVEInt] in {
1059    def : Pat<(i32 (int_arm_mve_vmldava
1060                            (i32 VTI.Unsigned),
1061                            (i32 bit_0) /* subtract */,
1062                            (i32 X) /* exchange */,
1063                            (i32 0) /* accumulator */,
1064                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
1065              (i32 (!cast<Instruction>(NAME # x # VTI.Suffix)
1066                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm)))>;
1067
1068    def : Pat<(i32 (int_arm_mve_vmldava_predicated
1069                            (i32 VTI.Unsigned),
1070                            (i32 bit_0) /* subtract */,
1071                            (i32 X) /* exchange */,
1072                            (i32 0) /* accumulator */,
1073                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
1074                            (VTI.Pred VCCR:$mask))),
1075              (i32 (!cast<Instruction>(NAME # x # VTI.Suffix)
1076                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
1077                             ARMVCCThen, (VTI.Pred VCCR:$mask)))>;
1078
1079    def : Pat<(i32 (int_arm_mve_vmldava
1080                            (i32 VTI.Unsigned),
1081                            (i32 bit_0) /* subtract */,
1082                            (i32 X) /* exchange */,
1083                            (i32 tGPREven:$RdaSrc),
1084                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
1085              (i32 (!cast<Instruction>(NAME # "a" # x # VTI.Suffix)
1086                            (i32 tGPREven:$RdaSrc),
1087                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm)))>;
1088
1089    def : Pat<(i32 (int_arm_mve_vmldava_predicated
1090                            (i32 VTI.Unsigned),
1091                            (i32 bit_0) /* subtract */,
1092                            (i32 X) /* exchange */,
1093                            (i32 tGPREven:$RdaSrc),
1094                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
1095                            (VTI.Pred VCCR:$mask))),
1096              (i32 (!cast<Instruction>(NAME # "a" # x # VTI.Suffix)
1097                            (i32 tGPREven:$RdaSrc),
1098                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
1099                             ARMVCCThen, (VTI.Pred VCCR:$mask)))>;
1100  }
1101}
1102
1103multiclass MVE_VMLAMLSDAV_AX<string iname, MVEVectorVTInfo VTI, bit sz,
1104                             bit bit_28, bit bit_8, bit bit_0> {
1105  defm "" : MVE_VMLAMLSDAV_A<iname, "", VTI, sz, bit_28,
1106                             0b0, bit_8, bit_0>;
1107  defm "" : MVE_VMLAMLSDAV_A<iname, "x", VTI, sz, bit_28,
1108                             0b1, bit_8, bit_0>;
1109}
1110
1111multiclass MVE_VMLADAV_multi<MVEVectorVTInfo SVTI, MVEVectorVTInfo UVTI,
1112                             bit sz, bit bit_8> {
1113  defm "" : MVE_VMLAMLSDAV_AX<"vmladav", SVTI,
1114                              sz, 0b0, bit_8, 0b0>;
1115  defm "" : MVE_VMLAMLSDAV_A<"vmladav", "", UVTI,
1116                             sz, 0b1, 0b0, bit_8, 0b0>;
1117}
1118
1119multiclass MVE_VMLSDAV_multi<MVEVectorVTInfo VTI, bit sz, bit bit_28> {
1120  defm "" : MVE_VMLAMLSDAV_AX<"vmlsdav", VTI,
1121                              sz, bit_28, 0b0, 0b1>;
1122}
1123
1124defm MVE_VMLADAV : MVE_VMLADAV_multi<MVE_v16s8, MVE_v16u8, 0b0, 0b1>;
1125defm MVE_VMLADAV : MVE_VMLADAV_multi<MVE_v8s16, MVE_v8u16, 0b0, 0b0>;
1126defm MVE_VMLADAV : MVE_VMLADAV_multi<MVE_v4s32, MVE_v4u32, 0b1, 0b0>;
1127
1128defm MVE_VMLSDAV : MVE_VMLSDAV_multi<MVE_v16s8, 0b0, 0b1>;
1129defm MVE_VMLSDAV : MVE_VMLSDAV_multi<MVE_v8s16, 0b0, 0b0>;
1130defm MVE_VMLSDAV : MVE_VMLSDAV_multi<MVE_v4s32, 0b1, 0b0>;
1131
1132def SDTVecReduce2 : SDTypeProfile<1, 2, [    // VMLAV
1133  SDTCisInt<0>, SDTCisVec<1>, SDTCisVec<2>
1134]>;
1135def SDTVecReduce2L : SDTypeProfile<2, 2, [    // VMLALV
1136  SDTCisInt<0>, SDTCisInt<1>, SDTCisVec<2>, SDTCisVec<3>
1137]>;
1138def SDTVecReduce2LA : SDTypeProfile<2, 4, [    // VMLALVA
1139  SDTCisInt<0>, SDTCisInt<1>, SDTCisInt<2>, SDTCisInt<3>,
1140  SDTCisVec<4>, SDTCisVec<5>
1141]>;
1142def SDTVecReduce2P : SDTypeProfile<1, 3, [    // VMLAV
1143  SDTCisInt<0>, SDTCisVec<1>, SDTCisVec<2>, SDTCisVec<3>
1144]>;
1145def SDTVecReduce2LP : SDTypeProfile<2, 3, [    // VMLALV
1146  SDTCisInt<0>, SDTCisInt<1>, SDTCisVec<2>, SDTCisVec<3>, SDTCisVec<4>
1147]>;
1148def SDTVecReduce2LAP : SDTypeProfile<2, 5, [    // VMLALVA
1149  SDTCisInt<0>, SDTCisInt<1>, SDTCisInt<2>, SDTCisInt<3>,
1150  SDTCisVec<4>, SDTCisVec<5>, SDTCisVec<6>
1151]>;
1152def ARMVMLAVs       : SDNode<"ARMISD::VMLAVs", SDTVecReduce2>;
1153def ARMVMLAVu       : SDNode<"ARMISD::VMLAVu", SDTVecReduce2>;
1154def ARMVMLALVs      : SDNode<"ARMISD::VMLALVs", SDTVecReduce2L>;
1155def ARMVMLALVu      : SDNode<"ARMISD::VMLALVu", SDTVecReduce2L>;
1156def ARMVMLALVAs     : SDNode<"ARMISD::VMLALVAs", SDTVecReduce2LA>;
1157def ARMVMLALVAu     : SDNode<"ARMISD::VMLALVAu", SDTVecReduce2LA>;
1158def ARMVMLAVps      : SDNode<"ARMISD::VMLAVps", SDTVecReduce2P>;
1159def ARMVMLAVpu      : SDNode<"ARMISD::VMLAVpu", SDTVecReduce2P>;
1160def ARMVMLALVps     : SDNode<"ARMISD::VMLALVps", SDTVecReduce2LP>;
1161def ARMVMLALVpu     : SDNode<"ARMISD::VMLALVpu", SDTVecReduce2LP>;
1162def ARMVMLALVAps    : SDNode<"ARMISD::VMLALVAps", SDTVecReduce2LAP>;
1163def ARMVMLALVApu    : SDNode<"ARMISD::VMLALVApu", SDTVecReduce2LAP>;
1164
1165let Predicates = [HasMVEInt] in {
1166  def : Pat<(i32 (vecreduce_add (mul (v4i32 MQPR:$src1), (v4i32 MQPR:$src2)))),
1167            (i32 (MVE_VMLADAVu32 $src1, $src2))>;
1168  def : Pat<(i32 (vecreduce_add (mul (v8i16 MQPR:$src1), (v8i16 MQPR:$src2)))),
1169            (i32 (MVE_VMLADAVu16 $src1, $src2))>;
1170  def : Pat<(i32 (ARMVMLAVs (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))),
1171            (i32 (MVE_VMLADAVs16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)))>;
1172  def : Pat<(i32 (ARMVMLAVu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))),
1173            (i32 (MVE_VMLADAVu16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)))>;
1174  def : Pat<(i32 (vecreduce_add (mul (v16i8 MQPR:$src1), (v16i8 MQPR:$src2)))),
1175            (i32 (MVE_VMLADAVu8 $src1, $src2))>;
1176  def : Pat<(i32 (ARMVMLAVs (v16i8 MQPR:$val1), (v16i8 MQPR:$val2))),
1177            (i32 (MVE_VMLADAVs8 (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)))>;
1178  def : Pat<(i32 (ARMVMLAVu (v16i8 MQPR:$val1), (v16i8 MQPR:$val2))),
1179            (i32 (MVE_VMLADAVu8 (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)))>;
1180
1181  def : Pat<(i32 (add (i32 (vecreduce_add (mul (v4i32 MQPR:$src1), (v4i32 MQPR:$src2)))),
1182                      (i32 tGPREven:$src3))),
1183            (i32 (MVE_VMLADAVau32 $src3, $src1, $src2))>;
1184  def : Pat<(i32 (add (i32 (vecreduce_add (mul (v8i16 MQPR:$src1), (v8i16 MQPR:$src2)))),
1185                      (i32 tGPREven:$src3))),
1186            (i32 (MVE_VMLADAVau16 $src3, $src1, $src2))>;
1187  def : Pat<(i32 (add (ARMVMLAVs (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)), tGPREven:$Rd)),
1188            (i32 (MVE_VMLADAVas16 tGPREven:$Rd, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)))>;
1189  def : Pat<(i32 (add (ARMVMLAVu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)), tGPREven:$Rd)),
1190            (i32 (MVE_VMLADAVau16 tGPREven:$Rd, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)))>;
1191  def : Pat<(i32 (add (i32 (vecreduce_add (mul (v16i8 MQPR:$src1), (v16i8 MQPR:$src2)))),
1192                      (i32 tGPREven:$src3))),
1193            (i32 (MVE_VMLADAVau8 $src3, $src1, $src2))>;
1194  def : Pat<(i32 (add (ARMVMLAVs (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)), tGPREven:$Rd)),
1195            (i32 (MVE_VMLADAVas8 tGPREven:$Rd, (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)))>;
1196  def : Pat<(i32 (add (ARMVMLAVu (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)), tGPREven:$Rd)),
1197            (i32 (MVE_VMLADAVau8 tGPREven:$Rd, (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)))>;
1198
1199  // Predicated
1200  def : Pat<(i32 (vecreduce_add (vselect (v4i1 VCCR:$pred),
1201                                         (mul (v4i32 MQPR:$src1), (v4i32 MQPR:$src2)),
1202                                         (v4i32 ARMimmAllZerosV)))),
1203            (i32 (MVE_VMLADAVu32 $src1, $src2, ARMVCCThen, $pred))>;
1204  def : Pat<(i32 (vecreduce_add (vselect (v8i1 VCCR:$pred),
1205                                         (mul (v8i16 MQPR:$src1), (v8i16 MQPR:$src2)),
1206                                         (v8i16 ARMimmAllZerosV)))),
1207            (i32 (MVE_VMLADAVu16 $src1, $src2, ARMVCCThen, $pred))>;
1208  def : Pat<(i32 (ARMVMLAVps (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred))),
1209            (i32 (MVE_VMLADAVs16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred))>;
1210  def : Pat<(i32 (ARMVMLAVpu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred))),
1211            (i32 (MVE_VMLADAVu16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred))>;
1212  def : Pat<(i32 (vecreduce_add (vselect (v16i1 VCCR:$pred),
1213                                         (mul (v16i8 MQPR:$src1), (v16i8 MQPR:$src2)),
1214                                         (v16i8 ARMimmAllZerosV)))),
1215            (i32 (MVE_VMLADAVu8 $src1, $src2, ARMVCCThen, $pred))>;
1216  def : Pat<(i32 (ARMVMLAVps (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), (v16i1 VCCR:$pred))),
1217            (i32 (MVE_VMLADAVs8 (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), ARMVCCThen, $pred))>;
1218  def : Pat<(i32 (ARMVMLAVpu (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), (v16i1 VCCR:$pred))),
1219            (i32 (MVE_VMLADAVu8 (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), ARMVCCThen, $pred))>;
1220
1221  def : Pat<(i32 (add (i32 (vecreduce_add (vselect (v4i1 VCCR:$pred),
1222                                                   (mul (v4i32 MQPR:$src1), (v4i32 MQPR:$src2)),
1223                                                   (v4i32 ARMimmAllZerosV)))),
1224                      (i32 tGPREven:$src3))),
1225            (i32 (MVE_VMLADAVau32 $src3, $src1, $src2, ARMVCCThen, $pred))>;
1226  def : Pat<(i32 (add (i32 (vecreduce_add (vselect (v8i1 VCCR:$pred),
1227                                                   (mul (v8i16 MQPR:$src1), (v8i16 MQPR:$src2)),
1228                                                   (v8i16 ARMimmAllZerosV)))),
1229                      (i32 tGPREven:$src3))),
1230            (i32 (MVE_VMLADAVau16 $src3, $src1, $src2, ARMVCCThen, $pred))>;
1231  def : Pat<(i32 (add (ARMVMLAVps (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)), tGPREven:$Rd)),
1232            (i32 (MVE_VMLADAVas16 tGPREven:$Rd, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred))>;
1233  def : Pat<(i32 (add (ARMVMLAVpu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)), tGPREven:$Rd)),
1234            (i32 (MVE_VMLADAVau16 tGPREven:$Rd, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred))>;
1235  def : Pat<(i32 (add (i32 (vecreduce_add (vselect (v16i1 VCCR:$pred),
1236                                                   (mul (v16i8 MQPR:$src1), (v16i8 MQPR:$src2)),
1237                                                   (v16i8 ARMimmAllZerosV)))),
1238                      (i32 tGPREven:$src3))),
1239            (i32 (MVE_VMLADAVau8 $src3, $src1, $src2, ARMVCCThen, $pred))>;
1240  def : Pat<(i32 (add (ARMVMLAVps (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), (v16i1 VCCR:$pred)), tGPREven:$Rd)),
1241            (i32 (MVE_VMLADAVas8 tGPREven:$Rd, (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), ARMVCCThen, $pred))>;
1242  def : Pat<(i32 (add (ARMVMLAVpu (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), (v16i1 VCCR:$pred)), tGPREven:$Rd)),
1243            (i32 (MVE_VMLADAVau8 tGPREven:$Rd, (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), ARMVCCThen, $pred))>;
1244}
1245
1246// vmlav aliases vmladav
1247foreach acc = ["", "a"] in {
1248  foreach suffix = ["s8", "s16", "s32", "u8", "u16", "u32"] in {
1249    def : MVEInstAlias<"vmlav"#acc#"${vp}."#suffix#"\t$RdaDest, $Qn, $Qm",
1250                       (!cast<Instruction>("MVE_VMLADAV"#acc#suffix)
1251                        tGPREven:$RdaDest, MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
1252  }
1253}
1254
1255// Base class for VMLALDAV and VMLSLDAV, VRMLALDAVH, VRMLSLDAVH
1256class MVE_VMLALDAVBase<string iname, string suffix, dag iops, string cstr,
1257                       bit sz, bit bit_28, bit A, bit X, bit bit_8, bit bit_0,
1258                       list<dag> pattern=[]>
1259  : MVE_rDest<(outs tGPREven:$RdaLoDest, tGPROdd:$RdaHiDest), iops, NoItinerary,
1260              iname, suffix, "$RdaLoDest, $RdaHiDest, $Qn, $Qm", cstr, pattern> {
1261  bits<4> RdaLoDest;
1262  bits<4> RdaHiDest;
1263  bits<3> Qm;
1264  bits<3> Qn;
1265
1266  let Inst{28} = bit_28;
1267  let Inst{22-20} = RdaHiDest{3-1};
1268  let Inst{19-17} = Qn{2-0};
1269  let Inst{16} = sz;
1270  let Inst{15-13} = RdaLoDest{3-1};
1271  let Inst{12} = X;
1272  let Inst{8} = bit_8;
1273  let Inst{7-6} = 0b00;
1274  let Inst{5} = A;
1275  let Inst{3-1} = Qm{2-0};
1276  let Inst{0} = bit_0;
1277  let horizontalReduction = 1;
1278  // Allow tail predication for non-exchanging versions. As this is also a
1279  // horizontalReduction, ARMLowOverheadLoops will also have to check that
1280  // the vector operands contain zeros in their false lanes for the instruction
1281  // to be properly valid.
1282  let validForTailPredication = !eq(X, 0);
1283
1284  let hasSideEffects = 0;
1285}
1286
1287multiclass MVE_VMLALDAVBase_A<string iname, string x, string suffix,
1288                               bit sz, bit bit_28, bit X, bit bit_8, bit bit_0,
1289                               list<dag> pattern=[]> {
1290  def ""#x#suffix : MVE_VMLALDAVBase<
1291     iname # x, suffix, (ins MQPR:$Qn, MQPR:$Qm), "",
1292     sz, bit_28, 0b0, X, bit_8, bit_0, pattern>;
1293  def "a"#x#suffix : MVE_VMLALDAVBase<
1294     iname # "a" # x, suffix,
1295     (ins tGPREven:$RdaLoSrc, tGPROdd:$RdaHiSrc, MQPR:$Qn, MQPR:$Qm),
1296     "$RdaLoDest = $RdaLoSrc,$RdaHiDest = $RdaHiSrc",
1297     sz, bit_28, 0b1, X, bit_8, bit_0, pattern>;
1298}
1299
1300
1301multiclass MVE_VMLALDAVBase_AX<string iname, string suffix, bit sz, bit bit_28,
1302                               bit bit_8, bit bit_0, list<dag> pattern=[]> {
1303  defm "" : MVE_VMLALDAVBase_A<iname, "", suffix, sz,
1304                               bit_28, 0b0, bit_8, bit_0, pattern>;
1305  defm "" : MVE_VMLALDAVBase_A<iname, "x", suffix, sz,
1306                               bit_28, 0b1, bit_8, bit_0, pattern>;
1307}
1308
1309multiclass MVE_VRMLALDAVH_multi<string suffix, list<dag> pattern=[]> {
1310  defm "" : MVE_VMLALDAVBase_AX<"vrmlaldavh", "s"#suffix,
1311                                0b0, 0b0, 0b1, 0b0, pattern>;
1312  defm "" : MVE_VMLALDAVBase_A<"vrmlaldavh", "", "u"#suffix,
1313                               0b0, 0b1, 0b0, 0b1, 0b0, pattern>;
1314}
1315
1316defm MVE_VRMLALDAVH : MVE_VRMLALDAVH_multi<"32">;
1317
1318// vrmlalvh aliases for vrmlaldavh
1319def : MVEInstAlias<"vrmlalvh${vp}.s32\t$RdaLo, $RdaHi, $Qn, $Qm",
1320                  (MVE_VRMLALDAVHs32
1321                   tGPREven:$RdaLo, tGPROdd:$RdaHi,
1322                   MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
1323def : MVEInstAlias<"vrmlalvha${vp}.s32\t$RdaLo, $RdaHi, $Qn, $Qm",
1324                  (MVE_VRMLALDAVHas32
1325                   tGPREven:$RdaLo, tGPROdd:$RdaHi,
1326                   MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
1327def : MVEInstAlias<"vrmlalvh${vp}.u32\t$RdaLo, $RdaHi, $Qn, $Qm",
1328                  (MVE_VRMLALDAVHu32
1329                   tGPREven:$RdaLo, tGPROdd:$RdaHi,
1330                   MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
1331def : MVEInstAlias<"vrmlalvha${vp}.u32\t$RdaLo, $RdaHi, $Qn, $Qm",
1332                  (MVE_VRMLALDAVHau32
1333                   tGPREven:$RdaLo, tGPROdd:$RdaHi,
1334                   MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
1335
1336multiclass MVE_VMLALDAV_multi<string suffix, bit sz, list<dag> pattern=[]> {
1337  defm "" : MVE_VMLALDAVBase_AX<"vmlaldav", "s"#suffix, sz, 0b0, 0b0, 0b0, pattern>;
1338  defm "" : MVE_VMLALDAVBase_A<"vmlaldav", "", "u"#suffix,
1339                               sz, 0b1, 0b0, 0b0, 0b0, pattern>;
1340}
1341
1342defm MVE_VMLALDAV : MVE_VMLALDAV_multi<"16", 0b0>;
1343defm MVE_VMLALDAV : MVE_VMLALDAV_multi<"32", 0b1>;
1344
1345let Predicates = [HasMVEInt] in {
1346  def : Pat<(ARMVMLALVs (v4i32 MQPR:$val1), (v4i32 MQPR:$val2)),
1347            (MVE_VMLALDAVs32 (v4i32 MQPR:$val1), (v4i32 MQPR:$val2))>;
1348  def : Pat<(ARMVMLALVu (v4i32 MQPR:$val1), (v4i32 MQPR:$val2)),
1349            (MVE_VMLALDAVu32 (v4i32 MQPR:$val1), (v4i32 MQPR:$val2))>;
1350  def : Pat<(ARMVMLALVs (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)),
1351            (MVE_VMLALDAVs16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))>;
1352  def : Pat<(ARMVMLALVu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)),
1353            (MVE_VMLALDAVu16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))>;
1354
1355  def : Pat<(ARMVMLALVAs tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2)),
1356            (MVE_VMLALDAVas32 tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2))>;
1357  def : Pat<(ARMVMLALVAu tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2)),
1358            (MVE_VMLALDAVau32 tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2))>;
1359  def : Pat<(ARMVMLALVAs tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)),
1360            (MVE_VMLALDAVas16 tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))>;
1361  def : Pat<(ARMVMLALVAu tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)),
1362            (MVE_VMLALDAVau16 tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))>;
1363
1364  // Predicated
1365  def : Pat<(ARMVMLALVps (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), (v4i1 VCCR:$pred)),
1366            (MVE_VMLALDAVs32 (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), ARMVCCThen, $pred)>;
1367  def : Pat<(ARMVMLALVpu (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), (v4i1 VCCR:$pred)),
1368            (MVE_VMLALDAVu32 (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), ARMVCCThen, $pred)>;
1369  def : Pat<(ARMVMLALVps (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)),
1370            (MVE_VMLALDAVs16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred)>;
1371  def : Pat<(ARMVMLALVpu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)),
1372            (MVE_VMLALDAVu16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred)>;
1373
1374  def : Pat<(ARMVMLALVAps tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), (v4i1 VCCR:$pred)),
1375            (MVE_VMLALDAVas32 tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), ARMVCCThen, $pred)>;
1376  def : Pat<(ARMVMLALVApu tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), (v4i1 VCCR:$pred)),
1377            (MVE_VMLALDAVau32 tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), ARMVCCThen, $pred)>;
1378  def : Pat<(ARMVMLALVAps tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)),
1379            (MVE_VMLALDAVas16 tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred)>;
1380  def : Pat<(ARMVMLALVApu tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)),
1381            (MVE_VMLALDAVau16 tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred)>;
1382}
1383
1384// vmlalv aliases vmlaldav
1385foreach acc = ["", "a"] in {
1386  foreach suffix = ["s16", "s32", "u16", "u32"] in {
1387    def : MVEInstAlias<"vmlalv" # acc # "${vp}." # suffix #
1388                          "\t$RdaLoDest, $RdaHiDest, $Qn, $Qm",
1389                       (!cast<Instruction>("MVE_VMLALDAV"#acc#suffix)
1390                       tGPREven:$RdaLoDest, tGPROdd:$RdaHiDest,
1391                       MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
1392  }
1393}
1394
1395multiclass MVE_VMLSLDAV_multi<string iname, string suffix, bit sz,
1396                              bit bit_28, list<dag> pattern=[]> {
1397  defm "" : MVE_VMLALDAVBase_AX<iname, suffix, sz, bit_28, 0b0, 0b1, pattern>;
1398}
1399
1400defm MVE_VMLSLDAV   : MVE_VMLSLDAV_multi<"vmlsldav", "s16", 0b0, 0b0>;
1401defm MVE_VMLSLDAV   : MVE_VMLSLDAV_multi<"vmlsldav", "s32", 0b1, 0b0>;
1402defm MVE_VRMLSLDAVH : MVE_VMLSLDAV_multi<"vrmlsldavh", "s32", 0b0, 0b1>;
1403
1404// end of mve_rDest instructions
1405
1406// start of mve_comp instructions
1407
1408class MVE_comp<InstrItinClass itin, string iname, string suffix,
1409               string cstr, list<dag> pattern=[]>
1410  : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm), itin, iname, suffix,
1411           "$Qd, $Qn, $Qm", vpred_r, cstr, pattern> {
1412  bits<4> Qd;
1413  bits<4> Qn;
1414  bits<4> Qm;
1415
1416  let Inst{22} = Qd{3};
1417  let Inst{19-17} = Qn{2-0};
1418  let Inst{16} = 0b0;
1419  let Inst{15-13} = Qd{2-0};
1420  let Inst{12} = 0b0;
1421  let Inst{10-9} = 0b11;
1422  let Inst{7} = Qn{3};
1423  let Inst{5} = Qm{3};
1424  let Inst{3-1} = Qm{2-0};
1425  let Inst{0} = 0b0;
1426}
1427
1428class MVE_VMINMAXNM<string iname, string suffix, bit sz, bit bit_21,
1429                    list<dag> pattern=[]>
1430  : MVE_comp<NoItinerary, iname, suffix, "", pattern> {
1431
1432  let Inst{28} = 0b1;
1433  let Inst{25-24} = 0b11;
1434  let Inst{23} = 0b0;
1435  let Inst{21} = bit_21;
1436  let Inst{20} = sz;
1437  let Inst{11} = 0b1;
1438  let Inst{8} = 0b1;
1439  let Inst{6} = 0b1;
1440  let Inst{4} = 0b1;
1441
1442  let Predicates = [HasMVEFloat];
1443}
1444
1445multiclass MVE_VMINMAXNM_m<string iname, bit bit_4, MVEVectorVTInfo VTI, SDNode Op, Intrinsic PredInt> {
1446  def "" : MVE_VMINMAXNM<iname, VTI.Suffix, VTI.Size{0}, bit_4>;
1447
1448  let Predicates = [HasMVEFloat] in {
1449    defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? (i32 0)), !cast<Instruction>(NAME)>;
1450  }
1451}
1452
1453defm MVE_VMAXNMf32 : MVE_VMINMAXNM_m<"vmaxnm", 0b0, MVE_v4f32, fmaxnum, int_arm_mve_max_predicated>;
1454defm MVE_VMAXNMf16 : MVE_VMINMAXNM_m<"vmaxnm", 0b0, MVE_v8f16, fmaxnum, int_arm_mve_max_predicated>;
1455defm MVE_VMINNMf32 : MVE_VMINMAXNM_m<"vminnm", 0b1, MVE_v4f32, fminnum, int_arm_mve_min_predicated>;
1456defm MVE_VMINNMf16 : MVE_VMINMAXNM_m<"vminnm", 0b1, MVE_v8f16, fminnum, int_arm_mve_min_predicated>;
1457
1458
1459class MVE_VMINMAX<string iname, string suffix, bit U, bits<2> size,
1460              bit bit_4, list<dag> pattern=[]>
1461  : MVE_comp<NoItinerary, iname, suffix, "", pattern> {
1462
1463  let Inst{28} = U;
1464  let Inst{25-24} = 0b11;
1465  let Inst{23} = 0b0;
1466  let Inst{21-20} = size{1-0};
1467  let Inst{11} = 0b0;
1468  let Inst{8} = 0b0;
1469  let Inst{6} = 0b1;
1470  let Inst{4} = bit_4;
1471  let validForTailPredication = 1;
1472}
1473
1474multiclass MVE_VMINMAX_m<string iname, bit bit_4, MVEVectorVTInfo VTI,
1475                      SDNode Op, Intrinsic PredInt> {
1476  def "" : MVE_VMINMAX<iname, VTI.Suffix, VTI.Unsigned, VTI.Size, bit_4>;
1477
1478  let Predicates = [HasMVEInt] in {
1479    defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? (i32 VTI.Unsigned)), !cast<Instruction>(NAME)>;
1480  }
1481}
1482
1483multiclass MVE_VMAX<MVEVectorVTInfo VTI>
1484  : MVE_VMINMAX_m<"vmax", 0b0, VTI, !if(VTI.Unsigned, umax, smax), int_arm_mve_max_predicated>;
1485multiclass MVE_VMIN<MVEVectorVTInfo VTI>
1486  : MVE_VMINMAX_m<"vmin", 0b1, VTI, !if(VTI.Unsigned, umin, smin), int_arm_mve_min_predicated>;
1487
1488defm MVE_VMINs8   : MVE_VMIN<MVE_v16s8>;
1489defm MVE_VMINs16  : MVE_VMIN<MVE_v8s16>;
1490defm MVE_VMINs32  : MVE_VMIN<MVE_v4s32>;
1491defm MVE_VMINu8   : MVE_VMIN<MVE_v16u8>;
1492defm MVE_VMINu16  : MVE_VMIN<MVE_v8u16>;
1493defm MVE_VMINu32  : MVE_VMIN<MVE_v4u32>;
1494
1495defm MVE_VMAXs8   : MVE_VMAX<MVE_v16s8>;
1496defm MVE_VMAXs16  : MVE_VMAX<MVE_v8s16>;
1497defm MVE_VMAXs32  : MVE_VMAX<MVE_v4s32>;
1498defm MVE_VMAXu8   : MVE_VMAX<MVE_v16u8>;
1499defm MVE_VMAXu16  : MVE_VMAX<MVE_v8u16>;
1500defm MVE_VMAXu32  : MVE_VMAX<MVE_v4u32>;
1501
1502// end of mve_comp instructions
1503
1504// start of mve_bit instructions
1505
1506class MVE_bit_arith<dag oops, dag iops, string iname, string suffix,
1507                    string ops, string cstr, list<dag> pattern=[]>
1508  : MVE_p<oops, iops, NoItinerary, iname, suffix, ops, vpred_r, cstr, pattern> {
1509  bits<4> Qd;
1510  bits<4> Qm;
1511
1512  let Inst{22} = Qd{3};
1513  let Inst{15-13} = Qd{2-0};
1514  let Inst{5} = Qm{3};
1515  let Inst{3-1} = Qm{2-0};
1516}
1517
1518def MVE_VBIC : MVE_bit_arith<(outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm),
1519                             "vbic", "", "$Qd, $Qn, $Qm", ""> {
1520  bits<4> Qn;
1521
1522  let Inst{28} = 0b0;
1523  let Inst{25-23} = 0b110;
1524  let Inst{21-20} = 0b01;
1525  let Inst{19-17} = Qn{2-0};
1526  let Inst{16} = 0b0;
1527  let Inst{12-8} = 0b00001;
1528  let Inst{7} = Qn{3};
1529  let Inst{6} = 0b1;
1530  let Inst{4} = 0b1;
1531  let Inst{0} = 0b0;
1532  let validForTailPredication = 1;
1533}
1534
1535class MVE_VREV<string iname, string suffix, bits<2> size, bits<2> bit_8_7, string cstr="">
1536  : MVE_bit_arith<(outs MQPR:$Qd), (ins MQPR:$Qm), iname,
1537                  suffix, "$Qd, $Qm", cstr> {
1538
1539  let Inst{28} = 0b1;
1540  let Inst{25-23} = 0b111;
1541  let Inst{21-20} = 0b11;
1542  let Inst{19-18} = size;
1543  let Inst{17-16} = 0b00;
1544  let Inst{12-9} = 0b0000;
1545  let Inst{8-7} = bit_8_7;
1546  let Inst{6} = 0b1;
1547  let Inst{4} = 0b0;
1548  let Inst{0} = 0b0;
1549}
1550
1551def MVE_VREV64_8  : MVE_VREV<"vrev64", "8", 0b00, 0b00, "@earlyclobber $Qd">;
1552def MVE_VREV64_16 : MVE_VREV<"vrev64", "16", 0b01, 0b00, "@earlyclobber $Qd">;
1553def MVE_VREV64_32 : MVE_VREV<"vrev64", "32", 0b10, 0b00, "@earlyclobber $Qd">;
1554
1555def MVE_VREV32_8  : MVE_VREV<"vrev32", "8", 0b00, 0b01>;
1556def MVE_VREV32_16 : MVE_VREV<"vrev32", "16", 0b01, 0b01>;
1557
1558def MVE_VREV16_8  : MVE_VREV<"vrev16", "8", 0b00, 0b10>;
1559
1560let Predicates = [HasMVEInt] in {
1561  def : Pat<(v8i16 (bswap (v8i16 MQPR:$src))),
1562            (v8i16 (MVE_VREV16_8 (v8i16 MQPR:$src)))>;
1563  def : Pat<(v4i32 (bswap (v4i32 MQPR:$src))),
1564            (v4i32 (MVE_VREV32_8 (v4i32 MQPR:$src)))>;
1565}
1566
1567multiclass MVE_VREV_basic_patterns<int revbits, list<MVEVectorVTInfo> VTIs,
1568                                   Instruction Inst> {
1569  defvar unpred_op = !cast<SDNode>("ARMvrev" # revbits);
1570
1571  foreach VTI = VTIs in {
1572    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$src))),
1573              (VTI.Vec (Inst (VTI.Vec MQPR:$src)))>;
1574    def : Pat<(VTI.Vec (int_arm_mve_vrev_predicated (VTI.Vec MQPR:$src),
1575                  revbits, (VTI.Pred VCCR:$pred), (VTI.Vec MQPR:$inactive))),
1576              (VTI.Vec (Inst (VTI.Vec MQPR:$src), ARMVCCThen,
1577                  (VTI.Pred VCCR:$pred), (VTI.Vec MQPR:$inactive)))>;
1578  }
1579}
1580
1581let Predicates = [HasMVEInt] in {
1582  defm: MVE_VREV_basic_patterns<64, [MVE_v4i32, MVE_v4f32], MVE_VREV64_32>;
1583  defm: MVE_VREV_basic_patterns<64, [MVE_v8i16, MVE_v8f16], MVE_VREV64_16>;
1584  defm: MVE_VREV_basic_patterns<64, [MVE_v16i8           ], MVE_VREV64_8>;
1585
1586  defm: MVE_VREV_basic_patterns<32, [MVE_v8i16, MVE_v8f16], MVE_VREV32_16>;
1587  defm: MVE_VREV_basic_patterns<32, [MVE_v16i8           ], MVE_VREV32_8>;
1588
1589  defm: MVE_VREV_basic_patterns<16, [MVE_v16i8           ], MVE_VREV16_8>;
1590}
1591
1592def MVE_VMVN : MVE_bit_arith<(outs MQPR:$Qd), (ins MQPR:$Qm),
1593                             "vmvn", "", "$Qd, $Qm", ""> {
1594  let Inst{28} = 0b1;
1595  let Inst{25-23} = 0b111;
1596  let Inst{21-16} = 0b110000;
1597  let Inst{12-6} = 0b0010111;
1598  let Inst{4} = 0b0;
1599  let Inst{0} = 0b0;
1600  let validForTailPredication = 1;
1601}
1602
1603let Predicates = [HasMVEInt] in {
1604  foreach VTI = [ MVE_v16i8, MVE_v8i16, MVE_v4i32, MVE_v2i64 ] in {
1605    def : Pat<(VTI.Vec (vnotq    (VTI.Vec MQPR:$val1))),
1606              (VTI.Vec (MVE_VMVN (VTI.Vec MQPR:$val1)))>;
1607    def : Pat<(VTI.Vec (int_arm_mve_mvn_predicated (VTI.Vec MQPR:$val1),
1608                       (VTI.Pred VCCR:$pred), (VTI.Vec MQPR:$inactive))),
1609              (VTI.Vec (MVE_VMVN (VTI.Vec MQPR:$val1), ARMVCCThen,
1610                       (VTI.Pred VCCR:$pred), (VTI.Vec MQPR:$inactive)))>;
1611  }
1612}
1613
1614class MVE_bit_ops<string iname, bits<2> bit_21_20, bit bit_28>
1615  : MVE_bit_arith<(outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm),
1616                  iname, "", "$Qd, $Qn, $Qm", ""> {
1617  bits<4> Qn;
1618
1619  let Inst{28} = bit_28;
1620  let Inst{25-23} = 0b110;
1621  let Inst{21-20} = bit_21_20;
1622  let Inst{19-17} = Qn{2-0};
1623  let Inst{16} = 0b0;
1624  let Inst{12-8} = 0b00001;
1625  let Inst{7} = Qn{3};
1626  let Inst{6} = 0b1;
1627  let Inst{4} = 0b1;
1628  let Inst{0} = 0b0;
1629  let validForTailPredication = 1;
1630}
1631
1632def MVE_VEOR : MVE_bit_ops<"veor", 0b00, 0b1>;
1633def MVE_VORN : MVE_bit_ops<"vorn", 0b11, 0b0>;
1634def MVE_VORR : MVE_bit_ops<"vorr", 0b10, 0b0>;
1635def MVE_VAND : MVE_bit_ops<"vand", 0b00, 0b0>;
1636
1637// add ignored suffixes as aliases
1638
1639foreach s=["s8", "s16", "s32", "u8", "u16", "u32", "i8", "i16", "i32", "f16", "f32"] in {
1640  def : MVEInstAlias<"vbic${vp}." # s # "\t$QdSrc, $QnSrc, $QmSrc",
1641        (MVE_VBIC MQPR:$QdSrc, MQPR:$QnSrc, MQPR:$QmSrc, vpred_r:$vp)>;
1642  def : MVEInstAlias<"veor${vp}." # s # "\t$QdSrc, $QnSrc, $QmSrc",
1643        (MVE_VEOR MQPR:$QdSrc, MQPR:$QnSrc, MQPR:$QmSrc, vpred_r:$vp)>;
1644  def : MVEInstAlias<"vorn${vp}." # s # "\t$QdSrc, $QnSrc, $QmSrc",
1645        (MVE_VORN MQPR:$QdSrc, MQPR:$QnSrc, MQPR:$QmSrc, vpred_r:$vp)>;
1646  def : MVEInstAlias<"vorr${vp}." # s # "\t$QdSrc, $QnSrc, $QmSrc",
1647        (MVE_VORR MQPR:$QdSrc, MQPR:$QnSrc, MQPR:$QmSrc, vpred_r:$vp)>;
1648  def : MVEInstAlias<"vand${vp}." # s # "\t$QdSrc, $QnSrc, $QmSrc",
1649        (MVE_VAND MQPR:$QdSrc, MQPR:$QnSrc, MQPR:$QmSrc, vpred_r:$vp)>;
1650}
1651
1652let Predicates = [HasMVEInt] in {
1653  defm : MVE_TwoOpPattern<MVE_v16i8, and, int_arm_mve_and_predicated, (? ), MVE_VAND, ARMimmAllOnesV>;
1654  defm : MVE_TwoOpPattern<MVE_v8i16, and, int_arm_mve_and_predicated, (? ), MVE_VAND, ARMimmAllOnesV>;
1655  defm : MVE_TwoOpPattern<MVE_v4i32, and, int_arm_mve_and_predicated, (? ), MVE_VAND, ARMimmAllOnesV>;
1656  defm : MVE_TwoOpPattern<MVE_v2i64, and, int_arm_mve_and_predicated, (? ), MVE_VAND, ARMimmAllOnesV>;
1657
1658  defm : MVE_TwoOpPattern<MVE_v16i8, or, int_arm_mve_orr_predicated, (? ), MVE_VORR, ARMimmAllZerosV>;
1659  defm : MVE_TwoOpPattern<MVE_v8i16, or, int_arm_mve_orr_predicated, (? ), MVE_VORR, ARMimmAllZerosV>;
1660  defm : MVE_TwoOpPattern<MVE_v4i32, or, int_arm_mve_orr_predicated, (? ), MVE_VORR, ARMimmAllZerosV>;
1661  defm : MVE_TwoOpPattern<MVE_v2i64, or, int_arm_mve_orr_predicated, (? ), MVE_VORR, ARMimmAllZerosV>;
1662
1663  defm : MVE_TwoOpPattern<MVE_v16i8, xor, int_arm_mve_eor_predicated, (? ), MVE_VEOR, ARMimmAllZerosV>;
1664  defm : MVE_TwoOpPattern<MVE_v8i16, xor, int_arm_mve_eor_predicated, (? ), MVE_VEOR, ARMimmAllZerosV>;
1665  defm : MVE_TwoOpPattern<MVE_v4i32, xor, int_arm_mve_eor_predicated, (? ), MVE_VEOR, ARMimmAllZerosV>;
1666  defm : MVE_TwoOpPattern<MVE_v2i64, xor, int_arm_mve_eor_predicated, (? ), MVE_VEOR, ARMimmAllZerosV>;
1667
1668  defm : MVE_TwoOpPattern<MVE_v16i8, BinOpFrag<(and node:$LHS, (vnotq node:$RHS))>,
1669                          int_arm_mve_bic_predicated, (? ), MVE_VBIC>;
1670  defm : MVE_TwoOpPattern<MVE_v8i16, BinOpFrag<(and node:$LHS, (vnotq node:$RHS))>,
1671                          int_arm_mve_bic_predicated, (? ), MVE_VBIC>;
1672  defm : MVE_TwoOpPattern<MVE_v4i32, BinOpFrag<(and node:$LHS, (vnotq node:$RHS))>,
1673                          int_arm_mve_bic_predicated, (? ), MVE_VBIC>;
1674  defm : MVE_TwoOpPattern<MVE_v2i64, BinOpFrag<(and node:$LHS, (vnotq node:$RHS))>,
1675                          int_arm_mve_bic_predicated, (? ), MVE_VBIC>;
1676
1677  defm : MVE_TwoOpPattern<MVE_v16i8, BinOpFrag<(or node:$LHS, (vnotq node:$RHS))>,
1678                          int_arm_mve_orn_predicated, (? ), MVE_VORN>;
1679  defm : MVE_TwoOpPattern<MVE_v8i16, BinOpFrag<(or node:$LHS, (vnotq node:$RHS))>,
1680                          int_arm_mve_orn_predicated, (? ), MVE_VORN>;
1681  defm : MVE_TwoOpPattern<MVE_v4i32, BinOpFrag<(or node:$LHS, (vnotq node:$RHS))>,
1682                          int_arm_mve_orn_predicated, (? ), MVE_VORN>;
1683  defm : MVE_TwoOpPattern<MVE_v2i64, BinOpFrag<(or node:$LHS, (vnotq node:$RHS))>,
1684                          int_arm_mve_orn_predicated, (? ), MVE_VORN>;
1685}
1686
1687class MVE_bit_cmode<string iname, string suffix, bit halfword, dag inOps>
1688  : MVE_p<(outs MQPR:$Qd), inOps, NoItinerary,
1689          iname, suffix, "$Qd, $imm", vpred_n, "$Qd = $Qd_src"> {
1690  bits<12> imm;
1691  bits<4> Qd;
1692
1693  let Inst{28} = imm{7};
1694  let Inst{27-23} = 0b11111;
1695  let Inst{22} = Qd{3};
1696  let Inst{21-19} = 0b000;
1697  let Inst{18-16} = imm{6-4};
1698  let Inst{15-13} = Qd{2-0};
1699  let Inst{12} = 0b0;
1700  let Inst{11} = halfword;
1701  let Inst{10} = !if(halfword, 0, imm{10});
1702  let Inst{9} = imm{9};
1703  let Inst{8} = 0b1;
1704  let Inst{7-6} = 0b01;
1705  let Inst{4} = 0b1;
1706  let Inst{3-0} = imm{3-0};
1707}
1708
1709multiclass MVE_bit_cmode_p<string iname, bit opcode,
1710                           MVEVectorVTInfo VTI, Operand imm_type, SDNode op> {
1711  def "" : MVE_bit_cmode<iname, VTI.Suffix, VTI.Size{0},
1712                         (ins MQPR:$Qd_src, imm_type:$imm)> {
1713    let Inst{5} = opcode;
1714    let validForTailPredication = 1;
1715  }
1716
1717  defvar Inst = !cast<Instruction>(NAME);
1718  defvar UnpredPat = (VTI.Vec (op (VTI.Vec MQPR:$src), timm:$simm));
1719
1720  let Predicates = [HasMVEInt] in {
1721    def : Pat<UnpredPat,
1722              (VTI.Vec (Inst (VTI.Vec MQPR:$src), imm_type:$simm))>;
1723    def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
1724                          UnpredPat, (VTI.Vec MQPR:$src))),
1725              (VTI.Vec (Inst (VTI.Vec MQPR:$src), imm_type:$simm,
1726                             ARMVCCThen, (VTI.Pred VCCR:$pred)))>;
1727  }
1728}
1729
1730multiclass MVE_VORRimm<MVEVectorVTInfo VTI, Operand imm_type> {
1731  defm "": MVE_bit_cmode_p<"vorr", 0, VTI, imm_type, ARMvorrImm>;
1732}
1733multiclass MVE_VBICimm<MVEVectorVTInfo VTI, Operand imm_type> {
1734  defm "": MVE_bit_cmode_p<"vbic", 1, VTI, imm_type, ARMvbicImm>;
1735}
1736
1737defm MVE_VORRimmi16 : MVE_VORRimm<MVE_v8i16, nImmSplatI16>;
1738defm MVE_VORRimmi32 : MVE_VORRimm<MVE_v4i32, nImmSplatI32>;
1739defm MVE_VBICimmi16 : MVE_VBICimm<MVE_v8i16, nImmSplatI16>;
1740defm MVE_VBICimmi32 : MVE_VBICimm<MVE_v4i32, nImmSplatI32>;
1741
1742def MVE_VORNimmi16 : MVEInstAlias<"vorn${vp}.i16\t$Qd, $imm",
1743    (MVE_VORRimmi16 MQPR:$Qd, nImmSplatNotI16:$imm, vpred_n:$vp), 0>;
1744def MVE_VORNimmi32 : MVEInstAlias<"vorn${vp}.i32\t$Qd, $imm",
1745    (MVE_VORRimmi32 MQPR:$Qd, nImmSplatNotI32:$imm, vpred_n:$vp), 0>;
1746
1747def MVE_VANDimmi16 : MVEInstAlias<"vand${vp}.i16\t$Qd, $imm",
1748    (MVE_VBICimmi16 MQPR:$Qd, nImmSplatNotI16:$imm, vpred_n:$vp), 0>;
1749def MVE_VANDimmi32 : MVEInstAlias<"vand${vp}.i32\t$Qd, $imm",
1750    (MVE_VBICimmi32 MQPR:$Qd, nImmSplatNotI32:$imm, vpred_n:$vp), 0>;
1751
1752def MVE_VMOV : MVEInstAlias<"vmov${vp}\t$Qd, $Qm",
1753    (MVE_VORR MQPR:$Qd, MQPR:$Qm, MQPR:$Qm, vpred_r:$vp)>;
1754
1755class MVE_VMOV_lane_direction {
1756  bit bit_20;
1757  dag oops;
1758  dag iops;
1759  string ops;
1760  string cstr;
1761}
1762def MVE_VMOV_from_lane : MVE_VMOV_lane_direction {
1763  let bit_20 = 0b1;
1764  let oops = (outs rGPR:$Rt);
1765  let iops = (ins MQPR:$Qd);
1766  let ops = "$Rt, $Qd$Idx";
1767  let cstr = "";
1768}
1769def MVE_VMOV_to_lane : MVE_VMOV_lane_direction {
1770  let bit_20 = 0b0;
1771  let oops = (outs MQPR:$Qd);
1772  let iops = (ins MQPR:$Qd_src, rGPR:$Rt);
1773  let ops = "$Qd$Idx, $Rt";
1774  let cstr = "$Qd = $Qd_src";
1775}
1776
1777class MVE_VMOV_lane<string suffix, bit U, dag indexop,
1778                    MVE_VMOV_lane_direction dir>
1779  : MVE_VMOV_lane_base<dir.oops, !con(dir.iops, indexop), NoItinerary,
1780                       "vmov", suffix, dir.ops, dir.cstr, []> {
1781  bits<4> Qd;
1782  bits<4> Rt;
1783
1784  let Inst{31-24} = 0b11101110;
1785  let Inst{23} = U;
1786  let Inst{20} = dir.bit_20;
1787  let Inst{19-17} = Qd{2-0};
1788  let Inst{15-12} = Rt{3-0};
1789  let Inst{11-8} = 0b1011;
1790  let Inst{7} = Qd{3};
1791  let Inst{4-0} = 0b10000;
1792
1793  let hasSideEffects = 0;
1794}
1795
1796class MVE_VMOV_lane_32<MVE_VMOV_lane_direction dir>
1797    : MVE_VMOV_lane<"32", 0b0, (ins MVEVectorIndex<4>:$Idx), dir> {
1798  bits<2> Idx;
1799  let Inst{22} = 0b0;
1800  let Inst{6-5} = 0b00;
1801  let Inst{16} = Idx{1};
1802  let Inst{21} = Idx{0};
1803
1804  let Predicates = [HasFPRegsV8_1M];
1805}
1806
1807class MVE_VMOV_lane_16<string suffix, bit U, MVE_VMOV_lane_direction dir>
1808  : MVE_VMOV_lane<suffix, U, (ins MVEVectorIndex<8>:$Idx), dir> {
1809  bits<3> Idx;
1810  let Inst{22} = 0b0;
1811  let Inst{5} = 0b1;
1812  let Inst{16} = Idx{2};
1813  let Inst{21} = Idx{1};
1814  let Inst{6} = Idx{0};
1815}
1816
1817class MVE_VMOV_lane_8<string suffix, bit U, MVE_VMOV_lane_direction dir>
1818  : MVE_VMOV_lane<suffix, U, (ins MVEVectorIndex<16>:$Idx), dir> {
1819  bits<4> Idx;
1820  let Inst{22} = 0b1;
1821  let Inst{16} = Idx{3};
1822  let Inst{21} = Idx{2};
1823  let Inst{6} = Idx{1};
1824  let Inst{5} = Idx{0};
1825}
1826
1827def MVE_VMOV_from_lane_32  : MVE_VMOV_lane_32<            MVE_VMOV_from_lane>;
1828def MVE_VMOV_from_lane_s16 : MVE_VMOV_lane_16<"s16", 0b0, MVE_VMOV_from_lane>;
1829def MVE_VMOV_from_lane_u16 : MVE_VMOV_lane_16<"u16", 0b1, MVE_VMOV_from_lane>;
1830def MVE_VMOV_from_lane_s8  : MVE_VMOV_lane_8 < "s8", 0b0, MVE_VMOV_from_lane>;
1831def MVE_VMOV_from_lane_u8  : MVE_VMOV_lane_8 < "u8", 0b1, MVE_VMOV_from_lane>;
1832let isInsertSubreg = 1 in
1833def MVE_VMOV_to_lane_32    : MVE_VMOV_lane_32<            MVE_VMOV_to_lane>;
1834def MVE_VMOV_to_lane_16    : MVE_VMOV_lane_16< "16", 0b0, MVE_VMOV_to_lane>;
1835def MVE_VMOV_to_lane_8     : MVE_VMOV_lane_8 <  "8", 0b0, MVE_VMOV_to_lane>;
1836
1837// This is the same as insertelt but allows the inserted value to be an i32 as
1838// will be used when it is the only legal type.
1839def ARMVecInsert : SDTypeProfile<1, 3, [
1840  SDTCisVT<2, i32>, SDTCisSameAs<0, 1>, SDTCisPtrTy<3>
1841]>;
1842def ARMinsertelt  : SDNode<"ISD::INSERT_VECTOR_ELT", ARMVecInsert>;
1843
1844let Predicates = [HasMVEInt] in {
1845  def : Pat<(extractelt (v2f64 MQPR:$src), imm:$lane),
1846            (f64 (EXTRACT_SUBREG MQPR:$src, (DSubReg_f64_reg imm:$lane)))>;
1847  def : Pat<(insertelt (v2f64 MQPR:$src1), DPR:$src2, imm:$lane),
1848            (INSERT_SUBREG (v2f64 (COPY_TO_REGCLASS MQPR:$src1, MQPR)), DPR:$src2, (DSubReg_f64_reg imm:$lane))>;
1849
1850  def : Pat<(extractelt (v4i32 MQPR:$src), imm:$lane),
1851            (COPY_TO_REGCLASS
1852              (i32 (EXTRACT_SUBREG MQPR:$src, (SSubReg_f32_reg imm:$lane))), rGPR)>;
1853  def : Pat<(insertelt (v4i32 MQPR:$src1), rGPR:$src2, imm:$lane),
1854            (MVE_VMOV_to_lane_32 MQPR:$src1, rGPR:$src2, imm:$lane)>;
1855  // This tries to copy from one lane to another, without going via GPR regs
1856  def : Pat<(insertelt (v4i32 MQPR:$src1), (extractelt (v4i32 MQPR:$src2), imm:$extlane), imm:$inslane),
1857            (v4i32 (COPY_TO_REGCLASS
1858                     (INSERT_SUBREG (v4f32 (COPY_TO_REGCLASS (v4i32 MQPR:$src1), MQPR)),
1859                                    (f32 (EXTRACT_SUBREG (v4f32 (COPY_TO_REGCLASS (v4i32 MQPR:$src2), MQPR)),
1860                                                         (SSubReg_f32_reg imm:$extlane))),
1861                                    (SSubReg_f32_reg imm:$inslane)),
1862                      MQPR))>;
1863
1864  def : Pat<(vector_insert (v16i8 MQPR:$src1), rGPR:$src2, imm:$lane),
1865            (MVE_VMOV_to_lane_8  MQPR:$src1, rGPR:$src2, imm:$lane)>;
1866  def : Pat<(vector_insert (v8i16 MQPR:$src1), rGPR:$src2, imm:$lane),
1867            (MVE_VMOV_to_lane_16 MQPR:$src1, rGPR:$src2, imm:$lane)>;
1868
1869  def : Pat<(ARMvgetlanes (v16i8 MQPR:$src), imm:$lane),
1870            (MVE_VMOV_from_lane_s8 MQPR:$src, imm:$lane)>;
1871  def : Pat<(ARMvgetlanes (v8i16 MQPR:$src), imm:$lane),
1872            (MVE_VMOV_from_lane_s16 MQPR:$src, imm:$lane)>;
1873  def : Pat<(ARMvgetlanes (v8f16 MQPR:$src), imm:$lane),
1874            (MVE_VMOV_from_lane_s16 MQPR:$src, imm:$lane)>;
1875  def : Pat<(ARMvgetlaneu (v16i8 MQPR:$src), imm:$lane),
1876            (MVE_VMOV_from_lane_u8 MQPR:$src, imm:$lane)>;
1877  def : Pat<(ARMvgetlaneu (v8i16 MQPR:$src), imm:$lane),
1878            (MVE_VMOV_from_lane_u16 MQPR:$src, imm:$lane)>;
1879  def : Pat<(ARMvgetlaneu (v8f16 MQPR:$src), imm:$lane),
1880            (MVE_VMOV_from_lane_u16 MQPR:$src, imm:$lane)>;
1881  // For i16's inserts being extracted from low lanes, then may use VINS.
1882  def : Pat<(ARMinsertelt (v8i16 MQPR:$src1),
1883                          (ARMvgetlaneu (v8i16 MQPR:$src2), imm_even:$extlane),
1884                          imm_odd:$inslane),
1885            (COPY_TO_REGCLASS (INSERT_SUBREG (v4f32 (COPY_TO_REGCLASS MQPR:$src1, MQPR)),
1886                                (VINSH (EXTRACT_SUBREG MQPR:$src1, (SSubReg_f16_reg imm_odd:$inslane)),
1887                                       (EXTRACT_SUBREG MQPR:$src2, (SSubReg_f16_reg imm_even:$extlane))),
1888                                (SSubReg_f16_reg imm_odd:$inslane)), MQPR)>;
1889
1890  def : Pat<(v16i8 (scalar_to_vector GPR:$src)),
1891            (MVE_VMOV_to_lane_8  (v16i8 (IMPLICIT_DEF)), rGPR:$src, (i32 0))>;
1892  def : Pat<(v8i16 (scalar_to_vector GPR:$src)),
1893            (MVE_VMOV_to_lane_16 (v8i16 (IMPLICIT_DEF)), rGPR:$src, (i32 0))>;
1894  def : Pat<(v4i32 (scalar_to_vector GPR:$src)),
1895            (MVE_VMOV_to_lane_32 (v4i32 (IMPLICIT_DEF)), rGPR:$src, (i32 0))>;
1896
1897  // Floating point patterns, still enabled under HasMVEInt
1898  def : Pat<(extractelt (v4f32 MQPR:$src), imm:$lane),
1899            (COPY_TO_REGCLASS (f32 (EXTRACT_SUBREG MQPR:$src, (SSubReg_f32_reg imm:$lane))), SPR)>;
1900  def : Pat<(insertelt (v4f32 MQPR:$src1), (f32 SPR:$src2), imm:$lane),
1901            (INSERT_SUBREG (v4f32 (COPY_TO_REGCLASS MQPR:$src1, MQPR)), SPR:$src2, (SSubReg_f32_reg imm:$lane))>;
1902
1903  def : Pat<(insertelt (v8f16 MQPR:$src1), (f16 HPR:$src2), imm_even:$lane),
1904            (MVE_VMOV_to_lane_16 MQPR:$src1, (COPY_TO_REGCLASS (f16 HPR:$src2), rGPR), imm:$lane)>;
1905  def : Pat<(insertelt (v8f16 MQPR:$src1), (f16 HPR:$src2), imm_odd:$lane),
1906            (COPY_TO_REGCLASS (INSERT_SUBREG (v4f32 (COPY_TO_REGCLASS MQPR:$src1, MQPR)),
1907                                (VINSH (EXTRACT_SUBREG MQPR:$src1, (SSubReg_f16_reg imm_odd:$lane)),
1908                                       (COPY_TO_REGCLASS HPR:$src2, SPR)),
1909                                (SSubReg_f16_reg imm_odd:$lane)), MQPR)>;
1910  def : Pat<(extractelt (v8f16 MQPR:$src), imm_even:$lane),
1911            (EXTRACT_SUBREG MQPR:$src, (SSubReg_f16_reg imm_even:$lane))>;
1912  def : Pat<(extractelt (v8f16 MQPR:$src), imm_odd:$lane),
1913            (COPY_TO_REGCLASS
1914              (VMOVH (EXTRACT_SUBREG MQPR:$src, (SSubReg_f16_reg imm_odd:$lane))),
1915              HPR)>;
1916
1917  def : Pat<(v2f64 (scalar_to_vector (f64 DPR:$src))),
1918            (INSERT_SUBREG (v2f64 (IMPLICIT_DEF)), DPR:$src, dsub_0)>;
1919  def : Pat<(v4f32 (scalar_to_vector SPR:$src)),
1920            (INSERT_SUBREG (v4f32 (IMPLICIT_DEF)), SPR:$src, ssub_0)>;
1921  def : Pat<(v4f32 (scalar_to_vector GPR:$src)),
1922            (MVE_VMOV_to_lane_32 (v4f32 (IMPLICIT_DEF)), rGPR:$src, (i32 0))>;
1923  def : Pat<(v8f16 (scalar_to_vector (f16 HPR:$src))),
1924            (INSERT_SUBREG (v8f16 (IMPLICIT_DEF)), (f16 HPR:$src), ssub_0)>;
1925  def : Pat<(v8f16 (scalar_to_vector GPR:$src)),
1926            (MVE_VMOV_to_lane_16 (v8f16 (IMPLICIT_DEF)), rGPR:$src, (i32 0))>;
1927}
1928
1929// end of mve_bit instructions
1930
1931// start of MVE Integer instructions
1932
1933class MVE_int<string iname, string suffix, bits<2> size, list<dag> pattern=[]>
1934  : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm), NoItinerary,
1935          iname, suffix, "$Qd, $Qn, $Qm", vpred_r, "", pattern> {
1936  bits<4> Qd;
1937  bits<4> Qn;
1938  bits<4> Qm;
1939
1940  let Inst{22} = Qd{3};
1941  let Inst{21-20} = size;
1942  let Inst{19-17} = Qn{2-0};
1943  let Inst{15-13} = Qd{2-0};
1944  let Inst{7} = Qn{3};
1945  let Inst{6} = 0b1;
1946  let Inst{5} = Qm{3};
1947  let Inst{3-1} = Qm{2-0};
1948}
1949
1950class MVE_VMULt1<string iname, string suffix, bits<2> size,
1951                   list<dag> pattern=[]>
1952  : MVE_int<iname, suffix, size, pattern> {
1953
1954  let Inst{28} = 0b0;
1955  let Inst{25-23} = 0b110;
1956  let Inst{16} = 0b0;
1957  let Inst{12-8} = 0b01001;
1958  let Inst{4} = 0b1;
1959  let Inst{0} = 0b0;
1960  let validForTailPredication = 1;
1961}
1962
1963multiclass MVE_VMUL_m<MVEVectorVTInfo VTI> {
1964  def "" : MVE_VMULt1<"vmul", VTI.Suffix, VTI.Size>;
1965
1966  let Predicates = [HasMVEInt] in {
1967    defm : MVE_TwoOpPattern<VTI, mul, int_arm_mve_mul_predicated, (? ),
1968                            !cast<Instruction>(NAME), ARMimmOneV>;
1969  }
1970}
1971
1972defm MVE_VMULi8  : MVE_VMUL_m<MVE_v16i8>;
1973defm MVE_VMULi16 : MVE_VMUL_m<MVE_v8i16>;
1974defm MVE_VMULi32 : MVE_VMUL_m<MVE_v4i32>;
1975
1976class MVE_VQxDMULH_Base<string iname, string suffix, bits<2> size, bit rounding,
1977                  list<dag> pattern=[]>
1978  : MVE_int<iname, suffix, size, pattern> {
1979
1980  let Inst{28} = rounding;
1981  let Inst{25-23} = 0b110;
1982  let Inst{16} = 0b0;
1983  let Inst{12-8} = 0b01011;
1984  let Inst{4} = 0b0;
1985  let Inst{0} = 0b0;
1986  let validForTailPredication = 1;
1987}
1988
1989def MVEvqdmulh : SDNode<"ARMISD::VQDMULH", SDTIntBinOp>;
1990
1991multiclass MVE_VQxDMULH_m<string iname, MVEVectorVTInfo VTI,
1992                      SDNode Op, Intrinsic unpred_int, Intrinsic pred_int,
1993                      bit rounding> {
1994  def "" : MVE_VQxDMULH_Base<iname, VTI.Suffix, VTI.Size, rounding>;
1995  defvar Inst = !cast<Instruction>(NAME);
1996
1997  let Predicates = [HasMVEInt] in {
1998    defm : MVE_TwoOpPattern<VTI, Op, pred_int, (? ), Inst>;
1999
2000    // Extra unpredicated multiply intrinsic patterns
2001    def : Pat<(VTI.Vec (unpred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn))),
2002              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
2003  }
2004}
2005
2006multiclass MVE_VQxDMULH<string iname, MVEVectorVTInfo VTI, bit rounding>
2007  : MVE_VQxDMULH_m<iname, VTI, !if(rounding, null_frag,
2008                                             MVEvqdmulh),
2009                               !if(rounding, int_arm_mve_vqrdmulh,
2010                                             int_arm_mve_vqdmulh),
2011                               !if(rounding, int_arm_mve_qrdmulh_predicated,
2012                                             int_arm_mve_qdmulh_predicated),
2013                   rounding>;
2014
2015defm MVE_VQDMULHi8  : MVE_VQxDMULH<"vqdmulh", MVE_v16s8, 0b0>;
2016defm MVE_VQDMULHi16 : MVE_VQxDMULH<"vqdmulh", MVE_v8s16, 0b0>;
2017defm MVE_VQDMULHi32 : MVE_VQxDMULH<"vqdmulh", MVE_v4s32, 0b0>;
2018
2019defm MVE_VQRDMULHi8  : MVE_VQxDMULH<"vqrdmulh", MVE_v16s8, 0b1>;
2020defm MVE_VQRDMULHi16 : MVE_VQxDMULH<"vqrdmulh", MVE_v8s16, 0b1>;
2021defm MVE_VQRDMULHi32 : MVE_VQxDMULH<"vqrdmulh", MVE_v4s32, 0b1>;
2022
2023class MVE_VADDSUB<string iname, string suffix, bits<2> size, bit subtract,
2024                    list<dag> pattern=[]>
2025  : MVE_int<iname, suffix, size, pattern> {
2026
2027  let Inst{28} = subtract;
2028  let Inst{25-23} = 0b110;
2029  let Inst{16} = 0b0;
2030  let Inst{12-8} = 0b01000;
2031  let Inst{4} = 0b0;
2032  let Inst{0} = 0b0;
2033  let validForTailPredication = 1;
2034}
2035
2036multiclass MVE_VADDSUB_m<string iname, MVEVectorVTInfo VTI, bit subtract,
2037                         SDNode Op, Intrinsic PredInt> {
2038  def "" : MVE_VADDSUB<iname, VTI.Suffix, VTI.Size, subtract>;
2039  defvar Inst = !cast<Instruction>(NAME);
2040
2041  let Predicates = [HasMVEInt] in {
2042    defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? ), !cast<Instruction>(NAME), ARMimmAllZerosV>;
2043  }
2044}
2045
2046multiclass MVE_VADD<MVEVectorVTInfo VTI>
2047  : MVE_VADDSUB_m<"vadd", VTI, 0b0, add, int_arm_mve_add_predicated>;
2048multiclass MVE_VSUB<MVEVectorVTInfo VTI>
2049  : MVE_VADDSUB_m<"vsub", VTI, 0b1, sub, int_arm_mve_sub_predicated>;
2050
2051defm MVE_VADDi8  : MVE_VADD<MVE_v16i8>;
2052defm MVE_VADDi16 : MVE_VADD<MVE_v8i16>;
2053defm MVE_VADDi32 : MVE_VADD<MVE_v4i32>;
2054
2055defm MVE_VSUBi8  : MVE_VSUB<MVE_v16i8>;
2056defm MVE_VSUBi16 : MVE_VSUB<MVE_v8i16>;
2057defm MVE_VSUBi32 : MVE_VSUB<MVE_v4i32>;
2058
2059class MVE_VQADDSUB<string iname, string suffix, bit U, bit subtract,
2060                   bits<2> size>
2061  : MVE_int<iname, suffix, size, []> {
2062
2063  let Inst{28} = U;
2064  let Inst{25-23} = 0b110;
2065  let Inst{16} = 0b0;
2066  let Inst{12-10} = 0b000;
2067  let Inst{9} = subtract;
2068  let Inst{8} = 0b0;
2069  let Inst{4} = 0b1;
2070  let Inst{0} = 0b0;
2071  let validForTailPredication = 1;
2072}
2073
2074class MVE_VQADD_<string suffix, bit U, bits<2> size>
2075  : MVE_VQADDSUB<"vqadd", suffix, U, 0b0, size>;
2076class MVE_VQSUB_<string suffix, bit U, bits<2> size>
2077  : MVE_VQADDSUB<"vqsub", suffix, U, 0b1, size>;
2078
2079multiclass MVE_VQADD_m<MVEVectorVTInfo VTI,
2080                      SDNode Op, Intrinsic PredInt> {
2081  def "" : MVE_VQADD_<VTI.Suffix, VTI.Unsigned, VTI.Size>;
2082  defvar Inst = !cast<Instruction>(NAME);
2083
2084  let Predicates = [HasMVEInt] in {
2085    defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? (i32 VTI.Unsigned)),
2086                            !cast<Instruction>(NAME)>;
2087  }
2088}
2089
2090multiclass MVE_VQADD<MVEVectorVTInfo VTI, SDNode unpred_op>
2091  : MVE_VQADD_m<VTI, unpred_op, int_arm_mve_qadd_predicated>;
2092
2093defm MVE_VQADDs8  : MVE_VQADD<MVE_v16s8, saddsat>;
2094defm MVE_VQADDs16 : MVE_VQADD<MVE_v8s16, saddsat>;
2095defm MVE_VQADDs32 : MVE_VQADD<MVE_v4s32, saddsat>;
2096defm MVE_VQADDu8  : MVE_VQADD<MVE_v16u8, uaddsat>;
2097defm MVE_VQADDu16 : MVE_VQADD<MVE_v8u16, uaddsat>;
2098defm MVE_VQADDu32 : MVE_VQADD<MVE_v4u32, uaddsat>;
2099
2100multiclass MVE_VQSUB_m<MVEVectorVTInfo VTI,
2101                      SDNode Op, Intrinsic PredInt> {
2102  def "" : MVE_VQSUB_<VTI.Suffix, VTI.Unsigned, VTI.Size>;
2103  defvar Inst = !cast<Instruction>(NAME);
2104
2105  let Predicates = [HasMVEInt] in {
2106    defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? (i32 VTI.Unsigned)),
2107                            !cast<Instruction>(NAME)>;
2108  }
2109}
2110
2111multiclass MVE_VQSUB<MVEVectorVTInfo VTI, SDNode unpred_op>
2112  : MVE_VQSUB_m<VTI, unpred_op, int_arm_mve_qsub_predicated>;
2113
2114defm MVE_VQSUBs8  : MVE_VQSUB<MVE_v16s8, ssubsat>;
2115defm MVE_VQSUBs16 : MVE_VQSUB<MVE_v8s16, ssubsat>;
2116defm MVE_VQSUBs32 : MVE_VQSUB<MVE_v4s32, ssubsat>;
2117defm MVE_VQSUBu8  : MVE_VQSUB<MVE_v16u8, usubsat>;
2118defm MVE_VQSUBu16 : MVE_VQSUB<MVE_v8u16, usubsat>;
2119defm MVE_VQSUBu32 : MVE_VQSUB<MVE_v4u32, usubsat>;
2120
2121class MVE_VABD_int<string suffix, bit U, bits<2> size,
2122                     list<dag> pattern=[]>
2123  : MVE_int<"vabd", suffix, size, pattern> {
2124
2125  let Inst{28} = U;
2126  let Inst{25-23} = 0b110;
2127  let Inst{16} = 0b0;
2128  let Inst{12-8} = 0b00111;
2129  let Inst{4} = 0b0;
2130  let Inst{0} = 0b0;
2131  let validForTailPredication = 1;
2132}
2133
2134multiclass MVE_VABD_m<MVEVectorVTInfo VTI,
2135                      Intrinsic unpred_int, Intrinsic pred_int> {
2136  def "" : MVE_VABD_int<VTI.Suffix, VTI.Unsigned, VTI.Size>;
2137  defvar Inst = !cast<Instruction>(NAME);
2138
2139  let Predicates = [HasMVEInt] in {
2140    // Unpredicated absolute difference
2141    def : Pat<(VTI.Vec (unpred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
2142                            (i32 VTI.Unsigned))),
2143              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
2144
2145    // Predicated absolute difference
2146    def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
2147                            (i32 VTI.Unsigned), (VTI.Pred VCCR:$mask),
2148                            (VTI.Vec MQPR:$inactive))),
2149              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
2150                             ARMVCCThen, (VTI.Pred VCCR:$mask),
2151                             (VTI.Vec MQPR:$inactive)))>;
2152  }
2153}
2154
2155multiclass MVE_VABD<MVEVectorVTInfo VTI>
2156  : MVE_VABD_m<VTI, int_arm_mve_vabd, int_arm_mve_abd_predicated>;
2157
2158defm MVE_VABDs8  : MVE_VABD<MVE_v16s8>;
2159defm MVE_VABDs16 : MVE_VABD<MVE_v8s16>;
2160defm MVE_VABDs32 : MVE_VABD<MVE_v4s32>;
2161defm MVE_VABDu8  : MVE_VABD<MVE_v16u8>;
2162defm MVE_VABDu16 : MVE_VABD<MVE_v8u16>;
2163defm MVE_VABDu32 : MVE_VABD<MVE_v4u32>;
2164
2165class MVE_VRHADD_Base<string suffix, bit U, bits<2> size, list<dag> pattern=[]>
2166  : MVE_int<"vrhadd", suffix, size, pattern> {
2167
2168  let Inst{28} = U;
2169  let Inst{25-23} = 0b110;
2170  let Inst{16} = 0b0;
2171  let Inst{12-8} = 0b00001;
2172  let Inst{4} = 0b0;
2173  let Inst{0} = 0b0;
2174  let validForTailPredication = 1;
2175}
2176
2177def addnuw : PatFrag<(ops node:$lhs, node:$rhs),
2178                     (add node:$lhs, node:$rhs), [{
2179  return N->getFlags().hasNoUnsignedWrap();
2180}]>;
2181
2182def addnsw : PatFrag<(ops node:$lhs, node:$rhs),
2183                     (add node:$lhs, node:$rhs), [{
2184  return N->getFlags().hasNoSignedWrap();
2185}]>;
2186
2187def subnuw : PatFrag<(ops node:$lhs, node:$rhs),
2188                     (sub node:$lhs, node:$rhs), [{
2189  return N->getFlags().hasNoUnsignedWrap();
2190}]>;
2191
2192def subnsw : PatFrag<(ops node:$lhs, node:$rhs),
2193                     (sub node:$lhs, node:$rhs), [{
2194  return N->getFlags().hasNoSignedWrap();
2195}]>;
2196
2197multiclass MVE_VRHADD_m<MVEVectorVTInfo VTI,
2198                      SDNode unpred_op, Intrinsic pred_int> {
2199  def "" : MVE_VRHADD_Base<VTI.Suffix, VTI.Unsigned, VTI.Size>;
2200  defvar Inst = !cast<Instruction>(NAME);
2201
2202  let Predicates = [HasMVEInt] in {
2203    // Unpredicated rounding add-with-divide-by-two
2204    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
2205                            (i32 VTI.Unsigned))),
2206              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
2207
2208    // Predicated add-with-divide-by-two
2209    def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
2210                            (i32 VTI.Unsigned), (VTI.Pred VCCR:$mask),
2211                            (VTI.Vec MQPR:$inactive))),
2212              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
2213                             ARMVCCThen, (VTI.Pred VCCR:$mask),
2214                             (VTI.Vec MQPR:$inactive)))>;
2215  }
2216}
2217
2218multiclass MVE_VRHADD<MVEVectorVTInfo VTI>
2219  : MVE_VRHADD_m<VTI, int_arm_mve_vrhadd, int_arm_mve_rhadd_predicated>;
2220
2221defm MVE_VRHADDs8  : MVE_VRHADD<MVE_v16s8>;
2222defm MVE_VRHADDs16 : MVE_VRHADD<MVE_v8s16>;
2223defm MVE_VRHADDs32 : MVE_VRHADD<MVE_v4s32>;
2224defm MVE_VRHADDu8  : MVE_VRHADD<MVE_v16u8>;
2225defm MVE_VRHADDu16 : MVE_VRHADD<MVE_v8u16>;
2226defm MVE_VRHADDu32 : MVE_VRHADD<MVE_v4u32>;
2227
2228// Rounding Halving Add perform the arithemtic operation with an extra bit of
2229// precision, before performing the shift, to void clipping errors. We're not
2230// modelling that here with these patterns, but we're using no wrap forms of
2231// add to ensure that the extra bit of information is not needed for the
2232// arithmetic or the rounding.
2233let Predicates = [HasMVEInt] in {
2234  def : Pat<(v16i8 (ARMvshrsImm (addnsw (addnsw (v16i8 MQPR:$Qm), (v16i8 MQPR:$Qn)),
2235                                        (v16i8 (ARMvmovImm (i32 3585)))),
2236                                (i32 1))),
2237            (MVE_VRHADDs8 MQPR:$Qm, MQPR:$Qn)>;
2238  def : Pat<(v8i16 (ARMvshrsImm (addnsw (addnsw (v8i16 MQPR:$Qm), (v8i16 MQPR:$Qn)),
2239                                        (v8i16 (ARMvmovImm (i32 2049)))),
2240                                (i32 1))),
2241            (MVE_VRHADDs16 MQPR:$Qm, MQPR:$Qn)>;
2242  def : Pat<(v4i32 (ARMvshrsImm (addnsw (addnsw (v4i32 MQPR:$Qm), (v4i32 MQPR:$Qn)),
2243                                        (v4i32 (ARMvmovImm (i32 1)))),
2244                                (i32 1))),
2245            (MVE_VRHADDs32 MQPR:$Qm, MQPR:$Qn)>;
2246  def : Pat<(v16i8 (ARMvshruImm (addnuw (addnuw (v16i8 MQPR:$Qm), (v16i8 MQPR:$Qn)),
2247                                        (v16i8 (ARMvmovImm (i32 3585)))),
2248                                (i32 1))),
2249            (MVE_VRHADDu8 MQPR:$Qm, MQPR:$Qn)>;
2250  def : Pat<(v8i16 (ARMvshruImm (addnuw (addnuw (v8i16 MQPR:$Qm), (v8i16 MQPR:$Qn)),
2251                                        (v8i16 (ARMvmovImm (i32 2049)))),
2252                                (i32 1))),
2253            (MVE_VRHADDu16 MQPR:$Qm, MQPR:$Qn)>;
2254  def : Pat<(v4i32 (ARMvshruImm (addnuw (addnuw (v4i32 MQPR:$Qm), (v4i32 MQPR:$Qn)),
2255                                        (v4i32 (ARMvmovImm (i32 1)))),
2256                                (i32 1))),
2257            (MVE_VRHADDu32 MQPR:$Qm, MQPR:$Qn)>;
2258}
2259
2260
2261class MVE_VHADDSUB<string iname, string suffix, bit U, bit subtract,
2262                   bits<2> size, list<dag> pattern=[]>
2263  : MVE_int<iname, suffix, size, pattern> {
2264
2265  let Inst{28} = U;
2266  let Inst{25-23} = 0b110;
2267  let Inst{16} = 0b0;
2268  let Inst{12-10} = 0b000;
2269  let Inst{9} = subtract;
2270  let Inst{8} = 0b0;
2271  let Inst{4} = 0b0;
2272  let Inst{0} = 0b0;
2273  let validForTailPredication = 1;
2274}
2275
2276class MVE_VHADD_<string suffix, bit U, bits<2> size,
2277              list<dag> pattern=[]>
2278  : MVE_VHADDSUB<"vhadd", suffix, U, 0b0, size, pattern>;
2279class MVE_VHSUB_<string suffix, bit U, bits<2> size,
2280              list<dag> pattern=[]>
2281  : MVE_VHADDSUB<"vhsub", suffix, U, 0b1, size, pattern>;
2282
2283multiclass MVE_VHADD_m<MVEVectorVTInfo VTI,
2284                      SDNode unpred_op, Intrinsic pred_int, PatFrag add_op,
2285                      SDNode shift_op> {
2286  def "" : MVE_VHADD_<VTI.Suffix, VTI.Unsigned, VTI.Size>;
2287  defvar Inst = !cast<Instruction>(NAME);
2288
2289  let Predicates = [HasMVEInt] in {
2290    // Unpredicated add-and-divide-by-two
2291    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn), (i32 VTI.Unsigned))),
2292              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
2293
2294    def : Pat<(VTI.Vec (shift_op (add_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)), (i32 1))),
2295              (Inst MQPR:$Qm, MQPR:$Qn)>;
2296
2297    // Predicated add-and-divide-by-two
2298    def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn), (i32 VTI.Unsigned),
2299                            (VTI.Pred VCCR:$mask), (VTI.Vec MQPR:$inactive))),
2300              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
2301                             ARMVCCThen, (VTI.Pred VCCR:$mask),
2302                             (VTI.Vec MQPR:$inactive)))>;
2303  }
2304}
2305
2306multiclass MVE_VHADD<MVEVectorVTInfo VTI, PatFrag add_op, SDNode shift_op>
2307  : MVE_VHADD_m<VTI, int_arm_mve_vhadd, int_arm_mve_hadd_predicated, add_op,
2308                shift_op>;
2309
2310// Halving add/sub perform the arithemtic operation with an extra bit of
2311// precision, before performing the shift, to void clipping errors. We're not
2312// modelling that here with these patterns, but we're using no wrap forms of
2313// add/sub to ensure that the extra bit of information is not needed.
2314defm MVE_VHADDs8  : MVE_VHADD<MVE_v16s8, addnsw, ARMvshrsImm>;
2315defm MVE_VHADDs16 : MVE_VHADD<MVE_v8s16, addnsw, ARMvshrsImm>;
2316defm MVE_VHADDs32 : MVE_VHADD<MVE_v4s32, addnsw, ARMvshrsImm>;
2317defm MVE_VHADDu8  : MVE_VHADD<MVE_v16u8, addnuw, ARMvshruImm>;
2318defm MVE_VHADDu16 : MVE_VHADD<MVE_v8u16, addnuw, ARMvshruImm>;
2319defm MVE_VHADDu32 : MVE_VHADD<MVE_v4u32, addnuw, ARMvshruImm>;
2320
2321multiclass MVE_VHSUB_m<MVEVectorVTInfo VTI,
2322                      SDNode unpred_op, Intrinsic pred_int, PatFrag sub_op,
2323                      SDNode shift_op> {
2324  def "" : MVE_VHSUB_<VTI.Suffix, VTI.Unsigned, VTI.Size>;
2325  defvar Inst = !cast<Instruction>(NAME);
2326
2327  let Predicates = [HasMVEInt] in {
2328    // Unpredicated subtract-and-divide-by-two
2329    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
2330                            (i32 VTI.Unsigned))),
2331              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
2332
2333    def : Pat<(VTI.Vec (shift_op (sub_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)), (i32 1))),
2334              (Inst MQPR:$Qm, MQPR:$Qn)>;
2335
2336
2337    // Predicated subtract-and-divide-by-two
2338    def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
2339                            (i32 VTI.Unsigned), (VTI.Pred VCCR:$mask),
2340                            (VTI.Vec MQPR:$inactive))),
2341              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
2342                             ARMVCCThen, (VTI.Pred VCCR:$mask),
2343                             (VTI.Vec MQPR:$inactive)))>;
2344  }
2345}
2346
2347multiclass MVE_VHSUB<MVEVectorVTInfo VTI, PatFrag sub_op, SDNode shift_op>
2348  : MVE_VHSUB_m<VTI, int_arm_mve_vhsub, int_arm_mve_hsub_predicated, sub_op,
2349                shift_op>;
2350
2351defm MVE_VHSUBs8  : MVE_VHSUB<MVE_v16s8, subnsw, ARMvshrsImm>;
2352defm MVE_VHSUBs16 : MVE_VHSUB<MVE_v8s16, subnsw, ARMvshrsImm>;
2353defm MVE_VHSUBs32 : MVE_VHSUB<MVE_v4s32, subnsw, ARMvshrsImm>;
2354defm MVE_VHSUBu8  : MVE_VHSUB<MVE_v16u8, subnuw, ARMvshruImm>;
2355defm MVE_VHSUBu16 : MVE_VHSUB<MVE_v8u16, subnuw, ARMvshruImm>;
2356defm MVE_VHSUBu32 : MVE_VHSUB<MVE_v4u32, subnuw, ARMvshruImm>;
2357
2358class MVE_VDUP<string suffix, bit B, bit E, list<dag> pattern=[]>
2359  : MVE_p<(outs MQPR:$Qd), (ins rGPR:$Rt), NoItinerary,
2360          "vdup", suffix, "$Qd, $Rt", vpred_r, "", pattern> {
2361  bits<4> Qd;
2362  bits<4> Rt;
2363
2364  let Inst{28} = 0b0;
2365  let Inst{25-23} = 0b101;
2366  let Inst{22} = B;
2367  let Inst{21-20} = 0b10;
2368  let Inst{19-17} = Qd{2-0};
2369  let Inst{16} = 0b0;
2370  let Inst{15-12} = Rt;
2371  let Inst{11-8} = 0b1011;
2372  let Inst{7} = Qd{3};
2373  let Inst{6} = 0b0;
2374  let Inst{5} = E;
2375  let Inst{4-0} = 0b10000;
2376  let validForTailPredication = 1;
2377}
2378
2379def MVE_VDUP32 : MVE_VDUP<"32", 0b0, 0b0>;
2380def MVE_VDUP16 : MVE_VDUP<"16", 0b0, 0b1>;
2381def MVE_VDUP8  : MVE_VDUP<"8",  0b1, 0b0>;
2382
2383let Predicates = [HasMVEInt] in {
2384  def : Pat<(v16i8 (ARMvdup (i32 rGPR:$elem))),
2385            (MVE_VDUP8  rGPR:$elem)>;
2386  def : Pat<(v8i16 (ARMvdup (i32 rGPR:$elem))),
2387            (MVE_VDUP16 rGPR:$elem)>;
2388  def : Pat<(v4i32 (ARMvdup (i32 rGPR:$elem))),
2389            (MVE_VDUP32 rGPR:$elem)>;
2390
2391  def : Pat<(v8f16 (ARMvdup (i32 rGPR:$elem))),
2392            (MVE_VDUP16 rGPR:$elem)>;
2393  def : Pat<(v4f32 (ARMvdup (i32 rGPR:$elem))),
2394            (MVE_VDUP32 rGPR:$elem)>;
2395
2396  // Match a vselect with an ARMvdup as a predicated MVE_VDUP
2397  def : Pat<(v16i8 (vselect (v16i1 VCCR:$pred),
2398                            (v16i8 (ARMvdup (i32 rGPR:$elem))),
2399                            (v16i8 MQPR:$inactive))),
2400            (MVE_VDUP8  rGPR:$elem, ARMVCCThen, (v16i1 VCCR:$pred),
2401                        (v16i8 MQPR:$inactive))>;
2402  def : Pat<(v8i16 (vselect (v8i1 VCCR:$pred),
2403                            (v8i16 (ARMvdup (i32 rGPR:$elem))),
2404                            (v8i16 MQPR:$inactive))),
2405            (MVE_VDUP16 rGPR:$elem, ARMVCCThen, (v8i1 VCCR:$pred),
2406                            (v8i16 MQPR:$inactive))>;
2407  def : Pat<(v4i32 (vselect (v4i1 VCCR:$pred),
2408                            (v4i32 (ARMvdup (i32 rGPR:$elem))),
2409                            (v4i32 MQPR:$inactive))),
2410            (MVE_VDUP32 rGPR:$elem, ARMVCCThen, (v4i1 VCCR:$pred),
2411                            (v4i32 MQPR:$inactive))>;
2412  def : Pat<(v4f32 (vselect (v4i1 VCCR:$pred),
2413                            (v4f32 (ARMvdup (i32 rGPR:$elem))),
2414                            (v4f32 MQPR:$inactive))),
2415            (MVE_VDUP32 rGPR:$elem, ARMVCCThen, (v4i1 VCCR:$pred),
2416                            (v4f32 MQPR:$inactive))>;
2417  def : Pat<(v8f16 (vselect (v8i1 VCCR:$pred),
2418                            (v8f16 (ARMvdup (i32 rGPR:$elem))),
2419                            (v8f16 MQPR:$inactive))),
2420            (MVE_VDUP16 rGPR:$elem, ARMVCCThen, (v8i1 VCCR:$pred),
2421                            (v8f16 MQPR:$inactive))>;
2422}
2423
2424
2425class MVEIntSingleSrc<string iname, string suffix, bits<2> size,
2426                         list<dag> pattern=[]>
2427  : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qm), NoItinerary,
2428          iname, suffix, "$Qd, $Qm", vpred_r, "", pattern> {
2429  bits<4> Qd;
2430  bits<4> Qm;
2431
2432  let Inst{22} = Qd{3};
2433  let Inst{19-18} = size{1-0};
2434  let Inst{15-13} = Qd{2-0};
2435  let Inst{5} = Qm{3};
2436  let Inst{3-1} = Qm{2-0};
2437}
2438
2439class MVE_VCLSCLZ<string iname, string suffix, bits<2> size,
2440                   bit count_zeroes, list<dag> pattern=[]>
2441  : MVEIntSingleSrc<iname, suffix, size, pattern> {
2442
2443  let Inst{28} = 0b1;
2444  let Inst{25-23} = 0b111;
2445  let Inst{21-20} = 0b11;
2446  let Inst{17-16} = 0b00;
2447  let Inst{12-8} = 0b00100;
2448  let Inst{7} = count_zeroes;
2449  let Inst{6} = 0b1;
2450  let Inst{4} = 0b0;
2451  let Inst{0} = 0b0;
2452  let validForTailPredication = 1;
2453}
2454
2455multiclass MVE_VCLSCLZ_p<string opname, bit opcode, MVEVectorVTInfo VTI,
2456                         SDPatternOperator unpred_op> {
2457  def "": MVE_VCLSCLZ<"v"#opname, VTI.Suffix, VTI.Size, opcode>;
2458
2459  defvar Inst     = !cast<Instruction>(NAME);
2460  defvar pred_int = !cast<Intrinsic>("int_arm_mve_"#opname#"_predicated");
2461
2462  let Predicates = [HasMVEInt] in {
2463    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$val))),
2464              (VTI.Vec (Inst (VTI.Vec MQPR:$val)))>;
2465    def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$val), (VTI.Pred VCCR:$pred),
2466                                 (VTI.Vec MQPR:$inactive))),
2467              (VTI.Vec (Inst (VTI.Vec MQPR:$val), ARMVCCThen,
2468                             (VTI.Pred VCCR:$pred), (VTI.Vec MQPR:$inactive)))>;
2469  }
2470}
2471
2472defm MVE_VCLSs8  : MVE_VCLSCLZ_p<"cls", 0, MVE_v16s8, int_arm_mve_vcls>;
2473defm MVE_VCLSs16 : MVE_VCLSCLZ_p<"cls", 0, MVE_v8s16, int_arm_mve_vcls>;
2474defm MVE_VCLSs32 : MVE_VCLSCLZ_p<"cls", 0, MVE_v4s32, int_arm_mve_vcls>;
2475
2476defm MVE_VCLZs8  : MVE_VCLSCLZ_p<"clz", 1, MVE_v16i8, ctlz>;
2477defm MVE_VCLZs16 : MVE_VCLSCLZ_p<"clz", 1, MVE_v8i16, ctlz>;
2478defm MVE_VCLZs32 : MVE_VCLSCLZ_p<"clz", 1, MVE_v4i32, ctlz>;
2479
2480class MVE_VABSNEG_int<string iname, string suffix, bits<2> size, bit negate,
2481                      bit saturate, list<dag> pattern=[]>
2482  : MVEIntSingleSrc<iname, suffix, size, pattern> {
2483
2484  let Inst{28} = 0b1;
2485  let Inst{25-23} = 0b111;
2486  let Inst{21-20} = 0b11;
2487  let Inst{17} = 0b0;
2488  let Inst{16} = !eq(saturate, 0);
2489  let Inst{12-11} = 0b00;
2490  let Inst{10} = saturate;
2491  let Inst{9-8} = 0b11;
2492  let Inst{7} = negate;
2493  let Inst{6} = 0b1;
2494  let Inst{4} = 0b0;
2495  let Inst{0} = 0b0;
2496  let validForTailPredication = 1;
2497}
2498
2499multiclass MVE_VABSNEG_int_m<string iname, bit negate, bit saturate,
2500                             SDPatternOperator unpred_op, Intrinsic pred_int,
2501                             MVEVectorVTInfo VTI> {
2502  def "" : MVE_VABSNEG_int<iname, VTI.Suffix, VTI.Size, negate, saturate>;
2503  defvar Inst = !cast<Instruction>(NAME);
2504
2505  let Predicates = [HasMVEInt] in {
2506    // VQABS and VQNEG have more difficult isel patterns defined elsewhere
2507    if !not(saturate) then {
2508      def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$v))),
2509                (VTI.Vec (Inst $v))>;
2510    }
2511
2512    def : Pat<(VTI.Vec (pred_int  (VTI.Vec MQPR:$v), (VTI.Pred VCCR:$mask),
2513                                  (VTI.Vec MQPR:$inactive))),
2514              (VTI.Vec (Inst $v, ARMVCCThen, $mask, $inactive))>;
2515  }
2516}
2517
2518foreach VTI = [ MVE_v16s8, MVE_v8s16, MVE_v4s32 ] in {
2519  defm "MVE_VABS" # VTI.Suffix : MVE_VABSNEG_int_m<
2520     "vabs",  0, 0, abs,   int_arm_mve_abs_predicated,  VTI>;
2521  defm "MVE_VQABS" # VTI.Suffix : MVE_VABSNEG_int_m<
2522     "vqabs", 0, 1, ?,     int_arm_mve_qabs_predicated, VTI>;
2523  defm "MVE_VNEG" # VTI.Suffix : MVE_VABSNEG_int_m<
2524     "vneg",  1, 0, vnegq, int_arm_mve_neg_predicated,  VTI>;
2525  defm "MVE_VQNEG" # VTI.Suffix : MVE_VABSNEG_int_m<
2526     "vqneg", 1, 1, ?,     int_arm_mve_qneg_predicated, VTI>;
2527}
2528
2529// int_min/int_max: vector containing INT_MIN/INT_MAX VTI.Size times
2530// zero_vec: v4i32-initialized zero vector, potentially wrapped in a bitconvert
2531multiclass vqabsneg_pattern<MVEVectorVTInfo VTI, dag int_min, dag int_max,
2532                         dag zero_vec,  MVE_VABSNEG_int vqabs_instruction,
2533                         MVE_VABSNEG_int vqneg_instruction> {
2534  let Predicates = [HasMVEInt] in {
2535    // The below tree can be replaced by a vqabs instruction, as it represents
2536    // the following vectorized expression (r being the value in $reg):
2537    // r > 0 ? r : (r == INT_MIN ? INT_MAX : -r)
2538    def : Pat<(VTI.Vec (vselect
2539                      (VTI.Pred (ARMvcmpz (VTI.Vec MQPR:$reg), ARMCCgt)),
2540                      (VTI.Vec MQPR:$reg),
2541                      (VTI.Vec (vselect
2542                                (VTI.Pred (ARMvcmp (VTI.Vec MQPR:$reg), int_min, ARMCCeq)),
2543                                int_max,
2544                                (sub (VTI.Vec zero_vec), (VTI.Vec MQPR:$reg)))))),
2545            (VTI.Vec (vqabs_instruction (VTI.Vec MQPR:$reg)))>;
2546    // Similarly, this tree represents vqneg, i.e. the following vectorized expression:
2547    // r == INT_MIN ? INT_MAX : -r
2548    def : Pat<(VTI.Vec (vselect
2549                        (VTI.Pred (ARMvcmp (VTI.Vec MQPR:$reg), int_min, ARMCCeq)),
2550                        int_max,
2551                        (sub (VTI.Vec zero_vec), (VTI.Vec MQPR:$reg)))),
2552               (VTI.Vec (vqneg_instruction (VTI.Vec MQPR:$reg)))>;
2553  }
2554}
2555
2556defm MVE_VQABSNEG_Ps8  : vqabsneg_pattern<MVE_v16i8,
2557                                    (v16i8 (ARMvmovImm (i32 3712))),
2558                                    (v16i8 (ARMvmovImm (i32 3711))),
2559                                    (bitconvert (v4i32 (ARMvmovImm (i32 0)))),
2560                                    MVE_VQABSs8, MVE_VQNEGs8>;
2561defm MVE_VQABSNEG_Ps16 : vqabsneg_pattern<MVE_v8i16,
2562                                    (v8i16 (ARMvmovImm (i32 2688))),
2563                                    (v8i16 (ARMvmvnImm (i32 2688))),
2564                                    (bitconvert (v4i32 (ARMvmovImm (i32 0)))),
2565                                    MVE_VQABSs16, MVE_VQNEGs16>;
2566defm MVE_VQABSNEG_Ps32 : vqabsneg_pattern<MVE_v4i32,
2567                                    (v4i32 (ARMvmovImm (i32 1664))),
2568                                    (v4i32 (ARMvmvnImm (i32 1664))),
2569                                    (ARMvmovImm (i32 0)),
2570                                    MVE_VQABSs32, MVE_VQNEGs32>;
2571
2572class MVE_mod_imm<string iname, string suffix, bits<4> cmode, bit op,
2573                  dag iops, list<dag> pattern=[]>
2574  : MVE_p<(outs MQPR:$Qd), iops, NoItinerary, iname, suffix, "$Qd, $imm",
2575          vpred_r, "", pattern> {
2576  bits<13> imm;
2577  bits<4> Qd;
2578
2579  let Inst{28} = imm{7};
2580  let Inst{25-23} = 0b111;
2581  let Inst{22} = Qd{3};
2582  let Inst{21-19} = 0b000;
2583  let Inst{18-16} = imm{6-4};
2584  let Inst{15-13} = Qd{2-0};
2585  let Inst{12} = 0b0;
2586  let Inst{11-8} = cmode{3-0};
2587  let Inst{7-6} = 0b01;
2588  let Inst{5} = op;
2589  let Inst{4} = 0b1;
2590  let Inst{3-0} = imm{3-0};
2591
2592  let DecoderMethod = "DecodeMVEModImmInstruction";
2593  let validForTailPredication = 1;
2594}
2595
2596let isReMaterializable = 1 in {
2597let isAsCheapAsAMove = 1 in {
2598def MVE_VMOVimmi8  : MVE_mod_imm<"vmov", "i8",  {1,1,1,0}, 0b0, (ins nImmSplatI8:$imm)>;
2599def MVE_VMOVimmi16 : MVE_mod_imm<"vmov", "i16", {1,0,?,0}, 0b0, (ins nImmSplatI16:$imm)> {
2600  let Inst{9} = imm{9};
2601}
2602def MVE_VMOVimmi32 : MVE_mod_imm<"vmov", "i32", {?,?,?,?}, 0b0, (ins nImmVMOVI32:$imm)> {
2603  let Inst{11-8} = imm{11-8};
2604}
2605def MVE_VMOVimmi64 : MVE_mod_imm<"vmov", "i64", {1,1,1,0}, 0b1, (ins nImmSplatI64:$imm)>;
2606def MVE_VMOVimmf32 : MVE_mod_imm<"vmov", "f32", {1,1,1,1}, 0b0, (ins nImmVMOVF32:$imm)>;
2607} // let isAsCheapAsAMove = 1
2608
2609def MVE_VMVNimmi16 : MVE_mod_imm<"vmvn", "i16", {1,0,?,0}, 0b1, (ins nImmSplatI16:$imm)> {
2610  let Inst{9} = imm{9};
2611}
2612def MVE_VMVNimmi32 : MVE_mod_imm<"vmvn", "i32", {?,?,?,?}, 0b1, (ins nImmVMOVI32:$imm)> {
2613  let Inst{11-8} = imm{11-8};
2614}
2615} // let isReMaterializable = 1
2616
2617let Predicates = [HasMVEInt] in {
2618  def : Pat<(v16i8 (ARMvmovImm timm:$simm)),
2619            (v16i8 (MVE_VMOVimmi8  nImmSplatI8:$simm))>;
2620  def : Pat<(v8i16 (ARMvmovImm timm:$simm)),
2621            (v8i16 (MVE_VMOVimmi16 nImmSplatI16:$simm))>;
2622  def : Pat<(v4i32 (ARMvmovImm timm:$simm)),
2623            (v4i32 (MVE_VMOVimmi32 nImmVMOVI32:$simm))>;
2624  def : Pat<(v2i64 (ARMvmovImm timm:$simm)),
2625            (v2i64 (MVE_VMOVimmi64 nImmSplatI64:$simm))>;
2626
2627  def : Pat<(v8i16 (ARMvmvnImm timm:$simm)),
2628            (v8i16 (MVE_VMVNimmi16 nImmSplatI16:$simm))>;
2629  def : Pat<(v4i32 (ARMvmvnImm timm:$simm)),
2630            (v4i32 (MVE_VMVNimmi32 nImmVMOVI32:$simm))>;
2631
2632  def : Pat<(v4f32 (ARMvmovFPImm timm:$simm)),
2633            (v4f32 (MVE_VMOVimmf32 nImmVMOVF32:$simm))>;
2634
2635  def : Pat<(v8i16 (vselect (v8i1 VCCR:$pred), (ARMvmvnImm timm:$simm),
2636                            MQPR:$inactive)),
2637            (v8i16 (MVE_VMVNimmi16 nImmSplatI16:$simm,
2638                            ARMVCCThen, VCCR:$pred, MQPR:$inactive))>;
2639  def : Pat<(v4i32 (vselect (v4i1 VCCR:$pred), (ARMvmvnImm timm:$simm),
2640                            MQPR:$inactive)),
2641            (v4i32 (MVE_VMVNimmi32 nImmSplatI32:$simm,
2642                            ARMVCCThen, VCCR:$pred, MQPR:$inactive))>;
2643}
2644
2645class MVE_VMINMAXA<string iname, string suffix, bits<2> size,
2646                   bit bit_12, list<dag> pattern=[]>
2647  : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qd_src, MQPR:$Qm),
2648          NoItinerary, iname, suffix, "$Qd, $Qm", vpred_n, "$Qd = $Qd_src",
2649          pattern> {
2650  bits<4> Qd;
2651  bits<4> Qm;
2652
2653  let Inst{28} = 0b0;
2654  let Inst{25-23} = 0b100;
2655  let Inst{22} = Qd{3};
2656  let Inst{21-20} = 0b11;
2657  let Inst{19-18} = size;
2658  let Inst{17-16} = 0b11;
2659  let Inst{15-13} = Qd{2-0};
2660  let Inst{12} = bit_12;
2661  let Inst{11-6} = 0b111010;
2662  let Inst{5} = Qm{3};
2663  let Inst{4} = 0b0;
2664  let Inst{3-1} = Qm{2-0};
2665  let Inst{0} = 0b1;
2666  let validForTailPredication = 1;
2667}
2668
2669multiclass MVE_VMINMAXA_m<string iname, MVEVectorVTInfo VTI,
2670                      SDNode unpred_op, Intrinsic pred_int, bit bit_12> {
2671  def "" : MVE_VMINMAXA<iname, VTI.Suffix, VTI.Size, bit_12>;
2672  defvar Inst = !cast<Instruction>(NAME);
2673
2674  let Predicates = [HasMVEInt] in {
2675    // Unpredicated v(min|max)a
2676    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$Qd), (abs (VTI.Vec MQPR:$Qm)))),
2677              (VTI.Vec (Inst (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm)))>;
2678
2679    // Predicated v(min|max)a
2680    def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm),
2681                            (VTI.Pred VCCR:$mask))),
2682              (VTI.Vec (Inst (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm),
2683                            ARMVCCThen, (VTI.Pred VCCR:$mask)))>;
2684  }
2685}
2686
2687multiclass MVE_VMINA<MVEVectorVTInfo VTI>
2688  : MVE_VMINMAXA_m<"vmina", VTI, umin, int_arm_mve_vmina_predicated, 0b1>;
2689
2690defm MVE_VMINAs8  : MVE_VMINA<MVE_v16s8>;
2691defm MVE_VMINAs16 : MVE_VMINA<MVE_v8s16>;
2692defm MVE_VMINAs32 : MVE_VMINA<MVE_v4s32>;
2693
2694multiclass MVE_VMAXA<MVEVectorVTInfo VTI>
2695  : MVE_VMINMAXA_m<"vmaxa", VTI, umax, int_arm_mve_vmaxa_predicated, 0b0>;
2696
2697defm MVE_VMAXAs8  : MVE_VMAXA<MVE_v16s8>;
2698defm MVE_VMAXAs16 : MVE_VMAXA<MVE_v8s16>;
2699defm MVE_VMAXAs32 : MVE_VMAXA<MVE_v4s32>;
2700
2701// end of MVE Integer instructions
2702
2703// start of mve_imm_shift instructions
2704
2705def MVE_VSHLC : MVE_p<(outs rGPR:$RdmDest, MQPR:$Qd),
2706                      (ins MQPR:$QdSrc, rGPR:$RdmSrc, long_shift:$imm),
2707                      NoItinerary, "vshlc", "", "$QdSrc, $RdmSrc, $imm",
2708                      vpred_n, "$RdmDest = $RdmSrc,$Qd = $QdSrc"> {
2709  bits<5> imm;
2710  bits<4> Qd;
2711  bits<4> RdmDest;
2712
2713  let Inst{28} = 0b0;
2714  let Inst{25-23} = 0b101;
2715  let Inst{22} = Qd{3};
2716  let Inst{21} = 0b1;
2717  let Inst{20-16} = imm{4-0};
2718  let Inst{15-13} = Qd{2-0};
2719  let Inst{12-4} = 0b011111100;
2720  let Inst{3-0} = RdmDest{3-0};
2721}
2722
2723class MVE_shift_imm<dag oops, dag iops, string iname, string suffix,
2724                    string ops, vpred_ops vpred, string cstr,
2725                    list<dag> pattern=[]>
2726  : MVE_p<oops, iops, NoItinerary, iname, suffix, ops, vpred, cstr, pattern> {
2727  bits<4> Qd;
2728  bits<4> Qm;
2729
2730  let Inst{22} = Qd{3};
2731  let Inst{15-13} = Qd{2-0};
2732  let Inst{5} = Qm{3};
2733  let Inst{3-1} = Qm{2-0};
2734}
2735
2736class MVE_VMOVL<string iname, string suffix, bits<2> sz, bit U, bit top,
2737              list<dag> pattern=[]>
2738  : MVE_shift_imm<(outs MQPR:$Qd), (ins MQPR:$Qm),
2739                  iname, suffix, "$Qd, $Qm", vpred_r, "",
2740                  pattern> {
2741  let Inst{28} = U;
2742  let Inst{25-23} = 0b101;
2743  let Inst{21} = 0b1;
2744  let Inst{20-19} = sz{1-0};
2745  let Inst{18-16} = 0b000;
2746  let Inst{12} = top;
2747  let Inst{11-6} = 0b111101;
2748  let Inst{4} = 0b0;
2749  let Inst{0} = 0b0;
2750  let doubleWidthResult = 1;
2751}
2752
2753multiclass MVE_VMOVL_m<bit top, string chr, MVEVectorVTInfo OutVTI,
2754                       MVEVectorVTInfo InVTI> {
2755  def "": MVE_VMOVL<"vmovl" # chr, InVTI.Suffix, OutVTI.Size,
2756                    InVTI.Unsigned, top>;
2757  defvar Inst = !cast<Instruction>(NAME);
2758
2759  def : Pat<(OutVTI.Vec (int_arm_mve_vmovl_predicated (InVTI.Vec MQPR:$src),
2760                            (i32 InVTI.Unsigned), (i32 top),
2761                            (OutVTI.Pred VCCR:$pred),
2762                            (OutVTI.Vec MQPR:$inactive))),
2763            (OutVTI.Vec (Inst (InVTI.Vec MQPR:$src), ARMVCCThen,
2764                            (OutVTI.Pred VCCR:$pred),
2765                            (OutVTI.Vec MQPR:$inactive)))>;
2766}
2767
2768defm MVE_VMOVLs8bh  : MVE_VMOVL_m<0, "b", MVE_v8s16, MVE_v16s8>;
2769defm MVE_VMOVLs8th  : MVE_VMOVL_m<1, "t", MVE_v8s16, MVE_v16s8>;
2770defm MVE_VMOVLu8bh  : MVE_VMOVL_m<0, "b", MVE_v8u16, MVE_v16u8>;
2771defm MVE_VMOVLu8th  : MVE_VMOVL_m<1, "t", MVE_v8u16, MVE_v16u8>;
2772defm MVE_VMOVLs16bh : MVE_VMOVL_m<0, "b", MVE_v4s32, MVE_v8s16>;
2773defm MVE_VMOVLs16th : MVE_VMOVL_m<1, "t", MVE_v4s32, MVE_v8s16>;
2774defm MVE_VMOVLu16bh : MVE_VMOVL_m<0, "b", MVE_v4s32, MVE_v8u16>;
2775defm MVE_VMOVLu16th : MVE_VMOVL_m<1, "t", MVE_v4s32, MVE_v8u16>;
2776
2777let Predicates = [HasMVEInt] in {
2778  def : Pat<(sext_inreg (v4i32 MQPR:$src), v4i16),
2779            (MVE_VMOVLs16bh MQPR:$src)>;
2780  def : Pat<(sext_inreg (v8i16 MQPR:$src), v8i8),
2781            (MVE_VMOVLs8bh MQPR:$src)>;
2782  def : Pat<(sext_inreg (v4i32 MQPR:$src), v4i8),
2783            (MVE_VMOVLs16bh (MVE_VMOVLs8bh MQPR:$src))>;
2784
2785  def : Pat<(sext_inreg (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src)))), v8i8),
2786            (MVE_VMOVLs8th MQPR:$src)>;
2787  def : Pat<(sext_inreg (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src)))), v4i16),
2788            (MVE_VMOVLs16th MQPR:$src)>;
2789
2790  // zext_inreg 8 -> 16
2791  def : Pat<(ARMvbicImm (v8i16 MQPR:$src), (i32 0xAFF)),
2792            (MVE_VMOVLu8bh MQPR:$src)>;
2793  // zext_inreg 16 -> 32
2794  def : Pat<(and (v4i32 MQPR:$src), (v4i32 (ARMvmovImm (i32 0xCFF)))),
2795            (MVE_VMOVLu16bh MQPR:$src)>;
2796  // Same zext_inreg with vrevs, picking the top half
2797  def : Pat<(ARMvbicImm (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src)))), (i32 0xAFF)),
2798            (MVE_VMOVLu8th MQPR:$src)>;
2799  def : Pat<(and (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src)))),
2800                 (v4i32 (ARMvmovImm (i32 0xCFF)))),
2801            (MVE_VMOVLu16th MQPR:$src)>;
2802}
2803
2804
2805class MVE_VSHLL_imm<string iname, string suffix, bit U, bit th,
2806                    Operand immtype, list<dag> pattern=[]>
2807  : MVE_shift_imm<(outs MQPR:$Qd), (ins MQPR:$Qm, immtype:$imm),
2808                  iname, suffix, "$Qd, $Qm, $imm", vpred_r, "", pattern> {
2809  let Inst{28} = U;
2810  let Inst{25-23} = 0b101;
2811  let Inst{21} = 0b1;
2812  let Inst{12} = th;
2813  let Inst{11-6} = 0b111101;
2814  let Inst{4} = 0b0;
2815  let Inst{0} = 0b0;
2816
2817  // For the MVE_VSHLL_patterns multiclass to refer to
2818  Operand immediateType = immtype;
2819
2820  let doubleWidthResult = 1;
2821}
2822
2823// The immediate VSHLL instructions accept shift counts from 1 up to
2824// the lane width (8 or 16), but the full-width shifts have an
2825// entirely separate encoding, given below with 'lw' in the name.
2826
2827class MVE_VSHLL_imm8<string iname, string suffix,
2828                     bit U, bit th, list<dag> pattern=[]>
2829  : MVE_VSHLL_imm<iname, suffix, U, th, mve_shift_imm1_7, pattern> {
2830  bits<3> imm;
2831  let Inst{20-19} = 0b01;
2832  let Inst{18-16} = imm;
2833}
2834
2835class MVE_VSHLL_imm16<string iname, string suffix,
2836                      bit U, bit th, list<dag> pattern=[]>
2837  : MVE_VSHLL_imm<iname, suffix, U, th, mve_shift_imm1_15, pattern> {
2838  bits<4> imm;
2839  let Inst{20} = 0b1;
2840  let Inst{19-16} = imm;
2841}
2842
2843def MVE_VSHLL_imms8bh  : MVE_VSHLL_imm8 <"vshllb", "s8", 0b0, 0b0>;
2844def MVE_VSHLL_imms8th  : MVE_VSHLL_imm8 <"vshllt", "s8", 0b0, 0b1>;
2845def MVE_VSHLL_immu8bh  : MVE_VSHLL_imm8 <"vshllb", "u8", 0b1, 0b0>;
2846def MVE_VSHLL_immu8th  : MVE_VSHLL_imm8 <"vshllt", "u8", 0b1, 0b1>;
2847def MVE_VSHLL_imms16bh : MVE_VSHLL_imm16<"vshllb", "s16", 0b0, 0b0>;
2848def MVE_VSHLL_imms16th : MVE_VSHLL_imm16<"vshllt", "s16", 0b0, 0b1>;
2849def MVE_VSHLL_immu16bh : MVE_VSHLL_imm16<"vshllb", "u16", 0b1, 0b0>;
2850def MVE_VSHLL_immu16th : MVE_VSHLL_imm16<"vshllt", "u16", 0b1, 0b1>;
2851
2852class MVE_VSHLL_by_lane_width<string iname, string suffix, bits<2> size,
2853                              bit U, string ops, list<dag> pattern=[]>
2854  : MVE_shift_imm<(outs MQPR:$Qd), (ins MQPR:$Qm),
2855                  iname, suffix, ops, vpred_r, "", pattern> {
2856  let Inst{28} = U;
2857  let Inst{25-23} = 0b100;
2858  let Inst{21-20} = 0b11;
2859  let Inst{19-18} = size{1-0};
2860  let Inst{17-16} = 0b01;
2861  let Inst{11-6} = 0b111000;
2862  let Inst{4} = 0b0;
2863  let Inst{0} = 0b1;
2864  let doubleWidthResult = 1;
2865}
2866
2867multiclass MVE_VSHLL_lw<string iname, string suffix, bits<2> sz, bit U,
2868                              string ops, list<dag> pattern=[]> {
2869  def bh : MVE_VSHLL_by_lane_width<iname#"b", suffix, sz, U, ops, pattern> {
2870    let Inst{12} = 0b0;
2871  }
2872  def th : MVE_VSHLL_by_lane_width<iname#"t", suffix, sz, U, ops, pattern> {
2873    let Inst{12} = 0b1;
2874  }
2875}
2876
2877defm MVE_VSHLL_lws8  : MVE_VSHLL_lw<"vshll", "s8",  0b00, 0b0, "$Qd, $Qm, #8">;
2878defm MVE_VSHLL_lws16 : MVE_VSHLL_lw<"vshll", "s16", 0b01, 0b0, "$Qd, $Qm, #16">;
2879defm MVE_VSHLL_lwu8  : MVE_VSHLL_lw<"vshll", "u8",  0b00, 0b1, "$Qd, $Qm, #8">;
2880defm MVE_VSHLL_lwu16 : MVE_VSHLL_lw<"vshll", "u16", 0b01, 0b1, "$Qd, $Qm, #16">;
2881
2882multiclass MVE_VSHLL_patterns<MVEVectorVTInfo VTI, int top> {
2883  defvar suffix     = !strconcat(VTI.Suffix, !if(top, "th", "bh"));
2884  defvar inst_imm   = !cast<MVE_VSHLL_imm>("MVE_VSHLL_imm" # suffix);
2885  defvar inst_lw    = !cast<MVE_VSHLL_by_lane_width>("MVE_VSHLL_lw" # suffix);
2886  defvar unpred_int = int_arm_mve_vshll_imm;
2887  defvar pred_int   = int_arm_mve_vshll_imm_predicated;
2888  defvar imm        = inst_imm.immediateType;
2889
2890  def : Pat<(VTI.DblVec (unpred_int (VTI.Vec MQPR:$src), imm:$imm,
2891                                    (i32 VTI.Unsigned), (i32 top))),
2892            (VTI.DblVec (inst_imm   (VTI.Vec MQPR:$src), imm:$imm))>;
2893  def : Pat<(VTI.DblVec (unpred_int (VTI.Vec MQPR:$src), (i32 VTI.LaneBits),
2894                                    (i32 VTI.Unsigned), (i32 top))),
2895            (VTI.DblVec (inst_lw    (VTI.Vec MQPR:$src)))>;
2896
2897  def : Pat<(VTI.DblVec (pred_int   (VTI.Vec MQPR:$src), imm:$imm,
2898                                    (i32 VTI.Unsigned), (i32 top),
2899                                    (VTI.DblPred VCCR:$mask),
2900                                    (VTI.DblVec MQPR:$inactive))),
2901            (VTI.DblVec (inst_imm   (VTI.Vec MQPR:$src), imm:$imm,
2902                                    ARMVCCThen, (VTI.DblPred VCCR:$mask),
2903                                    (VTI.DblVec MQPR:$inactive)))>;
2904  def : Pat<(VTI.DblVec (pred_int   (VTI.Vec MQPR:$src), (i32 VTI.LaneBits),
2905                                    (i32 VTI.Unsigned), (i32 top),
2906                                    (VTI.DblPred VCCR:$mask),
2907                                    (VTI.DblVec MQPR:$inactive))),
2908            (VTI.DblVec (inst_lw    (VTI.Vec MQPR:$src), ARMVCCThen,
2909                                    (VTI.DblPred VCCR:$mask),
2910                                    (VTI.DblVec MQPR:$inactive)))>;
2911}
2912
2913foreach VTI = [MVE_v16s8, MVE_v8s16, MVE_v16u8, MVE_v8u16] in
2914  foreach top = [0, 1] in
2915    defm : MVE_VSHLL_patterns<VTI, top>;
2916
2917class MVE_shift_imm_partial<Operand imm, string iname, string suffix>
2918  : MVE_shift_imm<(outs MQPR:$Qd), (ins MQPR:$QdSrc, MQPR:$Qm, imm:$imm),
2919                  iname, suffix, "$Qd, $Qm, $imm", vpred_n, "$Qd = $QdSrc"> {
2920  Operand immediateType = imm;
2921}
2922
2923class MVE_VxSHRN<string iname, string suffix, bit bit_12, bit bit_28,
2924                 Operand imm, list<dag> pattern=[]>
2925  : MVE_shift_imm_partial<imm, iname, suffix> {
2926  bits<5> imm;
2927
2928  let Inst{28} = bit_28;
2929  let Inst{25-23} = 0b101;
2930  let Inst{21} = 0b0;
2931  let Inst{20-16} = imm{4-0};
2932  let Inst{12} = bit_12;
2933  let Inst{11-6} = 0b111111;
2934  let Inst{4} = 0b0;
2935  let Inst{0} = 0b1;
2936  let validForTailPredication = 1;
2937  let retainsPreviousHalfElement = 1;
2938}
2939
2940def MVE_VRSHRNi16bh : MVE_VxSHRN<"vrshrnb", "i16", 0b0, 0b1, shr_imm8> {
2941  let Inst{20-19} = 0b01;
2942}
2943def MVE_VRSHRNi16th : MVE_VxSHRN<"vrshrnt", "i16", 0b1, 0b1, shr_imm8> {
2944  let Inst{20-19} = 0b01;
2945}
2946def MVE_VRSHRNi32bh : MVE_VxSHRN<"vrshrnb", "i32", 0b0, 0b1, shr_imm16> {
2947  let Inst{20} = 0b1;
2948}
2949def MVE_VRSHRNi32th : MVE_VxSHRN<"vrshrnt", "i32", 0b1, 0b1, shr_imm16> {
2950  let Inst{20} = 0b1;
2951}
2952
2953def MVE_VSHRNi16bh : MVE_VxSHRN<"vshrnb", "i16", 0b0, 0b0, shr_imm8> {
2954  let Inst{20-19} = 0b01;
2955}
2956def MVE_VSHRNi16th : MVE_VxSHRN<"vshrnt", "i16", 0b1, 0b0, shr_imm8> {
2957  let Inst{20-19} = 0b01;
2958}
2959def MVE_VSHRNi32bh : MVE_VxSHRN<"vshrnb", "i32", 0b0, 0b0, shr_imm16> {
2960  let Inst{20} = 0b1;
2961}
2962def MVE_VSHRNi32th : MVE_VxSHRN<"vshrnt", "i32", 0b1, 0b0, shr_imm16> {
2963  let Inst{20} = 0b1;
2964}
2965
2966class MVE_VxQRSHRUN<string iname, string suffix, bit bit_28, bit bit_12,
2967                    Operand imm, list<dag> pattern=[]>
2968  : MVE_shift_imm_partial<imm, iname, suffix> {
2969  bits<5> imm;
2970
2971  let Inst{28} = bit_28;
2972  let Inst{25-23} = 0b101;
2973  let Inst{21} = 0b0;
2974  let Inst{20-16} = imm{4-0};
2975  let Inst{12} = bit_12;
2976  let Inst{11-6} = 0b111111;
2977  let Inst{4} = 0b0;
2978  let Inst{0} = 0b0;
2979  let validForTailPredication = 1;
2980  let retainsPreviousHalfElement = 1;
2981}
2982
2983def MVE_VQRSHRUNs16bh : MVE_VxQRSHRUN<
2984    "vqrshrunb", "s16", 0b1, 0b0, shr_imm8> {
2985  let Inst{20-19} = 0b01;
2986}
2987def MVE_VQRSHRUNs16th : MVE_VxQRSHRUN<
2988    "vqrshrunt", "s16", 0b1, 0b1, shr_imm8> {
2989  let Inst{20-19} = 0b01;
2990}
2991def MVE_VQRSHRUNs32bh : MVE_VxQRSHRUN<
2992    "vqrshrunb", "s32", 0b1, 0b0, shr_imm16> {
2993  let Inst{20} = 0b1;
2994}
2995def MVE_VQRSHRUNs32th : MVE_VxQRSHRUN<
2996    "vqrshrunt", "s32", 0b1, 0b1, shr_imm16> {
2997  let Inst{20} = 0b1;
2998}
2999
3000def MVE_VQSHRUNs16bh : MVE_VxQRSHRUN<
3001    "vqshrunb", "s16", 0b0, 0b0, shr_imm8> {
3002  let Inst{20-19} = 0b01;
3003}
3004def MVE_VQSHRUNs16th : MVE_VxQRSHRUN<
3005    "vqshrunt", "s16", 0b0, 0b1, shr_imm8> {
3006  let Inst{20-19} = 0b01;
3007}
3008def MVE_VQSHRUNs32bh : MVE_VxQRSHRUN<
3009    "vqshrunb", "s32", 0b0, 0b0, shr_imm16> {
3010  let Inst{20} = 0b1;
3011}
3012def MVE_VQSHRUNs32th : MVE_VxQRSHRUN<
3013    "vqshrunt", "s32", 0b0, 0b1, shr_imm16> {
3014  let Inst{20} = 0b1;
3015}
3016
3017class MVE_VxQRSHRN<string iname, string suffix, bit bit_0, bit bit_12,
3018                   Operand imm, list<dag> pattern=[]>
3019  : MVE_shift_imm_partial<imm, iname, suffix> {
3020  bits<5> imm;
3021
3022  let Inst{25-23} = 0b101;
3023  let Inst{21} = 0b0;
3024  let Inst{20-16} = imm{4-0};
3025  let Inst{12} = bit_12;
3026  let Inst{11-6} = 0b111101;
3027  let Inst{4} = 0b0;
3028  let Inst{0} = bit_0;
3029  let validForTailPredication = 1;
3030  let retainsPreviousHalfElement = 1;
3031}
3032
3033multiclass MVE_VxQRSHRN_types<string iname, bit bit_0, bit bit_12> {
3034  def s16 : MVE_VxQRSHRN<iname, "s16", bit_0, bit_12, shr_imm8> {
3035    let Inst{28} = 0b0;
3036    let Inst{20-19} = 0b01;
3037  }
3038  def u16 : MVE_VxQRSHRN<iname, "u16", bit_0, bit_12, shr_imm8> {
3039    let Inst{28} = 0b1;
3040    let Inst{20-19} = 0b01;
3041  }
3042  def s32 : MVE_VxQRSHRN<iname, "s32", bit_0, bit_12, shr_imm16> {
3043    let Inst{28} = 0b0;
3044    let Inst{20} = 0b1;
3045  }
3046  def u32 : MVE_VxQRSHRN<iname, "u32", bit_0, bit_12, shr_imm16> {
3047    let Inst{28} = 0b1;
3048    let Inst{20} = 0b1;
3049  }
3050}
3051
3052defm MVE_VQRSHRNbh : MVE_VxQRSHRN_types<"vqrshrnb", 0b1, 0b0>;
3053defm MVE_VQRSHRNth : MVE_VxQRSHRN_types<"vqrshrnt", 0b1, 0b1>;
3054defm MVE_VQSHRNbh  : MVE_VxQRSHRN_types<"vqshrnb", 0b0, 0b0>;
3055defm MVE_VQSHRNth  : MVE_VxQRSHRN_types<"vqshrnt", 0b0, 0b1>;
3056
3057multiclass MVE_VSHRN_patterns<MVE_shift_imm_partial inst,
3058                              MVEVectorVTInfo OutVTI, MVEVectorVTInfo InVTI,
3059                              bit q, bit r, bit top> {
3060  defvar inparams = (? (OutVTI.Vec MQPR:$QdSrc), (InVTI.Vec MQPR:$Qm),
3061                       (inst.immediateType:$imm), (i32 q), (i32 r),
3062                       (i32 OutVTI.Unsigned), (i32 InVTI.Unsigned), (i32 top));
3063  defvar outparams = (inst (OutVTI.Vec MQPR:$QdSrc), (InVTI.Vec MQPR:$Qm),
3064                           (imm:$imm));
3065
3066  def : Pat<(OutVTI.Vec !setdagop(inparams, int_arm_mve_vshrn)),
3067            (OutVTI.Vec outparams)>;
3068  def : Pat<(OutVTI.Vec !con(inparams, (int_arm_mve_vshrn_predicated
3069                                           (InVTI.Pred VCCR:$pred)))),
3070            (OutVTI.Vec !con(outparams, (? ARMVCCThen, VCCR:$pred)))>;
3071}
3072
3073defm : MVE_VSHRN_patterns<MVE_VSHRNi16bh,    MVE_v16s8, MVE_v8s16, 0,0,0>;
3074defm : MVE_VSHRN_patterns<MVE_VSHRNi16th,    MVE_v16s8, MVE_v8s16, 0,0,1>;
3075defm : MVE_VSHRN_patterns<MVE_VSHRNi32bh,    MVE_v8s16, MVE_v4s32, 0,0,0>;
3076defm : MVE_VSHRN_patterns<MVE_VSHRNi32th,    MVE_v8s16, MVE_v4s32, 0,0,1>;
3077defm : MVE_VSHRN_patterns<MVE_VSHRNi16bh,    MVE_v16u8, MVE_v8u16, 0,0,0>;
3078defm : MVE_VSHRN_patterns<MVE_VSHRNi16th,    MVE_v16u8, MVE_v8u16, 0,0,1>;
3079defm : MVE_VSHRN_patterns<MVE_VSHRNi32bh,    MVE_v8u16, MVE_v4u32, 0,0,0>;
3080defm : MVE_VSHRN_patterns<MVE_VSHRNi32th,    MVE_v8u16, MVE_v4u32, 0,0,1>;
3081defm : MVE_VSHRN_patterns<MVE_VRSHRNi16bh,   MVE_v16s8, MVE_v8s16, 0,1,0>;
3082defm : MVE_VSHRN_patterns<MVE_VRSHRNi16th,   MVE_v16s8, MVE_v8s16, 0,1,1>;
3083defm : MVE_VSHRN_patterns<MVE_VRSHRNi32bh,   MVE_v8s16, MVE_v4s32, 0,1,0>;
3084defm : MVE_VSHRN_patterns<MVE_VRSHRNi32th,   MVE_v8s16, MVE_v4s32, 0,1,1>;
3085defm : MVE_VSHRN_patterns<MVE_VRSHRNi16bh,   MVE_v16u8, MVE_v8u16, 0,1,0>;
3086defm : MVE_VSHRN_patterns<MVE_VRSHRNi16th,   MVE_v16u8, MVE_v8u16, 0,1,1>;
3087defm : MVE_VSHRN_patterns<MVE_VRSHRNi32bh,   MVE_v8u16, MVE_v4u32, 0,1,0>;
3088defm : MVE_VSHRN_patterns<MVE_VRSHRNi32th,   MVE_v8u16, MVE_v4u32, 0,1,1>;
3089defm : MVE_VSHRN_patterns<MVE_VQSHRNbhs16,   MVE_v16s8, MVE_v8s16, 1,0,0>;
3090defm : MVE_VSHRN_patterns<MVE_VQSHRNths16,   MVE_v16s8, MVE_v8s16, 1,0,1>;
3091defm : MVE_VSHRN_patterns<MVE_VQSHRNbhs32,   MVE_v8s16, MVE_v4s32, 1,0,0>;
3092defm : MVE_VSHRN_patterns<MVE_VQSHRNths32,   MVE_v8s16, MVE_v4s32, 1,0,1>;
3093defm : MVE_VSHRN_patterns<MVE_VQSHRNbhu16,   MVE_v16u8, MVE_v8u16, 1,0,0>;
3094defm : MVE_VSHRN_patterns<MVE_VQSHRNthu16,   MVE_v16u8, MVE_v8u16, 1,0,1>;
3095defm : MVE_VSHRN_patterns<MVE_VQSHRNbhu32,   MVE_v8u16, MVE_v4u32, 1,0,0>;
3096defm : MVE_VSHRN_patterns<MVE_VQSHRNthu32,   MVE_v8u16, MVE_v4u32, 1,0,1>;
3097defm : MVE_VSHRN_patterns<MVE_VQRSHRNbhs16,  MVE_v16s8, MVE_v8s16, 1,1,0>;
3098defm : MVE_VSHRN_patterns<MVE_VQRSHRNths16,  MVE_v16s8, MVE_v8s16, 1,1,1>;
3099defm : MVE_VSHRN_patterns<MVE_VQRSHRNbhs32,  MVE_v8s16, MVE_v4s32, 1,1,0>;
3100defm : MVE_VSHRN_patterns<MVE_VQRSHRNths32,  MVE_v8s16, MVE_v4s32, 1,1,1>;
3101defm : MVE_VSHRN_patterns<MVE_VQRSHRNbhu16,  MVE_v16u8, MVE_v8u16, 1,1,0>;
3102defm : MVE_VSHRN_patterns<MVE_VQRSHRNthu16,  MVE_v16u8, MVE_v8u16, 1,1,1>;
3103defm : MVE_VSHRN_patterns<MVE_VQRSHRNbhu32,  MVE_v8u16, MVE_v4u32, 1,1,0>;
3104defm : MVE_VSHRN_patterns<MVE_VQRSHRNthu32,  MVE_v8u16, MVE_v4u32, 1,1,1>;
3105defm : MVE_VSHRN_patterns<MVE_VQSHRUNs16bh,  MVE_v16u8, MVE_v8s16, 1,0,0>;
3106defm : MVE_VSHRN_patterns<MVE_VQSHRUNs16th,  MVE_v16u8, MVE_v8s16, 1,0,1>;
3107defm : MVE_VSHRN_patterns<MVE_VQSHRUNs32bh,  MVE_v8u16, MVE_v4s32, 1,0,0>;
3108defm : MVE_VSHRN_patterns<MVE_VQSHRUNs32th,  MVE_v8u16, MVE_v4s32, 1,0,1>;
3109defm : MVE_VSHRN_patterns<MVE_VQRSHRUNs16bh, MVE_v16u8, MVE_v8s16, 1,1,0>;
3110defm : MVE_VSHRN_patterns<MVE_VQRSHRUNs16th, MVE_v16u8, MVE_v8s16, 1,1,1>;
3111defm : MVE_VSHRN_patterns<MVE_VQRSHRUNs32bh, MVE_v8u16, MVE_v4s32, 1,1,0>;
3112defm : MVE_VSHRN_patterns<MVE_VQRSHRUNs32th, MVE_v8u16, MVE_v4s32, 1,1,1>;
3113
3114// end of mve_imm_shift instructions
3115
3116// start of mve_shift instructions
3117
3118class MVE_shift_by_vec<string iname, string suffix, bit U,
3119                       bits<2> size, bit bit_4, bit bit_8>
3120  : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qm, MQPR:$Qn), NoItinerary,
3121           iname, suffix, "$Qd, $Qm, $Qn", vpred_r, "", []> {
3122  // Shift instructions which take a vector of shift counts
3123  bits<4> Qd;
3124  bits<4> Qm;
3125  bits<4> Qn;
3126
3127  let Inst{28} = U;
3128  let Inst{25-24} = 0b11;
3129  let Inst{23} = 0b0;
3130  let Inst{22} = Qd{3};
3131  let Inst{21-20} = size;
3132  let Inst{19-17} = Qn{2-0};
3133  let Inst{16} = 0b0;
3134  let Inst{15-13} = Qd{2-0};
3135  let Inst{12-9} = 0b0010;
3136  let Inst{8} = bit_8;
3137  let Inst{7} = Qn{3};
3138  let Inst{6} = 0b1;
3139  let Inst{5} = Qm{3};
3140  let Inst{4} = bit_4;
3141  let Inst{3-1} = Qm{2-0};
3142  let Inst{0} = 0b0;
3143  let validForTailPredication = 1;
3144}
3145
3146multiclass MVE_shift_by_vec_p<string iname, MVEVectorVTInfo VTI, bit q, bit r> {
3147  def "" : MVE_shift_by_vec<iname, VTI.Suffix, VTI.Unsigned, VTI.Size, q, r>;
3148  defvar Inst = !cast<Instruction>(NAME);
3149
3150  def : Pat<(VTI.Vec (int_arm_mve_vshl_vector
3151                         (VTI.Vec MQPR:$in), (VTI.Vec MQPR:$sh),
3152                         (i32 q), (i32 r), (i32 VTI.Unsigned))),
3153            (VTI.Vec (Inst (VTI.Vec MQPR:$in), (VTI.Vec MQPR:$sh)))>;
3154
3155  def : Pat<(VTI.Vec (int_arm_mve_vshl_vector_predicated
3156                         (VTI.Vec MQPR:$in), (VTI.Vec MQPR:$sh),
3157                         (i32 q), (i32 r), (i32 VTI.Unsigned),
3158                         (VTI.Pred VCCR:$mask), (VTI.Vec MQPR:$inactive))),
3159            (VTI.Vec (Inst (VTI.Vec MQPR:$in), (VTI.Vec MQPR:$sh),
3160                           ARMVCCThen, (VTI.Pred VCCR:$mask),
3161                           (VTI.Vec MQPR:$inactive)))>;
3162}
3163
3164multiclass mve_shift_by_vec_multi<string iname, bit bit_4, bit bit_8> {
3165  defm s8  : MVE_shift_by_vec_p<iname, MVE_v16s8, bit_4, bit_8>;
3166  defm s16 : MVE_shift_by_vec_p<iname, MVE_v8s16, bit_4, bit_8>;
3167  defm s32 : MVE_shift_by_vec_p<iname, MVE_v4s32, bit_4, bit_8>;
3168  defm u8  : MVE_shift_by_vec_p<iname, MVE_v16u8, bit_4, bit_8>;
3169  defm u16 : MVE_shift_by_vec_p<iname, MVE_v8u16, bit_4, bit_8>;
3170  defm u32 : MVE_shift_by_vec_p<iname, MVE_v4u32, bit_4, bit_8>;
3171}
3172
3173defm MVE_VSHL_by_vec   : mve_shift_by_vec_multi<"vshl",   0b0, 0b0>;
3174defm MVE_VQSHL_by_vec  : mve_shift_by_vec_multi<"vqshl",  0b1, 0b0>;
3175defm MVE_VQRSHL_by_vec : mve_shift_by_vec_multi<"vqrshl", 0b1, 0b1>;
3176defm MVE_VRSHL_by_vec  : mve_shift_by_vec_multi<"vrshl",  0b0, 0b1>;
3177
3178let Predicates = [HasMVEInt] in {
3179  def : Pat<(v4i32 (ARMvshlu (v4i32 MQPR:$Qm), (v4i32 MQPR:$Qn))),
3180            (v4i32 (MVE_VSHL_by_vecu32 (v4i32 MQPR:$Qm), (v4i32 MQPR:$Qn)))>;
3181  def : Pat<(v8i16 (ARMvshlu (v8i16 MQPR:$Qm), (v8i16 MQPR:$Qn))),
3182            (v8i16 (MVE_VSHL_by_vecu16 (v8i16 MQPR:$Qm), (v8i16 MQPR:$Qn)))>;
3183  def : Pat<(v16i8 (ARMvshlu (v16i8 MQPR:$Qm), (v16i8 MQPR:$Qn))),
3184            (v16i8 (MVE_VSHL_by_vecu8 (v16i8 MQPR:$Qm), (v16i8 MQPR:$Qn)))>;
3185
3186  def : Pat<(v4i32 (ARMvshls (v4i32 MQPR:$Qm), (v4i32 MQPR:$Qn))),
3187            (v4i32 (MVE_VSHL_by_vecs32 (v4i32 MQPR:$Qm), (v4i32 MQPR:$Qn)))>;
3188  def : Pat<(v8i16 (ARMvshls (v8i16 MQPR:$Qm), (v8i16 MQPR:$Qn))),
3189            (v8i16 (MVE_VSHL_by_vecs16 (v8i16 MQPR:$Qm), (v8i16 MQPR:$Qn)))>;
3190  def : Pat<(v16i8 (ARMvshls (v16i8 MQPR:$Qm), (v16i8 MQPR:$Qn))),
3191            (v16i8 (MVE_VSHL_by_vecs8 (v16i8 MQPR:$Qm), (v16i8 MQPR:$Qn)))>;
3192}
3193
3194class MVE_shift_with_imm<string iname, string suffix, dag oops, dag iops,
3195                         string ops, vpred_ops vpred, string cstr,
3196                         list<dag> pattern=[]>
3197  : MVE_p<oops, iops, NoItinerary, iname, suffix, ops, vpred, cstr, pattern> {
3198  bits<4> Qd;
3199  bits<4> Qm;
3200
3201  let Inst{23} = 0b1;
3202  let Inst{22} = Qd{3};
3203  let Inst{15-13} = Qd{2-0};
3204  let Inst{12-11} = 0b00;
3205  let Inst{7-6} = 0b01;
3206  let Inst{5} = Qm{3};
3207  let Inst{4} = 0b1;
3208  let Inst{3-1} = Qm{2-0};
3209  let Inst{0} = 0b0;
3210  let validForTailPredication = 1;
3211
3212  // For the MVE_shift_imm_patterns multiclass to refer to
3213  MVEVectorVTInfo VTI;
3214  Operand immediateType;
3215  Intrinsic unpred_int;
3216  Intrinsic pred_int;
3217  dag unsignedFlag = (?);
3218}
3219
3220class MVE_VSxI_imm<string iname, string suffix, bit bit_8, Operand immType>
3221  : MVE_shift_with_imm<iname, suffix, (outs MQPR:$Qd),
3222                       (ins MQPR:$Qd_src, MQPR:$Qm, immType:$imm),
3223                       "$Qd, $Qm, $imm", vpred_n, "$Qd = $Qd_src"> {
3224  bits<6> imm;
3225  let Inst{28} = 0b1;
3226  let Inst{25-24} = 0b11;
3227  let Inst{21-16} = imm;
3228  let Inst{10-9} = 0b10;
3229  let Inst{8} = bit_8;
3230  let validForTailPredication = 1;
3231
3232  Operand immediateType = immType;
3233}
3234
3235def MVE_VSRIimm8 : MVE_VSxI_imm<"vsri", "8", 0b0, shr_imm8> {
3236  let Inst{21-19} = 0b001;
3237}
3238
3239def MVE_VSRIimm16 : MVE_VSxI_imm<"vsri", "16", 0b0, shr_imm16> {
3240  let Inst{21-20} = 0b01;
3241}
3242
3243def MVE_VSRIimm32 : MVE_VSxI_imm<"vsri", "32", 0b0, shr_imm32> {
3244  let Inst{21} = 0b1;
3245}
3246
3247def MVE_VSLIimm8 : MVE_VSxI_imm<"vsli", "8", 0b1, imm0_7> {
3248  let Inst{21-19} = 0b001;
3249}
3250
3251def MVE_VSLIimm16 : MVE_VSxI_imm<"vsli", "16", 0b1, imm0_15> {
3252  let Inst{21-20} = 0b01;
3253}
3254
3255def MVE_VSLIimm32 : MVE_VSxI_imm<"vsli", "32", 0b1,imm0_31> {
3256  let Inst{21} = 0b1;
3257}
3258
3259multiclass MVE_VSxI_patterns<MVE_VSxI_imm inst, string name,
3260                              MVEVectorVTInfo VTI> {
3261  defvar inparams = (? (VTI.Vec MQPR:$QdSrc), (VTI.Vec MQPR:$Qm),
3262                       (inst.immediateType:$imm));
3263  defvar outparams = (inst (VTI.Vec MQPR:$QdSrc), (VTI.Vec MQPR:$Qm),
3264                           (inst.immediateType:$imm));
3265  defvar unpred_int = !cast<Intrinsic>("int_arm_mve_" # name);
3266  defvar pred_int = !cast<Intrinsic>("int_arm_mve_" # name # "_predicated");
3267
3268  def : Pat<(VTI.Vec !setdagop(inparams, unpred_int)),
3269            (VTI.Vec outparams)>;
3270  def : Pat<(VTI.Vec !con(inparams, (pred_int (VTI.Pred VCCR:$pred)))),
3271            (VTI.Vec !con(outparams, (? ARMVCCThen, VCCR:$pred)))>;
3272}
3273
3274defm : MVE_VSxI_patterns<MVE_VSLIimm8,  "vsli", MVE_v16i8>;
3275defm : MVE_VSxI_patterns<MVE_VSLIimm16, "vsli", MVE_v8i16>;
3276defm : MVE_VSxI_patterns<MVE_VSLIimm32, "vsli", MVE_v4i32>;
3277defm : MVE_VSxI_patterns<MVE_VSRIimm8,  "vsri", MVE_v16i8>;
3278defm : MVE_VSxI_patterns<MVE_VSRIimm16, "vsri", MVE_v8i16>;
3279defm : MVE_VSxI_patterns<MVE_VSRIimm32, "vsri", MVE_v4i32>;
3280
3281class MVE_VQSHL_imm<MVEVectorVTInfo VTI_, Operand immType>
3282  : MVE_shift_with_imm<"vqshl", VTI_.Suffix, (outs MQPR:$Qd),
3283                       (ins MQPR:$Qm, immType:$imm), "$Qd, $Qm, $imm",
3284                       vpred_r, ""> {
3285  bits<6> imm;
3286
3287  let Inst{28} = VTI_.Unsigned;
3288  let Inst{25-24} = 0b11;
3289  let Inst{21-16} = imm;
3290  let Inst{10-8} = 0b111;
3291
3292  let VTI = VTI_;
3293  let immediateType = immType;
3294  let unsignedFlag = (? (i32 VTI.Unsigned));
3295}
3296
3297let unpred_int = int_arm_mve_vqshl_imm,
3298    pred_int = int_arm_mve_vqshl_imm_predicated in {
3299  def MVE_VQSHLimms8 : MVE_VQSHL_imm<MVE_v16s8, imm0_7> {
3300    let Inst{21-19} = 0b001;
3301  }
3302  def MVE_VQSHLimmu8 : MVE_VQSHL_imm<MVE_v16u8, imm0_7> {
3303    let Inst{21-19} = 0b001;
3304  }
3305
3306  def MVE_VQSHLimms16 : MVE_VQSHL_imm<MVE_v8s16, imm0_15> {
3307    let Inst{21-20} = 0b01;
3308  }
3309  def MVE_VQSHLimmu16 : MVE_VQSHL_imm<MVE_v8u16, imm0_15> {
3310    let Inst{21-20} = 0b01;
3311  }
3312
3313  def MVE_VQSHLimms32 : MVE_VQSHL_imm<MVE_v4s32, imm0_31> {
3314    let Inst{21} = 0b1;
3315  }
3316  def MVE_VQSHLimmu32 : MVE_VQSHL_imm<MVE_v4u32, imm0_31> {
3317    let Inst{21} = 0b1;
3318  }
3319}
3320
3321class MVE_VQSHLU_imm<MVEVectorVTInfo VTI_, Operand immType>
3322  : MVE_shift_with_imm<"vqshlu", VTI_.Suffix, (outs MQPR:$Qd),
3323                       (ins MQPR:$Qm, immType:$imm), "$Qd, $Qm, $imm",
3324                       vpred_r, ""> {
3325  bits<6> imm;
3326
3327  let Inst{28} = 0b1;
3328  let Inst{25-24} = 0b11;
3329  let Inst{21-16} = imm;
3330  let Inst{10-8} = 0b110;
3331
3332  let VTI = VTI_;
3333  let immediateType = immType;
3334}
3335
3336let unpred_int = int_arm_mve_vqshlu_imm,
3337    pred_int = int_arm_mve_vqshlu_imm_predicated in {
3338  def MVE_VQSHLU_imms8 : MVE_VQSHLU_imm<MVE_v16s8, imm0_7> {
3339    let Inst{21-19} = 0b001;
3340  }
3341
3342  def MVE_VQSHLU_imms16 : MVE_VQSHLU_imm<MVE_v8s16, imm0_15> {
3343    let Inst{21-20} = 0b01;
3344  }
3345
3346  def MVE_VQSHLU_imms32 : MVE_VQSHLU_imm<MVE_v4s32, imm0_31> {
3347    let Inst{21} = 0b1;
3348  }
3349}
3350
3351class MVE_VRSHR_imm<MVEVectorVTInfo VTI_, Operand immType>
3352  : MVE_shift_with_imm<"vrshr", VTI_.Suffix, (outs MQPR:$Qd),
3353                       (ins MQPR:$Qm, immType:$imm), "$Qd, $Qm, $imm",
3354                       vpred_r, ""> {
3355  bits<6> imm;
3356
3357  let Inst{28} = VTI_.Unsigned;
3358  let Inst{25-24} = 0b11;
3359  let Inst{21-16} = imm;
3360  let Inst{10-8} = 0b010;
3361
3362  let VTI = VTI_;
3363  let immediateType = immType;
3364  let unsignedFlag = (? (i32 VTI.Unsigned));
3365}
3366
3367let unpred_int = int_arm_mve_vrshr_imm,
3368    pred_int = int_arm_mve_vrshr_imm_predicated in {
3369  def MVE_VRSHR_imms8 : MVE_VRSHR_imm<MVE_v16s8, shr_imm8> {
3370    let Inst{21-19} = 0b001;
3371  }
3372
3373  def MVE_VRSHR_immu8 : MVE_VRSHR_imm<MVE_v16u8, shr_imm8> {
3374    let Inst{21-19} = 0b001;
3375  }
3376
3377  def MVE_VRSHR_imms16 : MVE_VRSHR_imm<MVE_v8s16, shr_imm16> {
3378    let Inst{21-20} = 0b01;
3379  }
3380
3381  def MVE_VRSHR_immu16 : MVE_VRSHR_imm<MVE_v8u16, shr_imm16> {
3382    let Inst{21-20} = 0b01;
3383  }
3384
3385  def MVE_VRSHR_imms32 : MVE_VRSHR_imm<MVE_v4s32, shr_imm32> {
3386    let Inst{21} = 0b1;
3387  }
3388
3389  def MVE_VRSHR_immu32 : MVE_VRSHR_imm<MVE_v4u32, shr_imm32> {
3390    let Inst{21} = 0b1;
3391  }
3392}
3393
3394multiclass MVE_shift_imm_patterns<MVE_shift_with_imm inst> {
3395  def : Pat<(inst.VTI.Vec !con((inst.unpred_int (inst.VTI.Vec MQPR:$src),
3396                                                inst.immediateType:$imm),
3397                               inst.unsignedFlag)),
3398            (inst.VTI.Vec (inst (inst.VTI.Vec MQPR:$src),
3399                                inst.immediateType:$imm))>;
3400
3401  def : Pat<(inst.VTI.Vec !con((inst.pred_int (inst.VTI.Vec MQPR:$src),
3402                                              inst.immediateType:$imm),
3403                               inst.unsignedFlag,
3404                               (? (inst.VTI.Pred VCCR:$mask),
3405                                  (inst.VTI.Vec MQPR:$inactive)))),
3406            (inst.VTI.Vec (inst (inst.VTI.Vec MQPR:$src),
3407                                inst.immediateType:$imm,
3408                                ARMVCCThen, (inst.VTI.Pred VCCR:$mask),
3409                                (inst.VTI.Vec MQPR:$inactive)))>;
3410}
3411
3412defm : MVE_shift_imm_patterns<MVE_VQSHLimms8>;
3413defm : MVE_shift_imm_patterns<MVE_VQSHLimmu8>;
3414defm : MVE_shift_imm_patterns<MVE_VQSHLimms16>;
3415defm : MVE_shift_imm_patterns<MVE_VQSHLimmu16>;
3416defm : MVE_shift_imm_patterns<MVE_VQSHLimms32>;
3417defm : MVE_shift_imm_patterns<MVE_VQSHLimmu32>;
3418defm : MVE_shift_imm_patterns<MVE_VQSHLU_imms8>;
3419defm : MVE_shift_imm_patterns<MVE_VQSHLU_imms16>;
3420defm : MVE_shift_imm_patterns<MVE_VQSHLU_imms32>;
3421defm : MVE_shift_imm_patterns<MVE_VRSHR_imms8>;
3422defm : MVE_shift_imm_patterns<MVE_VRSHR_immu8>;
3423defm : MVE_shift_imm_patterns<MVE_VRSHR_imms16>;
3424defm : MVE_shift_imm_patterns<MVE_VRSHR_immu16>;
3425defm : MVE_shift_imm_patterns<MVE_VRSHR_imms32>;
3426defm : MVE_shift_imm_patterns<MVE_VRSHR_immu32>;
3427
3428class MVE_VSHR_imm<string suffix, dag imm>
3429  : MVE_shift_with_imm<"vshr", suffix, (outs MQPR:$Qd),
3430                       !con((ins MQPR:$Qm), imm), "$Qd, $Qm, $imm",
3431                       vpred_r, ""> {
3432  bits<6> imm;
3433
3434  let Inst{25-24} = 0b11;
3435  let Inst{21-16} = imm;
3436  let Inst{10-8} = 0b000;
3437}
3438
3439def MVE_VSHR_imms8 : MVE_VSHR_imm<"s8", (ins shr_imm8:$imm)> {
3440  let Inst{28} = 0b0;
3441  let Inst{21-19} = 0b001;
3442}
3443
3444def MVE_VSHR_immu8 : MVE_VSHR_imm<"u8", (ins shr_imm8:$imm)> {
3445  let Inst{28} = 0b1;
3446  let Inst{21-19} = 0b001;
3447}
3448
3449def MVE_VSHR_imms16 : MVE_VSHR_imm<"s16", (ins shr_imm16:$imm)> {
3450  let Inst{28} = 0b0;
3451  let Inst{21-20} = 0b01;
3452}
3453
3454def MVE_VSHR_immu16 : MVE_VSHR_imm<"u16", (ins shr_imm16:$imm)> {
3455  let Inst{28} = 0b1;
3456  let Inst{21-20} = 0b01;
3457}
3458
3459def MVE_VSHR_imms32 : MVE_VSHR_imm<"s32", (ins shr_imm32:$imm)> {
3460  let Inst{28} = 0b0;
3461  let Inst{21} = 0b1;
3462}
3463
3464def MVE_VSHR_immu32 : MVE_VSHR_imm<"u32", (ins shr_imm32:$imm)> {
3465  let Inst{28} = 0b1;
3466  let Inst{21} = 0b1;
3467}
3468
3469class MVE_VSHL_imm<string suffix, dag imm>
3470  : MVE_shift_with_imm<"vshl", suffix, (outs MQPR:$Qd),
3471                       !con((ins MQPR:$Qm), imm), "$Qd, $Qm, $imm",
3472                       vpred_r, ""> {
3473  bits<6> imm;
3474
3475  let Inst{28} = 0b0;
3476  let Inst{25-24} = 0b11;
3477  let Inst{21-16} = imm;
3478  let Inst{10-8} = 0b101;
3479}
3480
3481def MVE_VSHL_immi8 : MVE_VSHL_imm<"i8", (ins imm0_7:$imm)> {
3482  let Inst{21-19} = 0b001;
3483}
3484
3485def MVE_VSHL_immi16 : MVE_VSHL_imm<"i16", (ins imm0_15:$imm)> {
3486  let Inst{21-20} = 0b01;
3487}
3488
3489def MVE_VSHL_immi32 : MVE_VSHL_imm<"i32", (ins imm0_31:$imm)> {
3490  let Inst{21} = 0b1;
3491}
3492
3493multiclass MVE_immediate_shift_patterns_inner<
3494    MVEVectorVTInfo VTI, Operand imm_operand_type, SDNode unpred_op,
3495    Intrinsic pred_int, Instruction inst, list<int> unsignedFlag = []> {
3496
3497  def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$src), imm_operand_type:$imm)),
3498            (VTI.Vec (inst (VTI.Vec MQPR:$src), imm_operand_type:$imm))>;
3499
3500  def : Pat<(VTI.Vec !con((pred_int (VTI.Vec MQPR:$src), imm_operand_type:$imm),
3501                          !dag(pred_int, unsignedFlag, ?),
3502                          (pred_int (VTI.Pred VCCR:$mask),
3503                                   (VTI.Vec MQPR:$inactive)))),
3504            (VTI.Vec (inst (VTI.Vec MQPR:$src), imm_operand_type:$imm,
3505                           ARMVCCThen, (VTI.Pred VCCR:$mask),
3506                           (VTI.Vec MQPR:$inactive)))>;
3507}
3508
3509multiclass MVE_immediate_shift_patterns<MVEVectorVTInfo VTI,
3510                                        Operand imm_operand_type> {
3511  defm : MVE_immediate_shift_patterns_inner<VTI, imm_operand_type,
3512      ARMvshlImm, int_arm_mve_shl_imm_predicated,
3513      !cast<Instruction>("MVE_VSHL_immi" # VTI.BitsSuffix)>;
3514  defm : MVE_immediate_shift_patterns_inner<VTI, imm_operand_type,
3515      ARMvshruImm, int_arm_mve_shr_imm_predicated,
3516      !cast<Instruction>("MVE_VSHR_immu" # VTI.BitsSuffix), [1]>;
3517  defm : MVE_immediate_shift_patterns_inner<VTI, imm_operand_type,
3518      ARMvshrsImm, int_arm_mve_shr_imm_predicated,
3519      !cast<Instruction>("MVE_VSHR_imms" # VTI.BitsSuffix), [0]>;
3520}
3521
3522let Predicates = [HasMVEInt] in {
3523  defm : MVE_immediate_shift_patterns<MVE_v16i8, imm0_7>;
3524  defm : MVE_immediate_shift_patterns<MVE_v8i16, imm0_15>;
3525  defm : MVE_immediate_shift_patterns<MVE_v4i32, imm0_31>;
3526}
3527
3528// end of mve_shift instructions
3529
3530// start of MVE Floating Point instructions
3531
3532class MVE_float<string iname, string suffix, dag oops, dag iops, string ops,
3533                vpred_ops vpred, string cstr, list<dag> pattern=[]>
3534  : MVE_f<oops, iops, NoItinerary, iname, suffix, ops, vpred, cstr, pattern> {
3535  bits<4> Qm;
3536
3537  let Inst{12} = 0b0;
3538  let Inst{6} = 0b1;
3539  let Inst{5} = Qm{3};
3540  let Inst{3-1} = Qm{2-0};
3541  let Inst{0} = 0b0;
3542}
3543
3544class MVE_VRINT<string rmode, bits<3> op, string suffix, bits<2> size,
3545                list<dag> pattern=[]>
3546  : MVE_float<!strconcat("vrint", rmode), suffix, (outs MQPR:$Qd),
3547              (ins MQPR:$Qm), "$Qd, $Qm", vpred_r, "", pattern> {
3548  bits<4> Qd;
3549
3550  let Inst{28} = 0b1;
3551  let Inst{25-23} = 0b111;
3552  let Inst{22} = Qd{3};
3553  let Inst{21-20} = 0b11;
3554  let Inst{19-18} = size;
3555  let Inst{17-16} = 0b10;
3556  let Inst{15-13} = Qd{2-0};
3557  let Inst{11-10} = 0b01;
3558  let Inst{9-7} = op{2-0};
3559  let Inst{4} = 0b0;
3560  let validForTailPredication = 1;
3561
3562}
3563
3564multiclass MVE_VRINT_m<MVEVectorVTInfo VTI, string suffix, bits<3> opcode,
3565                       SDPatternOperator unpred_op> {
3566  def "": MVE_VRINT<suffix, opcode, VTI.Suffix, VTI.Size>;
3567  defvar Inst = !cast<Instruction>(NAME);
3568  defvar pred_int = !cast<Intrinsic>("int_arm_mve_vrint"#suffix#"_predicated");
3569
3570  let Predicates = [HasMVEFloat] in {
3571    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$val))),
3572              (VTI.Vec (Inst (VTI.Vec MQPR:$val)))>;
3573    def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$val), (VTI.Pred VCCR:$pred),
3574                                 (VTI.Vec MQPR:$inactive))),
3575              (VTI.Vec (Inst (VTI.Vec MQPR:$val), ARMVCCThen,
3576                             (VTI.Pred VCCR:$pred), (VTI.Vec MQPR:$inactive)))>;
3577  }
3578}
3579
3580multiclass MVE_VRINT_ops<MVEVectorVTInfo VTI> {
3581  defm N : MVE_VRINT_m<VTI, "n", 0b000, int_arm_mve_vrintn>;
3582  defm X : MVE_VRINT_m<VTI, "x", 0b001, frint>;
3583  defm A : MVE_VRINT_m<VTI, "a", 0b010, fround>;
3584  defm Z : MVE_VRINT_m<VTI, "z", 0b011, ftrunc>;
3585  defm M : MVE_VRINT_m<VTI, "m", 0b101, ffloor>;
3586  defm P : MVE_VRINT_m<VTI, "p", 0b111, fceil>;
3587}
3588
3589defm MVE_VRINTf16 : MVE_VRINT_ops<MVE_v8f16>;
3590defm MVE_VRINTf32 : MVE_VRINT_ops<MVE_v4f32>;
3591
3592class MVEFloatArithNeon<string iname, string suffix, bit size,
3593                           dag oops, dag iops, string ops,
3594                           vpred_ops vpred, string cstr, list<dag> pattern=[]>
3595  : MVE_float<iname, suffix, oops, iops, ops, vpred, cstr, pattern> {
3596  let Inst{20} = size;
3597  let Inst{16} = 0b0;
3598}
3599
3600class MVE_VMUL_fp<string iname, string suffix, bit size, list<dag> pattern=[]>
3601  : MVEFloatArithNeon<iname, suffix, size, (outs MQPR:$Qd),
3602                      (ins MQPR:$Qn, MQPR:$Qm), "$Qd, $Qn, $Qm", vpred_r, "",
3603                      pattern> {
3604  bits<4> Qd;
3605  bits<4> Qn;
3606
3607  let Inst{28} = 0b1;
3608  let Inst{25-23} = 0b110;
3609  let Inst{22} = Qd{3};
3610  let Inst{21} = 0b0;
3611  let Inst{19-17} = Qn{2-0};
3612  let Inst{15-13} = Qd{2-0};
3613  let Inst{12-8} = 0b01101;
3614  let Inst{7} = Qn{3};
3615  let Inst{4} = 0b1;
3616  let validForTailPredication = 1;
3617}
3618
3619multiclass MVE_VMULT_fp_m<string iname, bit bit_21, MVEVectorVTInfo VTI,
3620                            SDNode Op, Intrinsic PredInt> {
3621  def "" : MVE_VMUL_fp<iname, VTI.Suffix, VTI.Size{0}>;
3622  defvar Inst = !cast<Instruction>(NAME);
3623
3624  let Predicates = [HasMVEFloat] in {
3625    defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? ), !cast<Instruction>(NAME)>;
3626  }
3627}
3628
3629multiclass MVE_VMUL_fp_m<MVEVectorVTInfo VTI>
3630  : MVE_VMULT_fp_m<"vmul", 0, VTI, fmul, int_arm_mve_mul_predicated>;
3631
3632defm MVE_VMULf32 : MVE_VMUL_fp_m<MVE_v4f32>;
3633defm MVE_VMULf16 : MVE_VMUL_fp_m<MVE_v8f16>;
3634
3635class MVE_VCMLA<string suffix, bit size>
3636  : MVEFloatArithNeon<"vcmla", suffix, size, (outs MQPR:$Qd),
3637                         (ins MQPR:$Qd_src, MQPR:$Qn, MQPR:$Qm, complexrotateop:$rot),
3638                         "$Qd, $Qn, $Qm, $rot", vpred_n, "$Qd = $Qd_src", []> {
3639  bits<4> Qd;
3640  bits<4> Qn;
3641  bits<2> rot;
3642
3643  let Inst{28} = 0b1;
3644  let Inst{25} = 0b0;
3645  let Inst{24-23} = rot;
3646  let Inst{22} = Qd{3};
3647  let Inst{21} = 0b1;
3648  let Inst{19-17} = Qn{2-0};
3649  let Inst{15-13} = Qd{2-0};
3650  let Inst{12-8} = 0b01000;
3651  let Inst{7} = Qn{3};
3652  let Inst{4} = 0b0;
3653}
3654
3655multiclass MVE_VCMLA_m<MVEVectorVTInfo VTI, bit size> {
3656  def "" : MVE_VCMLA<VTI.Suffix, size>;
3657  defvar Inst = !cast<Instruction>(NAME);
3658
3659  let Predicates = [HasMVEFloat] in {
3660    def : Pat<(VTI.Vec (int_arm_mve_vcmlaq
3661                            imm:$rot, (VTI.Vec MQPR:$Qd_src),
3662                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
3663              (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src),
3664                             (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
3665                             imm:$rot))>;
3666
3667    def : Pat<(VTI.Vec (int_arm_mve_vcmlaq_predicated
3668                            imm:$rot, (VTI.Vec MQPR:$Qd_src),
3669                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
3670                            (VTI.Pred VCCR:$mask))),
3671              (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src), (VTI.Vec MQPR:$Qn),
3672                             (VTI.Vec MQPR:$Qm), imm:$rot,
3673                             ARMVCCThen, (VTI.Pred VCCR:$mask)))>;
3674
3675  }
3676}
3677
3678defm MVE_VCMLAf16 : MVE_VCMLA_m<MVE_v8f16, 0b0>;
3679defm MVE_VCMLAf32 : MVE_VCMLA_m<MVE_v4f32, 0b1>;
3680
3681class MVE_VADDSUBFMA_fp<string iname, string suffix, bit size, bit bit_4,
3682                        bit bit_8, bit bit_21, dag iops=(ins),
3683                        vpred_ops vpred=vpred_r, string cstr="",
3684                        list<dag> pattern=[]>
3685  : MVEFloatArithNeon<iname, suffix, size, (outs MQPR:$Qd),
3686                      !con(iops, (ins MQPR:$Qn, MQPR:$Qm)), "$Qd, $Qn, $Qm",
3687                      vpred, cstr, pattern> {
3688  bits<4> Qd;
3689  bits<4> Qn;
3690
3691  let Inst{28} = 0b0;
3692  let Inst{25-23} = 0b110;
3693  let Inst{22} = Qd{3};
3694  let Inst{21} = bit_21;
3695  let Inst{19-17} = Qn{2-0};
3696  let Inst{15-13} = Qd{2-0};
3697  let Inst{11-9} = 0b110;
3698  let Inst{8} = bit_8;
3699  let Inst{7} = Qn{3};
3700  let Inst{4} = bit_4;
3701  let validForTailPredication = 1;
3702}
3703
3704multiclass MVE_VFMA_fp_multi<string iname, bit fms, MVEVectorVTInfo VTI> {
3705  def "" : MVE_VADDSUBFMA_fp<iname, VTI.Suffix, VTI.Size{0}, 0b1, 0b0, fms,
3706                             (ins MQPR:$Qd_src), vpred_n, "$Qd = $Qd_src">;
3707  defvar Inst = !cast<Instruction>(NAME);
3708  defvar pred_int = int_arm_mve_fma_predicated;
3709  defvar m1   = (VTI.Vec MQPR:$m1);
3710  defvar m2   = (VTI.Vec MQPR:$m2);
3711  defvar add  = (VTI.Vec MQPR:$add);
3712  defvar pred = (VTI.Pred VCCR:$pred);
3713
3714  let Predicates = [HasMVEFloat] in {
3715    if fms then {
3716      def : Pat<(VTI.Vec (fma (fneg m1), m2, add)),
3717                (Inst $add, $m1, $m2)>;
3718      def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
3719                                  (VTI.Vec (fma (fneg m1), m2, add)),
3720                                  add)),
3721                (Inst $add, $m1, $m2, ARMVCCThen, $pred)>;
3722      def : Pat<(VTI.Vec (pred_int (fneg m1), m2, add, pred)),
3723                (Inst $add, $m1, $m2, ARMVCCThen, $pred)>;
3724      def : Pat<(VTI.Vec (pred_int m1, (fneg m2), add, pred)),
3725                (Inst $add, $m1, $m2, ARMVCCThen, $pred)>;
3726    } else {
3727      def : Pat<(VTI.Vec (fma m1, m2, add)),
3728                (Inst $add, $m1, $m2)>;
3729      def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
3730                                  (VTI.Vec (fma m1, m2, add)),
3731                                  add)),
3732                (Inst $add, $m1, $m2, ARMVCCThen, $pred)>;
3733      def : Pat<(VTI.Vec (pred_int m1, m2, add, pred)),
3734                (Inst $add, $m1, $m2, ARMVCCThen, $pred)>;
3735    }
3736  }
3737}
3738
3739defm MVE_VFMAf32 : MVE_VFMA_fp_multi<"vfma", 0, MVE_v4f32>;
3740defm MVE_VFMAf16 : MVE_VFMA_fp_multi<"vfma", 0, MVE_v8f16>;
3741defm MVE_VFMSf32 : MVE_VFMA_fp_multi<"vfms", 1, MVE_v4f32>;
3742defm MVE_VFMSf16 : MVE_VFMA_fp_multi<"vfms", 1, MVE_v8f16>;
3743
3744multiclass MVE_VADDSUB_fp_m<string iname, bit bit_21, MVEVectorVTInfo VTI,
3745                            SDNode Op, Intrinsic PredInt> {
3746  def "" : MVE_VADDSUBFMA_fp<iname, VTI.Suffix, VTI.Size{0}, 0, 1, bit_21> {
3747    let validForTailPredication = 1;
3748  }
3749  defvar Inst = !cast<Instruction>(NAME);
3750
3751  let Predicates = [HasMVEFloat] in {
3752    defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? ), !cast<Instruction>(NAME)>;
3753  }
3754}
3755
3756multiclass MVE_VADD_fp_m<MVEVectorVTInfo VTI>
3757  : MVE_VADDSUB_fp_m<"vadd", 0, VTI, fadd, int_arm_mve_add_predicated>;
3758multiclass MVE_VSUB_fp_m<MVEVectorVTInfo VTI>
3759  : MVE_VADDSUB_fp_m<"vsub", 1, VTI, fsub, int_arm_mve_sub_predicated>;
3760
3761defm MVE_VADDf32 : MVE_VADD_fp_m<MVE_v4f32>;
3762defm MVE_VADDf16 : MVE_VADD_fp_m<MVE_v8f16>;
3763
3764defm MVE_VSUBf32 : MVE_VSUB_fp_m<MVE_v4f32>;
3765defm MVE_VSUBf16 : MVE_VSUB_fp_m<MVE_v8f16>;
3766
3767class MVE_VCADD<string suffix, bit size, string cstr="">
3768  : MVEFloatArithNeon<"vcadd", suffix, size, (outs MQPR:$Qd),
3769                         (ins MQPR:$Qn, MQPR:$Qm, complexrotateopodd:$rot),
3770                         "$Qd, $Qn, $Qm, $rot", vpred_r, cstr, []> {
3771  bits<4> Qd;
3772  bits<4> Qn;
3773  bit rot;
3774
3775  let Inst{28} = 0b1;
3776  let Inst{25} = 0b0;
3777  let Inst{24} = rot;
3778  let Inst{23} = 0b1;
3779  let Inst{22} = Qd{3};
3780  let Inst{21} = 0b0;
3781  let Inst{19-17} = Qn{2-0};
3782  let Inst{15-13} = Qd{2-0};
3783  let Inst{12-8} = 0b01000;
3784  let Inst{7} = Qn{3};
3785  let Inst{4} = 0b0;
3786}
3787
3788multiclass MVE_VCADD_m<MVEVectorVTInfo VTI, bit size, string cstr=""> {
3789  def "" : MVE_VCADD<VTI.Suffix, size, cstr>;
3790  defvar Inst = !cast<Instruction>(NAME);
3791
3792  let Predicates = [HasMVEFloat] in {
3793    def : Pat<(VTI.Vec (int_arm_mve_vcaddq (i32 1),
3794                            imm:$rot, (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
3795              (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
3796                             imm:$rot))>;
3797
3798    def : Pat<(VTI.Vec (int_arm_mve_vcaddq_predicated (i32 1),
3799                            imm:$rot, (VTI.Vec MQPR:$inactive),
3800                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
3801                            (VTI.Pred VCCR:$mask))),
3802              (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
3803                             imm:$rot, ARMVCCThen, (VTI.Pred VCCR:$mask),
3804                             (VTI.Vec MQPR:$inactive)))>;
3805
3806  }
3807}
3808
3809defm MVE_VCADDf16 : MVE_VCADD_m<MVE_v8f16, 0b0>;
3810defm MVE_VCADDf32 : MVE_VCADD_m<MVE_v4f32, 0b1, "@earlyclobber $Qd">;
3811
3812class MVE_VABD_fp<string suffix, bit size>
3813  : MVE_float<"vabd", suffix, (outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm),
3814              "$Qd, $Qn, $Qm", vpred_r, ""> {
3815  bits<4> Qd;
3816  bits<4> Qn;
3817
3818  let Inst{28} = 0b1;
3819  let Inst{25-23} = 0b110;
3820  let Inst{22} = Qd{3};
3821  let Inst{21} = 0b1;
3822  let Inst{20} = size;
3823  let Inst{19-17} = Qn{2-0};
3824  let Inst{16} = 0b0;
3825  let Inst{15-13} = Qd{2-0};
3826  let Inst{11-8} = 0b1101;
3827  let Inst{7} = Qn{3};
3828  let Inst{4} = 0b0;
3829  let validForTailPredication = 1;
3830}
3831
3832multiclass MVE_VABDT_fp_m<MVEVectorVTInfo VTI,
3833                            Intrinsic unpred_int, Intrinsic pred_int> {
3834  def "" : MVE_VABD_fp<VTI.Suffix, VTI.Size{0}>;
3835  defvar Inst = !cast<Instruction>(NAME);
3836
3837  let Predicates = [HasMVEFloat] in {
3838    def : Pat<(VTI.Vec (unpred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
3839                            (i32 0))),
3840              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
3841    def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
3842                            (i32 0), (VTI.Pred VCCR:$mask),
3843                            (VTI.Vec MQPR:$inactive))),
3844              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
3845                             ARMVCCThen, (VTI.Pred VCCR:$mask),
3846                             (VTI.Vec MQPR:$inactive)))>;
3847  }
3848}
3849
3850multiclass MVE_VABD_fp_m<MVEVectorVTInfo VTI>
3851  : MVE_VABDT_fp_m<VTI, int_arm_mve_vabd, int_arm_mve_abd_predicated>;
3852
3853defm MVE_VABDf32 : MVE_VABD_fp_m<MVE_v4f32>;
3854defm MVE_VABDf16 : MVE_VABD_fp_m<MVE_v8f16>;
3855
3856let Predicates = [HasMVEFloat] in {
3857  def : Pat<(v8f16 (fabs (fsub (v8f16 MQPR:$Qm), (v8f16 MQPR:$Qn)))),
3858            (MVE_VABDf16 MQPR:$Qm, MQPR:$Qn)>;
3859  def : Pat<(v4f32 (fabs (fsub (v4f32 MQPR:$Qm), (v4f32 MQPR:$Qn)))),
3860            (MVE_VABDf32 MQPR:$Qm, MQPR:$Qn)>;
3861}
3862
3863class MVE_VCVT_fix<string suffix, bit fsi, bit U, bit op,
3864                   Operand imm_operand_type>
3865  : MVE_float<"vcvt", suffix,
3866              (outs MQPR:$Qd), (ins MQPR:$Qm, imm_operand_type:$imm6),
3867              "$Qd, $Qm, $imm6", vpred_r, "", []> {
3868  bits<4> Qd;
3869  bits<6> imm6;
3870
3871  let Inst{28} = U;
3872  let Inst{25-23} = 0b111;
3873  let Inst{22} = Qd{3};
3874  let Inst{21} = 0b1;
3875  let Inst{19-16} = imm6{3-0};
3876  let Inst{15-13} = Qd{2-0};
3877  let Inst{11-10} = 0b11;
3878  let Inst{9} = fsi;
3879  let Inst{8} = op;
3880  let Inst{7} = 0b0;
3881  let Inst{4} = 0b1;
3882
3883  let DecoderMethod = "DecodeMVEVCVTt1fp";
3884  let validForTailPredication = 1;
3885}
3886
3887class MVE_VCVT_imm_asmop<int Bits> : AsmOperandClass {
3888  let PredicateMethod = "isImmediate<1," # Bits # ">";
3889  let DiagnosticString =
3890      "MVE fixed-point immediate operand must be between 1 and " # Bits;
3891  let Name = "MVEVcvtImm" # Bits;
3892  let RenderMethod = "addImmOperands";
3893}
3894class MVE_VCVT_imm<int Bits>: Operand<i32> {
3895  let ParserMatchClass = MVE_VCVT_imm_asmop<Bits>;
3896  let EncoderMethod = "getNEONVcvtImm32OpValue";
3897  let DecoderMethod = "DecodeVCVTImmOperand";
3898}
3899
3900class MVE_VCVT_fix_f32<string suffix, bit U, bit op>
3901    : MVE_VCVT_fix<suffix, 0b1, U, op, MVE_VCVT_imm<32>> {
3902  let Inst{20} = imm6{4};
3903}
3904class MVE_VCVT_fix_f16<string suffix, bit U, bit op>
3905    : MVE_VCVT_fix<suffix, 0b0, U, op, MVE_VCVT_imm<16>> {
3906  let Inst{20} = 0b1;
3907}
3908
3909multiclass MVE_VCVT_fix_patterns<Instruction Inst, bit U, MVEVectorVTInfo DestVTI,
3910                                 MVEVectorVTInfo SrcVTI> {
3911  let Predicates = [HasMVEFloat] in {
3912    def : Pat<(DestVTI.Vec (int_arm_mve_vcvt_fix
3913                              (i32 U), (SrcVTI.Vec MQPR:$Qm), imm:$scale)),
3914              (DestVTI.Vec (Inst (SrcVTI.Vec MQPR:$Qm), imm:$scale))>;
3915    def : Pat<(DestVTI.Vec (int_arm_mve_vcvt_fix_predicated (i32 U),
3916                              (DestVTI.Vec MQPR:$inactive),
3917                              (SrcVTI.Vec MQPR:$Qm),
3918                              imm:$scale,
3919                              (DestVTI.Pred VCCR:$mask))),
3920              (DestVTI.Vec (Inst (SrcVTI.Vec MQPR:$Qm), imm:$scale,
3921                             ARMVCCThen, (DestVTI.Pred VCCR:$mask),
3922                             (DestVTI.Vec MQPR:$inactive)))>;
3923  }
3924}
3925
3926multiclass MVE_VCVT_fix_f32_m<bit U, bit op,
3927                              MVEVectorVTInfo DestVTI, MVEVectorVTInfo SrcVTI> {
3928  def "" : MVE_VCVT_fix_f32<DestVTI.Suffix#"."#SrcVTI.Suffix, U, op>;
3929  defm : MVE_VCVT_fix_patterns<!cast<Instruction>(NAME), U, DestVTI, SrcVTI>;
3930}
3931
3932multiclass MVE_VCVT_fix_f16_m<bit U, bit op,
3933                              MVEVectorVTInfo DestVTI, MVEVectorVTInfo SrcVTI> {
3934  def "" : MVE_VCVT_fix_f16<DestVTI.Suffix#"."#SrcVTI.Suffix, U, op>;
3935  defm : MVE_VCVT_fix_patterns<!cast<Instruction>(NAME), U, DestVTI, SrcVTI>;
3936}
3937
3938defm MVE_VCVTf16s16_fix : MVE_VCVT_fix_f16_m<0b0, 0b0, MVE_v8f16, MVE_v8s16>;
3939defm MVE_VCVTs16f16_fix : MVE_VCVT_fix_f16_m<0b0, 0b1, MVE_v8s16, MVE_v8f16>;
3940defm MVE_VCVTf16u16_fix : MVE_VCVT_fix_f16_m<0b1, 0b0, MVE_v8f16, MVE_v8u16>;
3941defm MVE_VCVTu16f16_fix : MVE_VCVT_fix_f16_m<0b1, 0b1, MVE_v8u16, MVE_v8f16>;
3942defm MVE_VCVTf32s32_fix : MVE_VCVT_fix_f32_m<0b0, 0b0, MVE_v4f32, MVE_v4s32>;
3943defm MVE_VCVTs32f32_fix : MVE_VCVT_fix_f32_m<0b0, 0b1, MVE_v4s32, MVE_v4f32>;
3944defm MVE_VCVTf32u32_fix : MVE_VCVT_fix_f32_m<0b1, 0b0, MVE_v4f32, MVE_v4u32>;
3945defm MVE_VCVTu32f32_fix : MVE_VCVT_fix_f32_m<0b1, 0b1, MVE_v4u32, MVE_v4f32>;
3946
3947class MVE_VCVT_fp_int_anpm<string suffix, bits<2> size, bit op, string anpm,
3948                bits<2> rm, list<dag> pattern=[]>
3949  : MVE_float<!strconcat("vcvt", anpm), suffix, (outs MQPR:$Qd),
3950              (ins MQPR:$Qm), "$Qd, $Qm", vpred_r, "", pattern> {
3951  bits<4> Qd;
3952
3953  let Inst{28} = 0b1;
3954  let Inst{25-23} = 0b111;
3955  let Inst{22} = Qd{3};
3956  let Inst{21-20} = 0b11;
3957  let Inst{19-18} = size;
3958  let Inst{17-16} = 0b11;
3959  let Inst{15-13} = Qd{2-0};
3960  let Inst{12-10} = 0b000;
3961  let Inst{9-8} = rm;
3962  let Inst{7} = op;
3963  let Inst{4} = 0b0;
3964  let validForTailPredication = 1;
3965}
3966
3967multiclass MVE_VCVT_fp_int_anpm_inner<MVEVectorVTInfo Int, MVEVectorVTInfo Flt,
3968                                      string anpm, bits<2> rm> {
3969  def "": MVE_VCVT_fp_int_anpm<Int.Suffix # "." # Flt.Suffix, Int.Size,
3970                               Int.Unsigned, anpm, rm>;
3971
3972  defvar Inst         = !cast<Instruction>(NAME);
3973  defvar IntrBaseName = "int_arm_mve_vcvt" # anpm;
3974  defvar UnpredIntr   = !cast<Intrinsic>(IntrBaseName);
3975  defvar PredIntr     = !cast<Intrinsic>(IntrBaseName # "_predicated");
3976
3977  let Predicates = [HasMVEFloat] in {
3978    def : Pat<(Int.Vec (UnpredIntr (i32 Int.Unsigned), (Flt.Vec MQPR:$in))),
3979              (Int.Vec (Inst (Flt.Vec MQPR:$in)))>;
3980
3981    def : Pat<(Int.Vec (PredIntr (i32 Int.Unsigned), (Int.Vec MQPR:$inactive),
3982                                 (Flt.Vec MQPR:$in), (Flt.Pred VCCR:$pred))),
3983              (Int.Vec (Inst (Flt.Vec MQPR:$in), ARMVCCThen,
3984                             (Flt.Pred VCCR:$pred), (Int.Vec MQPR:$inactive)))>;
3985  }
3986}
3987
3988multiclass MVE_VCVT_fp_int_anpm_outer<MVEVectorVTInfo Int,
3989                                      MVEVectorVTInfo Flt> {
3990  defm a : MVE_VCVT_fp_int_anpm_inner<Int, Flt, "a", 0b00>;
3991  defm n : MVE_VCVT_fp_int_anpm_inner<Int, Flt, "n", 0b01>;
3992  defm p : MVE_VCVT_fp_int_anpm_inner<Int, Flt, "p", 0b10>;
3993  defm m : MVE_VCVT_fp_int_anpm_inner<Int, Flt, "m", 0b11>;
3994}
3995
3996// This defines instructions such as MVE_VCVTu16f16a, with an explicit
3997// rounding-mode suffix on the mnemonic. The class below will define
3998// the bare MVE_VCVTu16f16 (with implied rounding toward zero).
3999defm MVE_VCVTs16f16 : MVE_VCVT_fp_int_anpm_outer<MVE_v8s16, MVE_v8f16>;
4000defm MVE_VCVTu16f16 : MVE_VCVT_fp_int_anpm_outer<MVE_v8u16, MVE_v8f16>;
4001defm MVE_VCVTs32f32 : MVE_VCVT_fp_int_anpm_outer<MVE_v4s32, MVE_v4f32>;
4002defm MVE_VCVTu32f32 : MVE_VCVT_fp_int_anpm_outer<MVE_v4u32, MVE_v4f32>;
4003
4004class MVE_VCVT_fp_int<string suffix, bits<2> size, bit toint, bit unsigned,
4005                      list<dag> pattern=[]>
4006  : MVE_float<"vcvt", suffix, (outs MQPR:$Qd),
4007              (ins MQPR:$Qm), "$Qd, $Qm", vpred_r, "", pattern> {
4008  bits<4> Qd;
4009
4010  let Inst{28} = 0b1;
4011  let Inst{25-23} = 0b111;
4012  let Inst{22} = Qd{3};
4013  let Inst{21-20} = 0b11;
4014  let Inst{19-18} = size;
4015  let Inst{17-16} = 0b11;
4016  let Inst{15-13} = Qd{2-0};
4017  let Inst{12-9} = 0b0011;
4018  let Inst{8} = toint;
4019  let Inst{7} = unsigned;
4020  let Inst{4} = 0b0;
4021  let validForTailPredication = 1;
4022}
4023
4024multiclass MVE_VCVT_fp_int_m<MVEVectorVTInfo Dest, MVEVectorVTInfo Src,
4025                             SDNode unpred_op> {
4026  defvar Unsigned = !or(!eq(Dest.SuffixLetter,"u"), !eq(Src.SuffixLetter,"u"));
4027  defvar ToInt = !eq(Src.SuffixLetter,"f");
4028
4029  def "" : MVE_VCVT_fp_int<Dest.Suffix # "." # Src.Suffix, Dest.Size,
4030                           ToInt, Unsigned>;
4031  defvar Inst = !cast<Instruction>(NAME);
4032
4033  let Predicates = [HasMVEFloat] in {
4034    def : Pat<(Dest.Vec (unpred_op (Src.Vec MQPR:$src))),
4035              (Dest.Vec (Inst (Src.Vec MQPR:$src)))>;
4036    def : Pat<(Dest.Vec (int_arm_mve_vcvt_fp_int_predicated
4037                             (Src.Vec MQPR:$src), (i32 Unsigned),
4038                             (Src.Pred VCCR:$mask), (Dest.Vec MQPR:$inactive))),
4039              (Dest.Vec (Inst (Src.Vec MQPR:$src), ARMVCCThen,
4040                              (Src.Pred VCCR:$mask),
4041                              (Dest.Vec MQPR:$inactive)))>;
4042  }
4043}
4044// The unsuffixed VCVT for float->int implicitly rounds toward zero,
4045// which I reflect here in the llvm instruction names
4046defm MVE_VCVTs16f16z : MVE_VCVT_fp_int_m<MVE_v8s16, MVE_v8f16, fp_to_sint>;
4047defm MVE_VCVTu16f16z : MVE_VCVT_fp_int_m<MVE_v8u16, MVE_v8f16, fp_to_uint>;
4048defm MVE_VCVTs32f32z : MVE_VCVT_fp_int_m<MVE_v4s32, MVE_v4f32, fp_to_sint>;
4049defm MVE_VCVTu32f32z : MVE_VCVT_fp_int_m<MVE_v4u32, MVE_v4f32, fp_to_uint>;
4050// Whereas VCVT for int->float rounds to nearest
4051defm MVE_VCVTf16s16n : MVE_VCVT_fp_int_m<MVE_v8f16, MVE_v8s16, sint_to_fp>;
4052defm MVE_VCVTf16u16n : MVE_VCVT_fp_int_m<MVE_v8f16, MVE_v8u16, uint_to_fp>;
4053defm MVE_VCVTf32s32n : MVE_VCVT_fp_int_m<MVE_v4f32, MVE_v4s32, sint_to_fp>;
4054defm MVE_VCVTf32u32n : MVE_VCVT_fp_int_m<MVE_v4f32, MVE_v4u32, uint_to_fp>;
4055
4056class MVE_VABSNEG_fp<string iname, string suffix, bits<2> size, bit negate,
4057                   list<dag> pattern=[]>
4058  : MVE_float<iname, suffix, (outs MQPR:$Qd),
4059              (ins MQPR:$Qm), "$Qd, $Qm", vpred_r, "", pattern> {
4060  bits<4> Qd;
4061
4062  let Inst{28} = 0b1;
4063  let Inst{25-23} = 0b111;
4064  let Inst{22} = Qd{3};
4065  let Inst{21-20} = 0b11;
4066  let Inst{19-18} = size;
4067  let Inst{17-16} = 0b01;
4068  let Inst{15-13} = Qd{2-0};
4069  let Inst{11-8} = 0b0111;
4070  let Inst{7} = negate;
4071  let Inst{4} = 0b0;
4072  let validForTailPredication = 1;
4073}
4074
4075multiclass MVE_VABSNEG_fp_m<string iname, SDNode unpred_op, Intrinsic pred_int,
4076                            MVEVectorVTInfo VTI, bit opcode> {
4077  def "" : MVE_VABSNEG_fp<iname, VTI.Suffix, VTI.Size, opcode>;
4078  defvar Inst = !cast<Instruction>(NAME);
4079
4080  let Predicates = [HasMVEInt] in {
4081    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$v))),
4082              (VTI.Vec (Inst $v))>;
4083    def : Pat<(VTI.Vec (pred_int  (VTI.Vec MQPR:$v), (VTI.Pred VCCR:$mask),
4084                                  (VTI.Vec MQPR:$inactive))),
4085              (VTI.Vec (Inst $v, ARMVCCThen, $mask, $inactive))>;
4086  }
4087}
4088
4089defm MVE_VABSf16 : MVE_VABSNEG_fp_m<"vabs", fabs, int_arm_mve_abs_predicated,
4090                                    MVE_v8f16, 0>;
4091defm MVE_VABSf32 : MVE_VABSNEG_fp_m<"vabs", fabs, int_arm_mve_abs_predicated,
4092                                    MVE_v4f32, 0>;
4093defm MVE_VNEGf16 : MVE_VABSNEG_fp_m<"vneg", fneg, int_arm_mve_neg_predicated,
4094                                    MVE_v8f16, 1>;
4095defm MVE_VNEGf32 : MVE_VABSNEG_fp_m<"vneg", fneg, int_arm_mve_neg_predicated,
4096                                    MVE_v4f32, 1>;
4097
4098class MVE_VMAXMINNMA<string iname, string suffix, bit size, bit bit_12,
4099                     list<dag> pattern=[]>
4100  : MVE_f<(outs MQPR:$Qd), (ins MQPR:$Qd_src, MQPR:$Qm),
4101          NoItinerary, iname, suffix, "$Qd, $Qm", vpred_n, "$Qd = $Qd_src",
4102          pattern> {
4103  bits<4> Qd;
4104  bits<4> Qm;
4105
4106  let Inst{28} = size;
4107  let Inst{25-23} = 0b100;
4108  let Inst{22} = Qd{3};
4109  let Inst{21-16} = 0b111111;
4110  let Inst{15-13} = Qd{2-0};
4111  let Inst{12} = bit_12;
4112  let Inst{11-6} = 0b111010;
4113  let Inst{5} = Qm{3};
4114  let Inst{4} = 0b0;
4115  let Inst{3-1} = Qm{2-0};
4116  let Inst{0} = 0b1;
4117
4118  let isCommutable = 1;
4119}
4120
4121multiclass MVE_VMAXMINNMA_m<string iname, MVEVectorVTInfo VTI,
4122                      SDNode unpred_op, Intrinsic pred_int,
4123                      bit bit_12> {
4124  def "" : MVE_VMAXMINNMA<iname, VTI.Suffix, VTI.Size{0}, bit_12>;
4125  defvar Inst = !cast<Instruction>(NAME);
4126
4127  let Predicates = [HasMVEInt] in {
4128    // Unpredicated v(max|min)nma
4129    def : Pat<(VTI.Vec (unpred_op (fabs (VTI.Vec MQPR:$Qd)),
4130                                  (fabs (VTI.Vec MQPR:$Qm)))),
4131              (VTI.Vec (Inst (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm)))>;
4132
4133    // Predicated v(max|min)nma
4134    def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm),
4135                            (VTI.Pred VCCR:$mask))),
4136              (VTI.Vec (Inst (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm),
4137                            ARMVCCThen, (VTI.Pred VCCR:$mask)))>;
4138  }
4139}
4140
4141multiclass MVE_VMAXNMA<MVEVectorVTInfo VTI, bit bit_12>
4142  : MVE_VMAXMINNMA_m<"vmaxnma", VTI, fmaxnum, int_arm_mve_vmaxnma_predicated, bit_12>;
4143
4144defm MVE_VMAXNMAf32 : MVE_VMAXNMA<MVE_v4f32, 0b0>;
4145defm MVE_VMAXNMAf16 : MVE_VMAXNMA<MVE_v8f16, 0b0>;
4146
4147multiclass MVE_VMINNMA<MVEVectorVTInfo VTI, bit bit_12>
4148  : MVE_VMAXMINNMA_m<"vminnma", VTI, fminnum, int_arm_mve_vminnma_predicated, bit_12>;
4149
4150defm MVE_VMINNMAf32 : MVE_VMINNMA<MVE_v4f32, 0b1>;
4151defm MVE_VMINNMAf16 : MVE_VMINNMA<MVE_v8f16, 0b1>;
4152
4153// end of MVE Floating Point instructions
4154
4155// start of MVE compares
4156
4157class MVE_VCMPqq<string suffix, bit bit_28, bits<2> bits_21_20,
4158                 VCMPPredicateOperand predtype, list<dag> pattern=[]>
4159  : MVE_p<(outs VCCR:$P0), (ins MQPR:$Qn, MQPR:$Qm, predtype:$fc),
4160           NoItinerary, "vcmp", suffix, "$fc, $Qn, $Qm", vpred_n, "", pattern> {
4161  // Base class for comparing two vector registers
4162  bits<3> fc;
4163  bits<4> Qn;
4164  bits<4> Qm;
4165
4166  let Inst{28} = bit_28;
4167  let Inst{25-22} = 0b1000;
4168  let Inst{21-20} = bits_21_20;
4169  let Inst{19-17} = Qn{2-0};
4170  let Inst{16-13} = 0b1000;
4171  let Inst{12} = fc{2};
4172  let Inst{11-8} = 0b1111;
4173  let Inst{7} = fc{0};
4174  let Inst{6} = 0b0;
4175  let Inst{5} = Qm{3};
4176  let Inst{4} = 0b0;
4177  let Inst{3-1} = Qm{2-0};
4178  let Inst{0} = fc{1};
4179
4180  let Constraints = "";
4181
4182  // We need a custom decoder method for these instructions because of
4183  // the output VCCR operand, which isn't encoded in the instruction
4184  // bits anywhere (there is only one choice for it) but has to be
4185  // included in the MC operands so that codegen will be able to track
4186  // its data flow between instructions, spill/reload it when
4187  // necessary, etc. There seems to be no way to get the Tablegen
4188  // decoder to emit an operand that isn't affected by any instruction
4189  // bit.
4190  let DecoderMethod = "DecodeMVEVCMP<false," # predtype.DecoderMethod # ">";
4191  let validForTailPredication = 1;
4192}
4193
4194class MVE_VCMPqqf<string suffix, bit size>
4195    : MVE_VCMPqq<suffix, size, 0b11, pred_basic_fp> {
4196  let Predicates = [HasMVEFloat];
4197}
4198
4199class MVE_VCMPqqi<string suffix, bits<2> size>
4200    : MVE_VCMPqq<suffix, 0b1, size, pred_basic_i> {
4201  let Inst{12} = 0b0;
4202  let Inst{0} = 0b0;
4203}
4204
4205class MVE_VCMPqqu<string suffix, bits<2> size>
4206    : MVE_VCMPqq<suffix, 0b1, size, pred_basic_u> {
4207  let Inst{12} = 0b0;
4208  let Inst{0} = 0b1;
4209}
4210
4211class MVE_VCMPqqs<string suffix, bits<2> size>
4212    : MVE_VCMPqq<suffix, 0b1, size, pred_basic_s> {
4213  let Inst{12} = 0b1;
4214}
4215
4216def MVE_VCMPf32 : MVE_VCMPqqf<"f32", 0b0>;
4217def MVE_VCMPf16 : MVE_VCMPqqf<"f16", 0b1>;
4218
4219def MVE_VCMPi8  : MVE_VCMPqqi<"i8",  0b00>;
4220def MVE_VCMPi16 : MVE_VCMPqqi<"i16", 0b01>;
4221def MVE_VCMPi32 : MVE_VCMPqqi<"i32", 0b10>;
4222
4223def MVE_VCMPu8  : MVE_VCMPqqu<"u8",  0b00>;
4224def MVE_VCMPu16 : MVE_VCMPqqu<"u16", 0b01>;
4225def MVE_VCMPu32 : MVE_VCMPqqu<"u32", 0b10>;
4226
4227def MVE_VCMPs8  : MVE_VCMPqqs<"s8",  0b00>;
4228def MVE_VCMPs16 : MVE_VCMPqqs<"s16", 0b01>;
4229def MVE_VCMPs32 : MVE_VCMPqqs<"s32", 0b10>;
4230
4231class MVE_VCMPqr<string suffix, bit bit_28, bits<2> bits_21_20,
4232                 VCMPPredicateOperand predtype, list<dag> pattern=[]>
4233  : MVE_p<(outs VCCR:$P0), (ins MQPR:$Qn, GPRwithZR:$Rm, predtype:$fc),
4234           NoItinerary, "vcmp", suffix, "$fc, $Qn, $Rm", vpred_n, "", pattern> {
4235  // Base class for comparing a vector register with a scalar
4236  bits<3> fc;
4237  bits<4> Qn;
4238  bits<4> Rm;
4239
4240  let Inst{28} = bit_28;
4241  let Inst{25-22} = 0b1000;
4242  let Inst{21-20} = bits_21_20;
4243  let Inst{19-17} = Qn{2-0};
4244  let Inst{16-13} = 0b1000;
4245  let Inst{12} = fc{2};
4246  let Inst{11-8} = 0b1111;
4247  let Inst{7} = fc{0};
4248  let Inst{6} = 0b1;
4249  let Inst{5} = fc{1};
4250  let Inst{4} = 0b0;
4251  let Inst{3-0} = Rm{3-0};
4252
4253  let Constraints = "";
4254  // Custom decoder method, for the same reason as MVE_VCMPqq
4255  let DecoderMethod = "DecodeMVEVCMP<true," # predtype.DecoderMethod # ">";
4256  let validForTailPredication = 1;
4257}
4258
4259class MVE_VCMPqrf<string suffix, bit size>
4260    : MVE_VCMPqr<suffix, size, 0b11, pred_basic_fp> {
4261  let Predicates = [HasMVEFloat];
4262}
4263
4264class MVE_VCMPqri<string suffix, bits<2> size>
4265    : MVE_VCMPqr<suffix, 0b1, size, pred_basic_i> {
4266  let Inst{12} = 0b0;
4267  let Inst{5} = 0b0;
4268}
4269
4270class MVE_VCMPqru<string suffix, bits<2> size>
4271    : MVE_VCMPqr<suffix, 0b1, size, pred_basic_u> {
4272  let Inst{12} = 0b0;
4273  let Inst{5} = 0b1;
4274}
4275
4276class MVE_VCMPqrs<string suffix, bits<2> size>
4277    : MVE_VCMPqr<suffix, 0b1, size, pred_basic_s> {
4278  let Inst{12} = 0b1;
4279}
4280
4281def MVE_VCMPf32r : MVE_VCMPqrf<"f32", 0b0>;
4282def MVE_VCMPf16r : MVE_VCMPqrf<"f16", 0b1>;
4283
4284def MVE_VCMPi8r  : MVE_VCMPqri<"i8",  0b00>;
4285def MVE_VCMPi16r : MVE_VCMPqri<"i16", 0b01>;
4286def MVE_VCMPi32r : MVE_VCMPqri<"i32", 0b10>;
4287
4288def MVE_VCMPu8r  : MVE_VCMPqru<"u8",  0b00>;
4289def MVE_VCMPu16r : MVE_VCMPqru<"u16", 0b01>;
4290def MVE_VCMPu32r : MVE_VCMPqru<"u32", 0b10>;
4291
4292def MVE_VCMPs8r  : MVE_VCMPqrs<"s8",  0b00>;
4293def MVE_VCMPs16r : MVE_VCMPqrs<"s16", 0b01>;
4294def MVE_VCMPs32r : MVE_VCMPqrs<"s32", 0b10>;
4295
4296multiclass unpred_vcmp_z<string suffix, PatLeaf fc> {
4297  def i8  : Pat<(v16i1 (ARMvcmpz (v16i8 MQPR:$v1), fc)),
4298                (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8r") (v16i8 MQPR:$v1), ZR, fc))>;
4299  def i16 : Pat<(v8i1 (ARMvcmpz (v8i16 MQPR:$v1), fc)),
4300                (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16r") (v8i16 MQPR:$v1), ZR, fc))>;
4301  def i32 : Pat<(v4i1 (ARMvcmpz (v4i32 MQPR:$v1), fc)),
4302                (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32r") (v4i32 MQPR:$v1), ZR, fc))>;
4303
4304  def : Pat<(v16i1 (and (v16i1 VCCR:$p1), (v16i1 (ARMvcmpz (v16i8 MQPR:$v1), fc)))),
4305            (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8r") (v16i8 MQPR:$v1), ZR, fc, ARMVCCThen, VCCR:$p1))>;
4306  def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmpz (v8i16 MQPR:$v1), fc)))),
4307            (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16r") (v8i16 MQPR:$v1), ZR, fc, ARMVCCThen, VCCR:$p1))>;
4308  def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmpz (v4i32 MQPR:$v1), fc)))),
4309            (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32r") (v4i32 MQPR:$v1), ZR, fc, ARMVCCThen, VCCR:$p1))>;
4310}
4311
4312multiclass unpred_vcmp_r<string suffix, PatLeaf fc> {
4313  def i8  : Pat<(v16i1 (ARMvcmp (v16i8 MQPR:$v1), (v16i8 MQPR:$v2), fc)),
4314                (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8") (v16i8 MQPR:$v1), (v16i8 MQPR:$v2), fc))>;
4315  def i16 : Pat<(v8i1 (ARMvcmp (v8i16 MQPR:$v1), (v8i16 MQPR:$v2), fc)),
4316                (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16") (v8i16 MQPR:$v1), (v8i16 MQPR:$v2), fc))>;
4317  def i32 : Pat<(v4i1 (ARMvcmp (v4i32 MQPR:$v1), (v4i32 MQPR:$v2), fc)),
4318                (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32") (v4i32 MQPR:$v1), (v4i32 MQPR:$v2), fc))>;
4319
4320  def i8r  : Pat<(v16i1 (ARMvcmp (v16i8 MQPR:$v1), (v16i8 (ARMvdup rGPR:$v2)), fc)),
4321                 (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8r") (v16i8 MQPR:$v1), (i32 rGPR:$v2), fc))>;
4322  def i16r : Pat<(v8i1 (ARMvcmp (v8i16 MQPR:$v1), (v8i16 (ARMvdup rGPR:$v2)), fc)),
4323                 (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16r") (v8i16 MQPR:$v1), (i32 rGPR:$v2), fc))>;
4324  def i32r : Pat<(v4i1 (ARMvcmp (v4i32 MQPR:$v1), (v4i32 (ARMvdup rGPR:$v2)), fc)),
4325                 (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32r") (v4i32 MQPR:$v1), (i32 rGPR:$v2), fc))>;
4326
4327  def : Pat<(v16i1 (and (v16i1 VCCR:$p1), (v16i1 (ARMvcmp (v16i8 MQPR:$v1), (v16i8 MQPR:$v2), fc)))),
4328            (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8") (v16i8 MQPR:$v1), (v16i8 MQPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
4329  def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmp (v8i16 MQPR:$v1), (v8i16 MQPR:$v2), fc)))),
4330            (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16") (v8i16 MQPR:$v1), (v8i16 MQPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
4331  def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmp (v4i32 MQPR:$v1), (v4i32 MQPR:$v2), fc)))),
4332            (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32") (v4i32 MQPR:$v1), (v4i32 MQPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
4333
4334  def : Pat<(v16i1 (and (v16i1 VCCR:$p1), (v16i1 (ARMvcmp (v16i8 MQPR:$v1), (v16i8 (ARMvdup rGPR:$v2)), fc)))),
4335            (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8r") (v16i8 MQPR:$v1), (i32 rGPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
4336  def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmp (v8i16 MQPR:$v1), (v8i16 (ARMvdup rGPR:$v2)), fc)))),
4337            (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16r") (v8i16 MQPR:$v1), (i32 rGPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
4338  def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmp (v4i32 MQPR:$v1), (v4i32 (ARMvdup rGPR:$v2)), fc)))),
4339            (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32r") (v4i32 MQPR:$v1), (i32 rGPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
4340}
4341
4342multiclass unpred_vcmpf_z<PatLeaf fc> {
4343  def f16 : Pat<(v8i1 (ARMvcmpz (v8f16 MQPR:$v1), fc)),
4344                (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), ZR, fc))>;
4345  def f32 : Pat<(v4i1 (ARMvcmpz (v4f32 MQPR:$v1), fc)),
4346                (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), ZR, fc))>;
4347
4348  def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmpz (v8f16 MQPR:$v1), fc)))),
4349            (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), ZR, fc, ARMVCCThen, VCCR:$p1))>;
4350  def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmpz (v4f32 MQPR:$v1), fc)))),
4351            (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), ZR, fc, ARMVCCThen, VCCR:$p1))>;
4352}
4353
4354multiclass unpred_vcmpf_r<PatLeaf fc> {
4355  def : Pat<(v8i1 (ARMvcmp (v8f16 MQPR:$v1), (v8f16 MQPR:$v2), fc)),
4356            (v8i1 (MVE_VCMPf16 (v8f16 MQPR:$v1), (v8f16 MQPR:$v2), fc))>;
4357  def : Pat<(v4i1 (ARMvcmp (v4f32 MQPR:$v1), (v4f32 MQPR:$v2), fc)),
4358            (v4i1 (MVE_VCMPf32 (v4f32 MQPR:$v1), (v4f32 MQPR:$v2), fc))>;
4359
4360  def : Pat<(v8i1 (ARMvcmp (v8f16 MQPR:$v1), (v8f16 (ARMvdup rGPR:$v2)), fc)),
4361            (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), (i32 rGPR:$v2), fc))>;
4362  def : Pat<(v4i1 (ARMvcmp (v4f32 MQPR:$v1), (v4f32 (ARMvdup rGPR:$v2)), fc)),
4363            (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), (i32 rGPR:$v2), fc))>;
4364
4365  def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmp (v8f16 MQPR:$v1), (v8f16 MQPR:$v2), fc)))),
4366            (v8i1 (MVE_VCMPf16 (v8f16 MQPR:$v1), (v8f16 MQPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
4367  def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmp (v4f32 MQPR:$v1), (v4f32 MQPR:$v2), fc)))),
4368            (v4i1 (MVE_VCMPf32 (v4f32 MQPR:$v1), (v4f32 MQPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
4369
4370  def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmp (v8f16 MQPR:$v1), (v8f16 (ARMvdup rGPR:$v2)), fc)))),
4371            (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), (i32 rGPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
4372  def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmp (v4f32 MQPR:$v1), (v4f32 (ARMvdup rGPR:$v2)), fc)))),
4373            (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), (i32 rGPR:$v2), fc, ARMVCCThen, VCCR:$p1))>;
4374}
4375
4376let Predicates = [HasMVEInt] in {
4377  defm MVE_VCEQZ  : unpred_vcmp_z<"i", ARMCCeq>;
4378  defm MVE_VCNEZ  : unpred_vcmp_z<"i", ARMCCne>;
4379  defm MVE_VCGEZ  : unpred_vcmp_z<"s", ARMCCge>;
4380  defm MVE_VCLTZ  : unpred_vcmp_z<"s", ARMCClt>;
4381  defm MVE_VCGTZ  : unpred_vcmp_z<"s", ARMCCgt>;
4382  defm MVE_VCLEZ  : unpred_vcmp_z<"s", ARMCCle>;
4383  defm MVE_VCGTUZ : unpred_vcmp_z<"u", ARMCChi>;
4384  defm MVE_VCGEUZ : unpred_vcmp_z<"u", ARMCChs>;
4385
4386  defm MVE_VCEQ   : unpred_vcmp_r<"i", ARMCCeq>;
4387  defm MVE_VCNE   : unpred_vcmp_r<"i", ARMCCne>;
4388  defm MVE_VCGE   : unpred_vcmp_r<"s", ARMCCge>;
4389  defm MVE_VCLT   : unpred_vcmp_r<"s", ARMCClt>;
4390  defm MVE_VCGT   : unpred_vcmp_r<"s", ARMCCgt>;
4391  defm MVE_VCLE   : unpred_vcmp_r<"s", ARMCCle>;
4392  defm MVE_VCGTU  : unpred_vcmp_r<"u", ARMCChi>;
4393  defm MVE_VCGEU  : unpred_vcmp_r<"u", ARMCChs>;
4394}
4395
4396let Predicates = [HasMVEFloat] in {
4397  defm MVE_VFCEQZ  : unpred_vcmpf_z<ARMCCeq>;
4398  defm MVE_VFCNEZ  : unpred_vcmpf_z<ARMCCne>;
4399  defm MVE_VFCGEZ  : unpred_vcmpf_z<ARMCCge>;
4400  defm MVE_VFCLTZ  : unpred_vcmpf_z<ARMCClt>;
4401  defm MVE_VFCGTZ  : unpred_vcmpf_z<ARMCCgt>;
4402  defm MVE_VFCLEZ  : unpred_vcmpf_z<ARMCCle>;
4403
4404  defm MVE_VFCEQ   : unpred_vcmpf_r<ARMCCeq>;
4405  defm MVE_VFCNE   : unpred_vcmpf_r<ARMCCne>;
4406  defm MVE_VFCGE   : unpred_vcmpf_r<ARMCCge>;
4407  defm MVE_VFCLT   : unpred_vcmpf_r<ARMCClt>;
4408  defm MVE_VFCGT   : unpred_vcmpf_r<ARMCCgt>;
4409  defm MVE_VFCLE   : unpred_vcmpf_r<ARMCCle>;
4410}
4411
4412
4413// Extra "worst case" and/or/xor patterns, going into and out of GRP
4414multiclass two_predops<SDPatternOperator opnode, Instruction insn> {
4415  def v16i1 : Pat<(v16i1 (opnode (v16i1 VCCR:$p1), (v16i1 VCCR:$p2))),
4416                  (v16i1 (COPY_TO_REGCLASS
4417                           (insn (i32 (COPY_TO_REGCLASS (v16i1 VCCR:$p1), rGPR)),
4418                                 (i32 (COPY_TO_REGCLASS (v16i1 VCCR:$p2), rGPR))),
4419                           VCCR))>;
4420  def v8i1  : Pat<(v8i1 (opnode (v8i1 VCCR:$p1), (v8i1 VCCR:$p2))),
4421                  (v8i1 (COPY_TO_REGCLASS
4422                          (insn (i32 (COPY_TO_REGCLASS (v8i1 VCCR:$p1), rGPR)),
4423                                (i32 (COPY_TO_REGCLASS (v8i1 VCCR:$p2), rGPR))),
4424                          VCCR))>;
4425  def v4i1  : Pat<(v4i1 (opnode (v4i1 VCCR:$p1), (v4i1 VCCR:$p2))),
4426                  (v4i1 (COPY_TO_REGCLASS
4427                          (insn (i32 (COPY_TO_REGCLASS (v4i1 VCCR:$p1), rGPR)),
4428                                (i32 (COPY_TO_REGCLASS (v4i1 VCCR:$p2), rGPR))),
4429                          VCCR))>;
4430}
4431
4432let Predicates = [HasMVEInt] in {
4433  defm POR    : two_predops<or,  t2ORRrr>;
4434  defm PAND   : two_predops<and, t2ANDrr>;
4435  defm PEOR   : two_predops<xor, t2EORrr>;
4436}
4437
4438// Occasionally we need to cast between a i32 and a boolean vector, for
4439// example when moving between rGPR and VPR.P0 as part of predicate vector
4440// shuffles. We also sometimes need to cast between different predicate
4441// vector types (v4i1<>v8i1, etc.) also as part of lowering vector shuffles.
4442def predicate_cast : SDNode<"ARMISD::PREDICATE_CAST", SDTUnaryOp>;
4443
4444def load_align4 : PatFrag<(ops node:$ptr), (load node:$ptr), [{
4445  return cast<LoadSDNode>(N)->getAlignment() >= 4;
4446}]>;
4447
4448let Predicates = [HasMVEInt] in {
4449  foreach VT = [ v4i1, v8i1, v16i1 ] in {
4450    def : Pat<(i32 (predicate_cast (VT VCCR:$src))),
4451              (i32 (COPY_TO_REGCLASS (VT VCCR:$src), VCCR))>;
4452    def : Pat<(VT  (predicate_cast (i32 VCCR:$src))),
4453              (VT  (COPY_TO_REGCLASS (i32 VCCR:$src), VCCR))>;
4454
4455    foreach VT2 = [ v4i1, v8i1, v16i1 ] in
4456      def : Pat<(VT  (predicate_cast (VT2 VCCR:$src))),
4457                (VT  (COPY_TO_REGCLASS (VT2 VCCR:$src), VCCR))>;
4458  }
4459
4460  // If we happen to be casting from a load we can convert that straight
4461  // into a predicate load, so long as the load is of the correct type.
4462  foreach VT = [ v4i1, v8i1, v16i1 ] in {
4463    def : Pat<(VT (predicate_cast (i32 (load_align4 taddrmode_imm7<2>:$addr)))),
4464              (VT (VLDR_P0_off taddrmode_imm7<2>:$addr))>;
4465  }
4466
4467  // Here we match the specific SDNode type 'ARMVectorRegCastImpl'
4468  // rather than the more general 'ARMVectorRegCast' which would also
4469  // match some bitconverts. If we use the latter in cases where the
4470  // input and output types are the same, the bitconvert gets elided
4471  // and we end up generating a nonsense match of nothing.
4472
4473  foreach VT = [ v16i8, v8i16, v8f16, v4i32, v4f32, v2i64, v2f64 ] in
4474    foreach VT2 = [ v16i8, v8i16, v8f16, v4i32, v4f32, v2i64, v2f64 ] in
4475      def : Pat<(VT (ARMVectorRegCastImpl (VT2 MQPR:$src))),
4476                (VT MQPR:$src)>;
4477}
4478
4479// end of MVE compares
4480
4481// start of MVE_qDest_qSrc
4482
4483class MVE_qDest_qSrc<string iname, string suffix, dag oops, dag iops,
4484                     string ops, vpred_ops vpred, string cstr,
4485                     list<dag> pattern=[]>
4486  : MVE_p<oops, iops, NoItinerary, iname, suffix,
4487          ops, vpred, cstr, pattern> {
4488  bits<4> Qd;
4489  bits<4> Qm;
4490
4491  let Inst{25-23} = 0b100;
4492  let Inst{22} = Qd{3};
4493  let Inst{15-13} = Qd{2-0};
4494  let Inst{11-9} = 0b111;
4495  let Inst{6} = 0b0;
4496  let Inst{5} = Qm{3};
4497  let Inst{4} = 0b0;
4498  let Inst{3-1} = Qm{2-0};
4499}
4500
4501class MVE_VQxDMLxDH<string iname, bit exch, bit round, bit subtract,
4502                    string suffix, bits<2> size, string cstr="", list<dag> pattern=[]>
4503  : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
4504                   (ins MQPR:$Qd_src, MQPR:$Qn, MQPR:$Qm), "$Qd, $Qn, $Qm",
4505                   vpred_n, "$Qd = $Qd_src"#cstr, pattern> {
4506  bits<4> Qn;
4507
4508  let Inst{28} = subtract;
4509  let Inst{21-20} = size;
4510  let Inst{19-17} = Qn{2-0};
4511  let Inst{16} = 0b0;
4512  let Inst{12} = exch;
4513  let Inst{8} = 0b0;
4514  let Inst{7} = Qn{3};
4515  let Inst{0} = round;
4516}
4517
4518multiclass MVE_VQxDMLxDH_p<string iname, bit exch, bit round, bit subtract,
4519                           MVEVectorVTInfo VTI> {
4520  def "": MVE_VQxDMLxDH<iname, exch, round, subtract, VTI.Suffix, VTI.Size,
4521                        !if(!eq(VTI.LaneBits, 32), ",@earlyclobber $Qd", "")>;
4522  defvar Inst = !cast<Instruction>(NAME);
4523  defvar ConstParams = (? (i32 exch), (i32 round), (i32 subtract));
4524  defvar unpred_intr = int_arm_mve_vqdmlad;
4525  defvar pred_intr = int_arm_mve_vqdmlad_predicated;
4526
4527  def : Pat<(VTI.Vec !con((unpred_intr (VTI.Vec MQPR:$a), (VTI.Vec MQPR:$b),
4528                                       (VTI.Vec MQPR:$c)), ConstParams)),
4529            (VTI.Vec (Inst (VTI.Vec MQPR:$a), (VTI.Vec MQPR:$b),
4530                           (VTI.Vec MQPR:$c)))>;
4531  def : Pat<(VTI.Vec !con((pred_intr (VTI.Vec MQPR:$a), (VTI.Vec MQPR:$b),
4532                                     (VTI.Vec MQPR:$c)), ConstParams,
4533                          (? (VTI.Pred VCCR:$pred)))),
4534            (VTI.Vec (Inst (VTI.Vec MQPR:$a), (VTI.Vec MQPR:$b),
4535                           (VTI.Vec MQPR:$c),
4536                           ARMVCCThen, (VTI.Pred VCCR:$pred)))>;
4537}
4538
4539multiclass MVE_VQxDMLxDH_multi<string iname, bit exch,
4540                               bit round, bit subtract> {
4541  defm s8  : MVE_VQxDMLxDH_p<iname, exch, round, subtract, MVE_v16s8>;
4542  defm s16 : MVE_VQxDMLxDH_p<iname, exch, round, subtract, MVE_v8s16>;
4543  defm s32 : MVE_VQxDMLxDH_p<iname, exch, round, subtract, MVE_v4s32>;
4544}
4545
4546defm MVE_VQDMLADH   : MVE_VQxDMLxDH_multi<"vqdmladh",   0b0, 0b0, 0b0>;
4547defm MVE_VQDMLADHX  : MVE_VQxDMLxDH_multi<"vqdmladhx",  0b1, 0b0, 0b0>;
4548defm MVE_VQRDMLADH  : MVE_VQxDMLxDH_multi<"vqrdmladh",  0b0, 0b1, 0b0>;
4549defm MVE_VQRDMLADHX : MVE_VQxDMLxDH_multi<"vqrdmladhx", 0b1, 0b1, 0b0>;
4550defm MVE_VQDMLSDH   : MVE_VQxDMLxDH_multi<"vqdmlsdh",   0b0, 0b0, 0b1>;
4551defm MVE_VQDMLSDHX  : MVE_VQxDMLxDH_multi<"vqdmlsdhx",  0b1, 0b0, 0b1>;
4552defm MVE_VQRDMLSDH  : MVE_VQxDMLxDH_multi<"vqrdmlsdh",  0b0, 0b1, 0b1>;
4553defm MVE_VQRDMLSDHX : MVE_VQxDMLxDH_multi<"vqrdmlsdhx", 0b1, 0b1, 0b1>;
4554
4555class MVE_VCMUL<string iname, string suffix, bit size, string cstr="">
4556  : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
4557                   (ins MQPR:$Qn, MQPR:$Qm, complexrotateop:$rot),
4558                   "$Qd, $Qn, $Qm, $rot", vpred_r, cstr, []> {
4559  bits<4> Qn;
4560  bits<2> rot;
4561
4562  let Inst{28} = size;
4563  let Inst{21-20} = 0b11;
4564  let Inst{19-17} = Qn{2-0};
4565  let Inst{16} = 0b0;
4566  let Inst{12} = rot{1};
4567  let Inst{8} = 0b0;
4568  let Inst{7} = Qn{3};
4569  let Inst{0} = rot{0};
4570
4571  let Predicates = [HasMVEFloat];
4572}
4573
4574multiclass MVE_VCMUL_m<string iname, MVEVectorVTInfo VTI,
4575                       bit size, string cstr=""> {
4576  def "" : MVE_VCMUL<iname, VTI.Suffix, size, cstr>;
4577  defvar Inst = !cast<Instruction>(NAME);
4578
4579  let Predicates = [HasMVEFloat] in {
4580    def : Pat<(VTI.Vec (int_arm_mve_vcmulq
4581                            imm:$rot, (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
4582              (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
4583                             imm:$rot))>;
4584
4585    def : Pat<(VTI.Vec (int_arm_mve_vcmulq_predicated
4586                            imm:$rot, (VTI.Vec MQPR:$inactive),
4587                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
4588                            (VTI.Pred VCCR:$mask))),
4589              (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
4590                             imm:$rot, ARMVCCThen, (VTI.Pred VCCR:$mask),
4591                             (VTI.Vec MQPR:$inactive)))>;
4592
4593  }
4594}
4595
4596defm MVE_VCMULf16 : MVE_VCMUL_m<"vcmul", MVE_v8f16, 0b0>;
4597defm MVE_VCMULf32 : MVE_VCMUL_m<"vcmul", MVE_v4f32, 0b1, "@earlyclobber $Qd">;
4598
4599class MVE_VMULL<string iname, string suffix, bit bit_28, bits<2> bits_21_20,
4600                bit T, string cstr, list<dag> pattern=[]>
4601  : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
4602                   (ins MQPR:$Qn, MQPR:$Qm), "$Qd, $Qn, $Qm",
4603                   vpred_r, cstr, pattern> {
4604  bits<4> Qd;
4605  bits<4> Qn;
4606  bits<4> Qm;
4607
4608  let Inst{28} = bit_28;
4609  let Inst{21-20} = bits_21_20;
4610  let Inst{19-17} = Qn{2-0};
4611  let Inst{16} = 0b1;
4612  let Inst{12} = T;
4613  let Inst{8} = 0b0;
4614  let Inst{7} = Qn{3};
4615  let Inst{0} = 0b0;
4616  let validForTailPredication = 1;
4617  let doubleWidthResult = 1;
4618}
4619
4620multiclass MVE_VMULL_m<MVEVectorVTInfo VTI,
4621                       SDPatternOperator unpred_op, Intrinsic pred_int,
4622                       bit Top, string cstr=""> {
4623  def "" : MVE_VMULL<"vmull" # !if(Top, "t", "b"), VTI.Suffix, VTI.Unsigned,
4624                     VTI.Size, Top, cstr>;
4625  defvar Inst = !cast<Instruction>(NAME);
4626
4627  let Predicates = [HasMVEInt] in {
4628    defvar uflag = !if(!eq(VTI.SuffixLetter, "p"), (?), (? (i32 VTI.Unsigned)));
4629
4630    // Unpredicated multiply
4631    def : Pat<(VTI.DblVec !con((unpred_op (VTI.Vec MQPR:$Qm),
4632                                          (VTI.Vec MQPR:$Qn)),
4633                               uflag, (? (i32 Top)))),
4634              (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
4635
4636    // Predicated multiply
4637    def : Pat<(VTI.DblVec !con((pred_int (VTI.Vec MQPR:$Qm),
4638                                         (VTI.Vec MQPR:$Qn)),
4639                               uflag, (? (i32 Top), (VTI.DblPred VCCR:$mask),
4640                                         (VTI.DblVec MQPR:$inactive)))),
4641              (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
4642                                ARMVCCThen, (VTI.DblPred VCCR:$mask),
4643                                (VTI.DblVec MQPR:$inactive)))>;
4644  }
4645}
4646
4647// For polynomial multiplies, the size bits take the unused value 0b11, and
4648// the unsigned bit switches to encoding the size.
4649
4650defm MVE_VMULLBs8  : MVE_VMULL_m<MVE_v16s8, int_arm_mve_vmull,
4651                                 int_arm_mve_mull_int_predicated, 0b0>;
4652defm MVE_VMULLTs8  : MVE_VMULL_m<MVE_v16s8, int_arm_mve_vmull,
4653                                 int_arm_mve_mull_int_predicated, 0b1>;
4654defm MVE_VMULLBs16 : MVE_VMULL_m<MVE_v8s16, int_arm_mve_vmull,
4655                                 int_arm_mve_mull_int_predicated, 0b0>;
4656defm MVE_VMULLTs16 : MVE_VMULL_m<MVE_v8s16, int_arm_mve_vmull,
4657                                 int_arm_mve_mull_int_predicated, 0b1>;
4658defm MVE_VMULLBs32 : MVE_VMULL_m<MVE_v4s32, int_arm_mve_vmull,
4659                                 int_arm_mve_mull_int_predicated, 0b0,
4660                                 "@earlyclobber $Qd">;
4661defm MVE_VMULLTs32 : MVE_VMULL_m<MVE_v4s32, int_arm_mve_vmull,
4662                                 int_arm_mve_mull_int_predicated, 0b1,
4663                                 "@earlyclobber $Qd">;
4664
4665defm MVE_VMULLBu8  : MVE_VMULL_m<MVE_v16u8, int_arm_mve_vmull,
4666                                 int_arm_mve_mull_int_predicated, 0b0>;
4667defm MVE_VMULLTu8  : MVE_VMULL_m<MVE_v16u8, int_arm_mve_vmull,
4668                                 int_arm_mve_mull_int_predicated, 0b1>;
4669defm MVE_VMULLBu16 : MVE_VMULL_m<MVE_v8u16, int_arm_mve_vmull,
4670                                 int_arm_mve_mull_int_predicated, 0b0>;
4671defm MVE_VMULLTu16 : MVE_VMULL_m<MVE_v8u16, int_arm_mve_vmull,
4672                                 int_arm_mve_mull_int_predicated, 0b1>;
4673defm MVE_VMULLBu32 : MVE_VMULL_m<MVE_v4u32, int_arm_mve_vmull,
4674                                 int_arm_mve_mull_int_predicated, 0b0,
4675                                 "@earlyclobber $Qd">;
4676defm MVE_VMULLTu32 : MVE_VMULL_m<MVE_v4u32, int_arm_mve_vmull,
4677                                 int_arm_mve_mull_int_predicated, 0b1,
4678                                 "@earlyclobber $Qd">;
4679
4680defm MVE_VMULLBp8  : MVE_VMULL_m<MVE_v16p8, int_arm_mve_vmull_poly,
4681                                 int_arm_mve_mull_poly_predicated, 0b0>;
4682defm MVE_VMULLTp8  : MVE_VMULL_m<MVE_v16p8, int_arm_mve_vmull_poly,
4683                                 int_arm_mve_mull_poly_predicated, 0b1>;
4684defm MVE_VMULLBp16 : MVE_VMULL_m<MVE_v8p16, int_arm_mve_vmull_poly,
4685                                 int_arm_mve_mull_poly_predicated, 0b0>;
4686defm MVE_VMULLTp16 : MVE_VMULL_m<MVE_v8p16, int_arm_mve_vmull_poly,
4687                                 int_arm_mve_mull_poly_predicated, 0b1>;
4688
4689let Predicates = [HasMVEInt] in {
4690  def : Pat<(v2i64 (ARMvmulls (v4i32 MQPR:$src1), (v4i32 MQPR:$src2))),
4691            (MVE_VMULLBs32 MQPR:$src1, MQPR:$src2)>;
4692  def : Pat<(v2i64 (ARMvmulls (v4i32 (ARMvrev64 (v4i32 MQPR:$src1))),
4693                              (v4i32 (ARMvrev64 (v4i32 MQPR:$src2))))),
4694            (MVE_VMULLTs32 MQPR:$src1, MQPR:$src2)>;
4695
4696  def : Pat<(mul (sext_inreg (v4i32 MQPR:$src1), v4i16),
4697                 (sext_inreg (v4i32 MQPR:$src2), v4i16)),
4698            (MVE_VMULLBs16 MQPR:$src1, MQPR:$src2)>;
4699  def : Pat<(mul (sext_inreg (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src1)))), v4i16),
4700                 (sext_inreg (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src2)))), v4i16)),
4701            (MVE_VMULLTs16 MQPR:$src1, MQPR:$src2)>;
4702
4703  def : Pat<(mul (sext_inreg (v8i16 MQPR:$src1), v8i8),
4704                 (sext_inreg (v8i16 MQPR:$src2), v8i8)),
4705            (MVE_VMULLBs8 MQPR:$src1, MQPR:$src2)>;
4706  def : Pat<(mul (sext_inreg (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src1)))), v8i8),
4707                 (sext_inreg (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src2)))), v8i8)),
4708            (MVE_VMULLTs8 MQPR:$src1, MQPR:$src2)>;
4709
4710  def : Pat<(v2i64 (ARMvmullu (v4i32 MQPR:$src1), (v4i32 MQPR:$src2))),
4711            (MVE_VMULLBu32 MQPR:$src1, MQPR:$src2)>;
4712  def : Pat<(v2i64 (ARMvmullu (v4i32 (ARMvrev64 (v4i32 MQPR:$src1))),
4713                              (v4i32 (ARMvrev64 (v4i32 MQPR:$src2))))),
4714            (MVE_VMULLTu32 MQPR:$src1, MQPR:$src2)>;
4715
4716  def : Pat<(mul (and (v4i32 MQPR:$src1), (v4i32 (ARMvmovImm (i32 0xCFF)))),
4717                 (and (v4i32 MQPR:$src2), (v4i32 (ARMvmovImm (i32 0xCFF))))),
4718            (MVE_VMULLBu16 MQPR:$src1, MQPR:$src2)>;
4719  def : Pat<(mul (and (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src1)))),
4720                      (v4i32 (ARMvmovImm (i32 0xCFF)))),
4721                 (and (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src2)))),
4722                      (v4i32 (ARMvmovImm (i32 0xCFF))))),
4723            (MVE_VMULLTu16 MQPR:$src1, MQPR:$src2)>;
4724
4725  def : Pat<(mul (ARMvbicImm (v8i16 MQPR:$src1), (i32 0xAFF)),
4726                 (ARMvbicImm (v8i16 MQPR:$src2), (i32 0xAFF))),
4727            (MVE_VMULLBu8 MQPR:$src1, MQPR:$src2)>;
4728  def : Pat<(mul (ARMvbicImm (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src1)))), (i32 0xAFF)),
4729                 (ARMvbicImm (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src2)))), (i32 0xAFF))),
4730            (MVE_VMULLTu8 MQPR:$src1, MQPR:$src2)>;
4731}
4732
4733class MVE_VxMULH<string iname, string suffix, bit U, bits<2> size, bit round,
4734                 list<dag> pattern=[]>
4735  : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
4736                   (ins MQPR:$Qn, MQPR:$Qm), "$Qd, $Qn, $Qm",
4737                   vpred_r, "", pattern> {
4738  bits<4> Qn;
4739
4740  let Inst{28} = U;
4741  let Inst{21-20} = size;
4742  let Inst{19-17} = Qn{2-0};
4743  let Inst{16} = 0b1;
4744  let Inst{12} = round;
4745  let Inst{8} = 0b0;
4746  let Inst{7} = Qn{3};
4747  let Inst{0} = 0b1;
4748}
4749
4750multiclass MVE_VxMULH_m<string iname, MVEVectorVTInfo VTI, SDNode unpred_op,
4751                        Intrinsic pred_int, bit round> {
4752  def "" : MVE_VxMULH<iname, VTI.Suffix, VTI.Unsigned, VTI.Size, round>;
4753  defvar Inst = !cast<Instruction>(NAME);
4754
4755  let Predicates = [HasMVEInt] in {
4756    // Unpredicated multiply returning high bits
4757    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
4758                            (i32 VTI.Unsigned))),
4759              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
4760
4761    // Predicated multiply returning high bits
4762    def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
4763                            (i32 VTI.Unsigned), (VTI.Pred VCCR:$mask),
4764                            (VTI.Vec MQPR:$inactive))),
4765              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
4766                             ARMVCCThen, (VTI.Pred VCCR:$mask),
4767                             (VTI.Vec MQPR:$inactive)))>;
4768  }
4769}
4770
4771multiclass MVE_VMULT<string iname, MVEVectorVTInfo VTI, bit round>
4772  : MVE_VxMULH_m<iname, VTI, !if(round, int_arm_mve_vrmulh, int_arm_mve_vmulh),
4773                 !if(round, int_arm_mve_rmulh_predicated,
4774                            int_arm_mve_mulh_predicated),
4775                 round>;
4776
4777defm MVE_VMULHs8   : MVE_VMULT<"vmulh",  MVE_v16s8, 0b0>;
4778defm MVE_VMULHs16  : MVE_VMULT<"vmulh",  MVE_v8s16, 0b0>;
4779defm MVE_VMULHs32  : MVE_VMULT<"vmulh",  MVE_v4s32, 0b0>;
4780defm MVE_VMULHu8   : MVE_VMULT<"vmulh",  MVE_v16u8, 0b0>;
4781defm MVE_VMULHu16  : MVE_VMULT<"vmulh",  MVE_v8u16, 0b0>;
4782defm MVE_VMULHu32  : MVE_VMULT<"vmulh",  MVE_v4u32, 0b0>;
4783
4784defm MVE_VRMULHs8  : MVE_VMULT<"vrmulh", MVE_v16s8, 0b1>;
4785defm MVE_VRMULHs16 : MVE_VMULT<"vrmulh", MVE_v8s16, 0b1>;
4786defm MVE_VRMULHs32 : MVE_VMULT<"vrmulh", MVE_v4s32, 0b1>;
4787defm MVE_VRMULHu8  : MVE_VMULT<"vrmulh", MVE_v16u8, 0b1>;
4788defm MVE_VRMULHu16 : MVE_VMULT<"vrmulh", MVE_v8u16, 0b1>;
4789defm MVE_VRMULHu32 : MVE_VMULT<"vrmulh", MVE_v4u32, 0b1>;
4790
4791class MVE_VxMOVxN<string iname, string suffix, bit bit_28, bit bit_17,
4792                  bits<2> size, bit T, list<dag> pattern=[]>
4793  : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
4794                   (ins MQPR:$Qd_src, MQPR:$Qm), "$Qd, $Qm",
4795                   vpred_n, "$Qd = $Qd_src", pattern> {
4796
4797  let Inst{28} = bit_28;
4798  let Inst{21-20} = 0b11;
4799  let Inst{19-18} = size;
4800  let Inst{17} = bit_17;
4801  let Inst{16} = 0b1;
4802  let Inst{12} = T;
4803  let Inst{8} = 0b0;
4804  let Inst{7} = !not(bit_17);
4805  let Inst{0} = 0b1;
4806  let validForTailPredication = 1;
4807  let retainsPreviousHalfElement = 1;
4808}
4809
4810multiclass MVE_VxMOVxN_halves<string iname, string suffix,
4811                              bit bit_28, bit bit_17, bits<2> size> {
4812  def bh : MVE_VxMOVxN<iname # "b", suffix, bit_28, bit_17, size, 0b0>;
4813  def th : MVE_VxMOVxN<iname # "t", suffix, bit_28, bit_17, size, 0b1>;
4814}
4815
4816defm MVE_VMOVNi16   : MVE_VxMOVxN_halves<"vmovn",   "i16", 0b1, 0b0, 0b00>;
4817defm MVE_VMOVNi32   : MVE_VxMOVxN_halves<"vmovn",   "i32", 0b1, 0b0, 0b01>;
4818defm MVE_VQMOVNs16  : MVE_VxMOVxN_halves<"vqmovn",  "s16", 0b0, 0b1, 0b00>;
4819defm MVE_VQMOVNs32  : MVE_VxMOVxN_halves<"vqmovn",  "s32", 0b0, 0b1, 0b01>;
4820defm MVE_VQMOVNu16  : MVE_VxMOVxN_halves<"vqmovn",  "u16", 0b1, 0b1, 0b00>;
4821defm MVE_VQMOVNu32  : MVE_VxMOVxN_halves<"vqmovn",  "u32", 0b1, 0b1, 0b01>;
4822defm MVE_VQMOVUNs16 : MVE_VxMOVxN_halves<"vqmovun", "s16", 0b0, 0b0, 0b00>;
4823defm MVE_VQMOVUNs32 : MVE_VxMOVxN_halves<"vqmovun", "s32", 0b0, 0b0, 0b01>;
4824
4825def MVEvmovn       : SDNode<"ARMISD::VMOVN", SDTARMVEXT>;
4826
4827multiclass MVE_VMOVN_p<Instruction Inst, bit top,
4828                       MVEVectorVTInfo VTI, MVEVectorVTInfo InVTI> {
4829  // Match the most obvious MVEvmovn(a,b,t), which overwrites the odd or even
4830  // lanes of a (depending on t) with the even lanes of b.
4831  def : Pat<(VTI.Vec (MVEvmovn (VTI.Vec MQPR:$Qd_src),
4832                               (VTI.Vec MQPR:$Qm), (i32 top))),
4833            (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src), (VTI.Vec MQPR:$Qm)))>;
4834
4835  if !not(top) then {
4836    // If we see MVEvmovn(a,ARMvrev(b),1), that wants to overwrite the odd
4837    // lanes of a with the odd lanes of b. In other words, the lanes we're
4838    // _keeping_ from a are the even ones. So we can flip it round and say that
4839    // this is the same as overwriting the even lanes of b with the even lanes
4840    // of a, i.e. it's a VMOVNB with the operands reversed.
4841    defvar vrev = !cast<SDNode>("ARMvrev" # InVTI.LaneBits);
4842    def : Pat<(VTI.Vec (MVEvmovn (VTI.Vec MQPR:$Qm),
4843                                 (VTI.Vec (vrev MQPR:$Qd_src)), (i32 1))),
4844              (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src), (VTI.Vec MQPR:$Qm)))>;
4845  }
4846
4847  // Match the IR intrinsic for a predicated VMOVN. This regards the Qm input
4848  // as having wider lanes that we're narrowing, instead of already-narrow
4849  // lanes that we're taking every other one of.
4850  def : Pat<(VTI.Vec (int_arm_mve_vmovn_predicated (VTI.Vec MQPR:$Qd_src),
4851                                  (InVTI.Vec MQPR:$Qm), (i32 top),
4852                                  (InVTI.Pred VCCR:$pred))),
4853            (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src),
4854                              (InVTI.Vec MQPR:$Qm),
4855                              ARMVCCThen, (InVTI.Pred VCCR:$pred)))>;
4856}
4857
4858defm : MVE_VMOVN_p<MVE_VMOVNi32bh, 0, MVE_v8i16, MVE_v4i32>;
4859defm : MVE_VMOVN_p<MVE_VMOVNi32th, 1, MVE_v8i16, MVE_v4i32>;
4860defm : MVE_VMOVN_p<MVE_VMOVNi16bh, 0, MVE_v16i8, MVE_v8i16>;
4861defm : MVE_VMOVN_p<MVE_VMOVNi16th, 1, MVE_v16i8, MVE_v8i16>;
4862
4863multiclass MVE_VQMOVN_p<Instruction Inst, bit outU, bit inU, bit top,
4864                        MVEVectorVTInfo VTI, MVEVectorVTInfo InVTI> {
4865  def : Pat<(VTI.Vec (int_arm_mve_vqmovn (VTI.Vec MQPR:$Qd_src),
4866                                  (InVTI.Vec MQPR:$Qm),
4867                                  (i32 outU), (i32 inU), (i32 top))),
4868            (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src),
4869                              (InVTI.Vec MQPR:$Qm)))>;
4870
4871  def : Pat<(VTI.Vec (int_arm_mve_vqmovn_predicated (VTI.Vec MQPR:$Qd_src),
4872                                  (InVTI.Vec MQPR:$Qm),
4873                                  (i32 outU), (i32 inU), (i32 top),
4874                                  (InVTI.Pred VCCR:$pred))),
4875            (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src),
4876                              (InVTI.Vec MQPR:$Qm),
4877                              ARMVCCThen, (InVTI.Pred VCCR:$pred)))>;
4878}
4879
4880defm : MVE_VQMOVN_p<MVE_VQMOVNs32bh,  0, 0, 0, MVE_v8i16, MVE_v4i32>;
4881defm : MVE_VQMOVN_p<MVE_VQMOVNs32th,  0, 0, 1, MVE_v8i16, MVE_v4i32>;
4882defm : MVE_VQMOVN_p<MVE_VQMOVNs16bh,  0, 0, 0, MVE_v16i8, MVE_v8i16>;
4883defm : MVE_VQMOVN_p<MVE_VQMOVNs16th,  0, 0, 1, MVE_v16i8, MVE_v8i16>;
4884defm : MVE_VQMOVN_p<MVE_VQMOVNu32bh,  1, 1, 0, MVE_v8i16, MVE_v4i32>;
4885defm : MVE_VQMOVN_p<MVE_VQMOVNu32th,  1, 1, 1, MVE_v8i16, MVE_v4i32>;
4886defm : MVE_VQMOVN_p<MVE_VQMOVNu16bh,  1, 1, 0, MVE_v16i8, MVE_v8i16>;
4887defm : MVE_VQMOVN_p<MVE_VQMOVNu16th,  1, 1, 1, MVE_v16i8, MVE_v8i16>;
4888defm : MVE_VQMOVN_p<MVE_VQMOVUNs32bh, 1, 0, 0, MVE_v8i16, MVE_v4i32>;
4889defm : MVE_VQMOVN_p<MVE_VQMOVUNs32th, 1, 0, 1, MVE_v8i16, MVE_v4i32>;
4890defm : MVE_VQMOVN_p<MVE_VQMOVUNs16bh, 1, 0, 0, MVE_v16i8, MVE_v8i16>;
4891defm : MVE_VQMOVN_p<MVE_VQMOVUNs16th, 1, 0, 1, MVE_v16i8, MVE_v8i16>;
4892
4893def SDTARMVMOVNQ : SDTypeProfile<1, 3, [SDTCisVec<0>, SDTCisSameAs<0, 1>,
4894                                        SDTCisVec<2>, SDTCisVT<3, i32>]>;
4895def MVEvqmovns   : SDNode<"ARMISD::VQMOVNs", SDTARMVMOVNQ>;
4896def MVEvqmovnu   : SDNode<"ARMISD::VQMOVNu", SDTARMVMOVNQ>;
4897
4898let Predicates = [HasMVEInt] in {
4899  def : Pat<(v8i16 (MVEvqmovns (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), (i32 0))),
4900            (v8i16 (MVE_VQMOVNs32bh (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm)))>;
4901  def : Pat<(v8i16 (MVEvqmovns (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), (i32 1))),
4902            (v8i16 (MVE_VQMOVNs32th (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm)))>;
4903  def : Pat<(v16i8 (MVEvqmovns (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), (i32 0))),
4904            (v16i8 (MVE_VQMOVNs16bh (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm)))>;
4905  def : Pat<(v16i8 (MVEvqmovns (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), (i32 1))),
4906            (v16i8 (MVE_VQMOVNs16th (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm)))>;
4907
4908  def : Pat<(v8i16 (MVEvqmovnu (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), (i32 0))),
4909            (v8i16 (MVE_VQMOVNu32bh (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm)))>;
4910  def : Pat<(v8i16 (MVEvqmovnu (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), (i32 1))),
4911            (v8i16 (MVE_VQMOVNu32th (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm)))>;
4912  def : Pat<(v16i8 (MVEvqmovnu (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), (i32 0))),
4913            (v16i8 (MVE_VQMOVNu16bh (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm)))>;
4914  def : Pat<(v16i8 (MVEvqmovnu (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), (i32 1))),
4915            (v16i8 (MVE_VQMOVNu16th (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm)))>;
4916
4917  def : Pat<(v8i16 (MVEvqmovns (v8i16 MQPR:$Qd_src), (v4i32 (ARMvshrsImm (v4i32 MQPR:$Qm), imm0_31:$imm)), (i32 0))),
4918            (v8i16 (MVE_VQSHRNbhs32 (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), imm0_31:$imm))>;
4919  def : Pat<(v16i8 (MVEvqmovns (v16i8 MQPR:$Qd_src), (v8i16 (ARMvshrsImm (v8i16 MQPR:$Qm), imm0_15:$imm)), (i32 0))),
4920            (v16i8 (MVE_VQSHRNbhs16 (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), imm0_15:$imm))>;
4921  def : Pat<(v8i16 (MVEvqmovns (v8i16 MQPR:$Qd_src), (v4i32 (ARMvshrsImm (v4i32 MQPR:$Qm), imm0_31:$imm)), (i32 1))),
4922            (v8i16 (MVE_VQSHRNths32 (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), imm0_31:$imm))>;
4923  def : Pat<(v16i8 (MVEvqmovns (v16i8 MQPR:$Qd_src), (v8i16 (ARMvshrsImm (v8i16 MQPR:$Qm), imm0_15:$imm)), (i32 1))),
4924            (v16i8 (MVE_VQSHRNths16 (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), imm0_15:$imm))>;
4925
4926  def : Pat<(v8i16 (MVEvqmovnu (v8i16 MQPR:$Qd_src), (v4i32 (ARMvshruImm (v4i32 MQPR:$Qm), imm0_31:$imm)), (i32 0))),
4927            (v8i16 (MVE_VQSHRNbhu32 (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), imm0_31:$imm))>;
4928  def : Pat<(v16i8 (MVEvqmovnu (v16i8 MQPR:$Qd_src), (v8i16 (ARMvshruImm (v8i16 MQPR:$Qm), imm0_15:$imm)), (i32 0))),
4929            (v16i8 (MVE_VQSHRNbhu16 (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), imm0_15:$imm))>;
4930  def : Pat<(v8i16 (MVEvqmovnu (v8i16 MQPR:$Qd_src), (v4i32 (ARMvshruImm (v4i32 MQPR:$Qm), imm0_31:$imm)), (i32 1))),
4931            (v8i16 (MVE_VQSHRNthu32 (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), imm0_31:$imm))>;
4932  def : Pat<(v16i8 (MVEvqmovnu (v16i8 MQPR:$Qd_src), (v8i16 (ARMvshruImm (v8i16 MQPR:$Qm), imm0_15:$imm)), (i32 1))),
4933            (v16i8 (MVE_VQSHRNthu16 (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), imm0_15:$imm))>;
4934}
4935
4936class MVE_VCVT_ff<string iname, string suffix, bit op, bit T,
4937                  dag iops_extra, vpred_ops vpred, string cstr>
4938  : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
4939                   !con(iops_extra, (ins MQPR:$Qm)), "$Qd, $Qm",
4940                   vpred, cstr, []> {
4941  let Inst{28} = op;
4942  let Inst{21-16} = 0b111111;
4943  let Inst{12} = T;
4944  let Inst{8-7} = 0b00;
4945  let Inst{0} = 0b1;
4946
4947  let Predicates = [HasMVEFloat];
4948  let retainsPreviousHalfElement = 1;
4949}
4950
4951def SDTARMVCVTL    : SDTypeProfile<1, 2, [SDTCisVec<0>, SDTCisVec<1>,
4952                                         SDTCisVT<2, i32>]>;
4953def MVEvcvtn       : SDNode<"ARMISD::VCVTN", SDTARMVMOVNQ>;
4954def MVEvcvtl       : SDNode<"ARMISD::VCVTL", SDTARMVCVTL>;
4955
4956multiclass MVE_VCVT_f2h_m<string iname, int half> {
4957  def "": MVE_VCVT_ff<iname, "f16.f32", 0b0, half,
4958                      (ins MQPR:$Qd_src), vpred_n, "$Qd = $Qd_src">;
4959  defvar Inst = !cast<Instruction>(NAME);
4960
4961  let Predicates = [HasMVEFloat] in {
4962    def : Pat<(v8f16 (int_arm_mve_vcvt_narrow
4963                         (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm), (i32 half))),
4964              (v8f16 (Inst (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm)))>;
4965    def : Pat<(v8f16 (int_arm_mve_vcvt_narrow_predicated
4966                         (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm), (i32 half),
4967                         (v4i1 VCCR:$mask))),
4968              (v8f16 (Inst (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm),
4969                           ARMVCCThen, (v4i1 VCCR:$mask)))>;
4970
4971    def : Pat<(v8f16 (MVEvcvtn (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm), (i32 half))),
4972              (v8f16 (Inst (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm)))>;
4973  }
4974}
4975
4976multiclass MVE_VCVT_h2f_m<string iname, int half> {
4977  def "": MVE_VCVT_ff<iname, "f32.f16", 0b1, half, (ins), vpred_r, "">;
4978  defvar Inst = !cast<Instruction>(NAME);
4979
4980  let Predicates = [HasMVEFloat] in {
4981    def : Pat<(v4f32 (int_arm_mve_vcvt_widen (v8f16 MQPR:$Qm), (i32 half))),
4982              (v4f32 (Inst (v8f16 MQPR:$Qm)))>;
4983    def : Pat<(v4f32 (int_arm_mve_vcvt_widen_predicated
4984                         (v4f32 MQPR:$inactive), (v8f16 MQPR:$Qm), (i32 half),
4985                         (v4i1 VCCR:$mask))),
4986              (v4f32 (Inst (v8f16 MQPR:$Qm), ARMVCCThen,
4987                           (v4i1 VCCR:$mask), (v4f32 MQPR:$inactive)))>;
4988
4989    def : Pat<(v4f32 (MVEvcvtl (v8f16 MQPR:$Qm), (i32 half))),
4990              (v4f32 (Inst (v8f16 MQPR:$Qm)))>;
4991  }
4992}
4993
4994defm MVE_VCVTf16f32bh : MVE_VCVT_f2h_m<"vcvtb", 0b0>;
4995defm MVE_VCVTf16f32th : MVE_VCVT_f2h_m<"vcvtt", 0b1>;
4996defm MVE_VCVTf32f16bh : MVE_VCVT_h2f_m<"vcvtb", 0b0>;
4997defm MVE_VCVTf32f16th : MVE_VCVT_h2f_m<"vcvtt", 0b1>;
4998
4999class MVE_VxCADD<string iname, string suffix, bits<2> size, bit halve,
5000                 string cstr="">
5001  : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
5002                   (ins MQPR:$Qn, MQPR:$Qm, complexrotateopodd:$rot),
5003                   "$Qd, $Qn, $Qm, $rot", vpred_r, cstr, []> {
5004  bits<4> Qn;
5005  bit rot;
5006
5007  let Inst{28} = halve;
5008  let Inst{21-20} = size;
5009  let Inst{19-17} = Qn{2-0};
5010  let Inst{16} = 0b0;
5011  let Inst{12} = rot;
5012  let Inst{8} = 0b1;
5013  let Inst{7} = Qn{3};
5014  let Inst{0} = 0b0;
5015}
5016
5017multiclass MVE_VxCADD_m<string iname, MVEVectorVTInfo VTI,
5018                        bit halve, string cstr=""> {
5019  def "" : MVE_VxCADD<iname, VTI.Suffix, VTI.Size, halve, cstr>;
5020  defvar Inst = !cast<Instruction>(NAME);
5021
5022  let Predicates = [HasMVEInt] in {
5023    def : Pat<(VTI.Vec (int_arm_mve_vcaddq halve,
5024                            imm:$rot, (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
5025              (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
5026                             imm:$rot))>;
5027
5028    def : Pat<(VTI.Vec (int_arm_mve_vcaddq_predicated halve,
5029                            imm:$rot, (VTI.Vec MQPR:$inactive),
5030                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
5031                            (VTI.Pred VCCR:$mask))),
5032              (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
5033                             imm:$rot, ARMVCCThen, (VTI.Pred VCCR:$mask),
5034                             (VTI.Vec MQPR:$inactive)))>;
5035
5036  }
5037}
5038
5039defm MVE_VCADDi8   : MVE_VxCADD_m<"vcadd", MVE_v16i8, 0b1>;
5040defm MVE_VCADDi16  : MVE_VxCADD_m<"vcadd", MVE_v8i16, 0b1>;
5041defm MVE_VCADDi32  : MVE_VxCADD_m<"vcadd", MVE_v4i32, 0b1, "@earlyclobber $Qd">;
5042
5043defm MVE_VHCADDs8  : MVE_VxCADD_m<"vhcadd", MVE_v16s8, 0b0>;
5044defm MVE_VHCADDs16 : MVE_VxCADD_m<"vhcadd", MVE_v8s16, 0b0>;
5045defm MVE_VHCADDs32 : MVE_VxCADD_m<"vhcadd", MVE_v4s32, 0b0, "@earlyclobber $Qd">;
5046
5047class MVE_VADCSBC<string iname, bit I, bit subtract,
5048                  dag carryin, list<dag> pattern=[]>
5049  : MVE_qDest_qSrc<iname, "i32", (outs MQPR:$Qd, cl_FPSCR_NZCV:$carryout),
5050                   !con((ins MQPR:$Qn, MQPR:$Qm), carryin),
5051                   "$Qd, $Qn, $Qm", vpred_r, "", pattern> {
5052  bits<4> Qn;
5053
5054  let Inst{28} = subtract;
5055  let Inst{21-20} = 0b11;
5056  let Inst{19-17} = Qn{2-0};
5057  let Inst{16} = 0b0;
5058  let Inst{12} = I;
5059  let Inst{8} = 0b1;
5060  let Inst{7} = Qn{3};
5061  let Inst{0} = 0b0;
5062
5063  // Custom decoder method in order to add the FPSCR operand(s), which
5064  // Tablegen won't do right
5065  let DecoderMethod = "DecodeMVEVADCInstruction";
5066}
5067
5068def MVE_VADC  : MVE_VADCSBC<"vadc",  0b0, 0b0, (ins cl_FPSCR_NZCV:$carryin)>;
5069def MVE_VADCI : MVE_VADCSBC<"vadci", 0b1, 0b0, (ins)>;
5070
5071def MVE_VSBC  : MVE_VADCSBC<"vsbc",  0b0, 0b1, (ins cl_FPSCR_NZCV:$carryin)>;
5072def MVE_VSBCI : MVE_VADCSBC<"vsbci", 0b1, 0b1, (ins)>;
5073
5074class MVE_VQDMULL<string iname, string suffix, bit size, bit T,
5075                  string cstr="", list<dag> pattern=[]>
5076  : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
5077                   (ins MQPR:$Qn, MQPR:$Qm), "$Qd, $Qn, $Qm",
5078                   vpred_r, cstr, pattern> {
5079  bits<4> Qn;
5080
5081  let Inst{28} = size;
5082  let Inst{21-20} = 0b11;
5083  let Inst{19-17} = Qn{2-0};
5084  let Inst{16} = 0b0;
5085  let Inst{12} = T;
5086  let Inst{8} = 0b1;
5087  let Inst{7} = Qn{3};
5088  let Inst{0} = 0b1;
5089  let validForTailPredication = 1;
5090  let doubleWidthResult = 1;
5091}
5092
5093multiclass MVE_VQDMULL_m<string iname, MVEVectorVTInfo VTI, bit size, bit T,
5094                         string cstr> {
5095  def "" : MVE_VQDMULL<iname, VTI.Suffix, size, T, cstr>;
5096  defvar Inst = !cast<Instruction>(NAME);
5097
5098  let Predicates = [HasMVEInt] in {
5099    // Unpredicated saturating multiply
5100    def : Pat<(VTI.DblVec (int_arm_mve_vqdmull (VTI.Vec MQPR:$Qm),
5101                                               (VTI.Vec MQPR:$Qn), (i32 T))),
5102              (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
5103    // Predicated saturating multiply
5104    def : Pat<(VTI.DblVec (int_arm_mve_vqdmull_predicated
5105                                    (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
5106                                    (i32 T), (VTI.DblPred VCCR:$mask),
5107                                    (VTI.DblVec MQPR:$inactive))),
5108              (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
5109                                ARMVCCThen, (VTI.DblPred VCCR:$mask),
5110                                (VTI.DblVec MQPR:$inactive)))>;
5111  }
5112}
5113
5114multiclass MVE_VQDMULL_halves<MVEVectorVTInfo VTI, bit size, string cstr=""> {
5115  defm bh : MVE_VQDMULL_m<"vqdmullb", VTI, size, 0b0, cstr>;
5116  defm th : MVE_VQDMULL_m<"vqdmullt", VTI, size, 0b1, cstr>;
5117}
5118
5119defm MVE_VQDMULLs16 : MVE_VQDMULL_halves<MVE_v8s16, 0b0>;
5120defm MVE_VQDMULLs32 : MVE_VQDMULL_halves<MVE_v4s32, 0b1, "@earlyclobber $Qd">;
5121
5122// end of mve_qDest_qSrc
5123
5124// start of mve_qDest_rSrc
5125
5126class MVE_qr_base<dag oops, dag iops, InstrItinClass itin, string iname,
5127                  string suffix, string ops, vpred_ops vpred, string cstr,
5128                  list<dag> pattern=[]>
5129   : MVE_p<oops, iops, NoItinerary, iname, suffix, ops, vpred, cstr, pattern> {
5130  bits<4> Qd;
5131  bits<4> Qn;
5132  bits<4> Rm;
5133
5134  let Inst{25-23} = 0b100;
5135  let Inst{22} = Qd{3};
5136  let Inst{19-17} = Qn{2-0};
5137  let Inst{15-13} = Qd{2-0};
5138  let Inst{11-9} = 0b111;
5139  let Inst{7} = Qn{3};
5140  let Inst{6} = 0b1;
5141  let Inst{4} = 0b0;
5142  let Inst{3-0} = Rm{3-0};
5143}
5144
5145class MVE_qDest_rSrc<string iname, string suffix, string cstr="", list<dag> pattern=[]>
5146  : MVE_qr_base<(outs MQPR:$Qd), (ins MQPR:$Qn, rGPR:$Rm),
5147          NoItinerary, iname, suffix, "$Qd, $Qn, $Rm", vpred_r, cstr,
5148           pattern>;
5149
5150class MVE_qDestSrc_rSrc<string iname, string suffix, list<dag> pattern=[]>
5151  : MVE_qr_base<(outs MQPR:$Qd), (ins MQPR:$Qd_src, MQPR:$Qn, rGPR:$Rm),
5152          NoItinerary, iname, suffix, "$Qd, $Qn, $Rm", vpred_n, "$Qd = $Qd_src",
5153           pattern>;
5154
5155class MVE_qDest_single_rSrc<string iname, string suffix, list<dag> pattern=[]>
5156  : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qd_src, rGPR:$Rm), NoItinerary, iname,
5157          suffix, "$Qd, $Rm", vpred_n, "$Qd = $Qd_src", pattern> {
5158  bits<4> Qd;
5159  bits<4> Rm;
5160
5161  let Inst{22} = Qd{3};
5162  let Inst{15-13} = Qd{2-0};
5163  let Inst{3-0} = Rm{3-0};
5164}
5165
5166// Patterns for vector-scalar instructions with integer operands
5167multiclass MVE_vec_scalar_int_pat_m<Instruction inst, MVEVectorVTInfo VTI,
5168                                    SDPatternOperator unpred_op,
5169                                    SDPatternOperator pred_op,
5170                                    bit unpred_has_sign = 0,
5171                                    bit pred_has_sign = 0> {
5172  defvar UnpredSign = !if(unpred_has_sign, (? (i32 VTI.Unsigned)), (?));
5173  defvar PredSign = !if(pred_has_sign, (? (i32 VTI.Unsigned)), (?));
5174
5175  let Predicates = [HasMVEInt] in {
5176    // Unpredicated version
5177    def : Pat<(VTI.Vec !con((unpred_op (VTI.Vec MQPR:$Qm),
5178                                       (VTI.Vec (ARMvdup rGPR:$val))),
5179                            UnpredSign)),
5180              (VTI.Vec (inst (VTI.Vec MQPR:$Qm), (i32 rGPR:$val)))>;
5181    // Predicated version
5182    def : Pat<(VTI.Vec !con((pred_op (VTI.Vec MQPR:$Qm),
5183                                     (VTI.Vec (ARMvdup rGPR:$val))),
5184                            PredSign,
5185                            (pred_op (VTI.Pred VCCR:$mask),
5186                                     (VTI.Vec MQPR:$inactive)))),
5187              (VTI.Vec (inst (VTI.Vec MQPR:$Qm), (i32 rGPR:$val),
5188                             ARMVCCThen, (VTI.Pred VCCR:$mask),
5189                             (VTI.Vec MQPR:$inactive)))>;
5190  }
5191}
5192
5193class MVE_VADDSUB_qr<string iname, string suffix, bits<2> size,
5194                     bit bit_5, bit bit_12, bit bit_16, bit bit_28>
5195  : MVE_qDest_rSrc<iname, suffix, ""> {
5196
5197  let Inst{28} = bit_28;
5198  let Inst{21-20} = size;
5199  let Inst{16} = bit_16;
5200  let Inst{12} = bit_12;
5201  let Inst{8} = 0b1;
5202  let Inst{5} = bit_5;
5203  let validForTailPredication = 1;
5204}
5205
5206// Vector-scalar add/sub
5207multiclass MVE_VADDSUB_qr_m<string iname, MVEVectorVTInfo VTI, bit subtract,
5208                            SDNode Op, Intrinsic PredInt> {
5209  def "" : MVE_VADDSUB_qr<iname, VTI.Suffix, VTI.Size, 0b0, subtract, 0b1, 0b0>;
5210  let Predicates = [HasMVEInt] in {
5211    defm : MVE_TwoOpPatternDup<VTI, Op, PredInt, (? ), !cast<Instruction>(NAME), ARMimmAllZerosV>;
5212  }
5213}
5214
5215multiclass MVE_VADD_qr_m<MVEVectorVTInfo VTI>
5216  : MVE_VADDSUB_qr_m<"vadd", VTI, 0b0, add, int_arm_mve_add_predicated>;
5217
5218multiclass MVE_VSUB_qr_m<MVEVectorVTInfo VTI>
5219  : MVE_VADDSUB_qr_m<"vsub", VTI, 0b1, sub, int_arm_mve_sub_predicated>;
5220
5221defm MVE_VADD_qr_i8  : MVE_VADD_qr_m<MVE_v16i8>;
5222defm MVE_VADD_qr_i16 : MVE_VADD_qr_m<MVE_v8i16>;
5223defm MVE_VADD_qr_i32 : MVE_VADD_qr_m<MVE_v4i32>;
5224
5225defm MVE_VSUB_qr_i8  : MVE_VSUB_qr_m<MVE_v16i8>;
5226defm MVE_VSUB_qr_i16 : MVE_VSUB_qr_m<MVE_v8i16>;
5227defm MVE_VSUB_qr_i32 : MVE_VSUB_qr_m<MVE_v4i32>;
5228
5229// Vector-scalar saturating add/sub
5230multiclass MVE_VQADDSUB_qr_m<string iname, MVEVectorVTInfo VTI, bit subtract,
5231                             SDNode Op, Intrinsic PredInt> {
5232  def "" : MVE_VADDSUB_qr<iname, VTI.Suffix, VTI.Size, 0b1, subtract,
5233                          0b0, VTI.Unsigned>;
5234
5235  let Predicates = [HasMVEInt] in {
5236    defm : MVE_TwoOpPatternDup<VTI, Op, PredInt, (? (i32 VTI.Unsigned)),
5237                               !cast<Instruction>(NAME)>;
5238  }
5239}
5240
5241multiclass MVE_VQADD_qr_m<MVEVectorVTInfo VTI, SDNode Op>
5242  : MVE_VQADDSUB_qr_m<"vqadd", VTI, 0b0, Op, int_arm_mve_qadd_predicated>;
5243
5244multiclass MVE_VQSUB_qr_m<MVEVectorVTInfo VTI, SDNode Op>
5245  : MVE_VQADDSUB_qr_m<"vqsub", VTI, 0b1, Op, int_arm_mve_qsub_predicated>;
5246
5247defm MVE_VQADD_qr_s8  : MVE_VQADD_qr_m<MVE_v16s8, saddsat>;
5248defm MVE_VQADD_qr_s16 : MVE_VQADD_qr_m<MVE_v8s16, saddsat>;
5249defm MVE_VQADD_qr_s32 : MVE_VQADD_qr_m<MVE_v4s32, saddsat>;
5250defm MVE_VQADD_qr_u8  : MVE_VQADD_qr_m<MVE_v16u8, uaddsat>;
5251defm MVE_VQADD_qr_u16 : MVE_VQADD_qr_m<MVE_v8u16, uaddsat>;
5252defm MVE_VQADD_qr_u32 : MVE_VQADD_qr_m<MVE_v4u32, uaddsat>;
5253
5254defm MVE_VQSUB_qr_s8  : MVE_VQSUB_qr_m<MVE_v16s8, ssubsat>;
5255defm MVE_VQSUB_qr_s16 : MVE_VQSUB_qr_m<MVE_v8s16, ssubsat>;
5256defm MVE_VQSUB_qr_s32 : MVE_VQSUB_qr_m<MVE_v4s32, ssubsat>;
5257defm MVE_VQSUB_qr_u8  : MVE_VQSUB_qr_m<MVE_v16u8, usubsat>;
5258defm MVE_VQSUB_qr_u16 : MVE_VQSUB_qr_m<MVE_v8u16, usubsat>;
5259defm MVE_VQSUB_qr_u32 : MVE_VQSUB_qr_m<MVE_v4u32, usubsat>;
5260
5261class MVE_VQDMULL_qr<string iname, string suffix, bit size,
5262                     bit T, string cstr="", list<dag> pattern=[]>
5263  : MVE_qDest_rSrc<iname, suffix, cstr, pattern> {
5264
5265  let Inst{28} = size;
5266  let Inst{21-20} = 0b11;
5267  let Inst{16} = 0b0;
5268  let Inst{12} = T;
5269  let Inst{8} = 0b1;
5270  let Inst{5} = 0b1;
5271  let validForTailPredication = 1;
5272  let doubleWidthResult = 1;
5273}
5274
5275multiclass MVE_VQDMULL_qr_m<string iname, MVEVectorVTInfo VTI, bit size,
5276                            bit T, string cstr> {
5277  def "" : MVE_VQDMULL_qr<iname, VTI.Suffix, size, T, cstr>;
5278  defvar Inst = !cast<Instruction>(NAME);
5279
5280  let Predicates = [HasMVEInt] in {
5281    // Unpredicated saturating multiply
5282    def : Pat<(VTI.DblVec (int_arm_mve_vqdmull (VTI.Vec MQPR:$Qm),
5283                                               (VTI.Vec (ARMvdup rGPR:$val)),
5284                                               (i32 T))),
5285              (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (i32 rGPR:$val)))>;
5286    // Predicated saturating multiply
5287    def : Pat<(VTI.DblVec (int_arm_mve_vqdmull_predicated
5288                                    (VTI.Vec MQPR:$Qm),
5289                                    (VTI.Vec (ARMvdup rGPR:$val)),
5290                                    (i32 T),
5291                                    (VTI.DblPred VCCR:$mask),
5292                                    (VTI.DblVec MQPR:$inactive))),
5293              (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (i32 rGPR:$val),
5294                             ARMVCCThen, (VTI.DblPred VCCR:$mask),
5295                             (VTI.DblVec MQPR:$inactive)))>;
5296  }
5297}
5298
5299multiclass MVE_VQDMULL_qr_halves<MVEVectorVTInfo VTI, bit size, string cstr=""> {
5300  defm bh : MVE_VQDMULL_qr_m<"vqdmullb", VTI, size, 0b0, cstr>;
5301  defm th : MVE_VQDMULL_qr_m<"vqdmullt", VTI, size, 0b1, cstr>;
5302}
5303
5304defm MVE_VQDMULL_qr_s16 : MVE_VQDMULL_qr_halves<MVE_v8s16, 0b0>;
5305defm MVE_VQDMULL_qr_s32 : MVE_VQDMULL_qr_halves<MVE_v4s32, 0b1, "@earlyclobber $Qd">;
5306
5307class MVE_VxADDSUB_qr<string iname, string suffix,
5308                      bit bit_28, bits<2> bits_21_20, bit subtract,
5309                      list<dag> pattern=[]>
5310  : MVE_qDest_rSrc<iname, suffix, "", pattern> {
5311
5312  let Inst{28} = bit_28;
5313  let Inst{21-20} = bits_21_20;
5314  let Inst{16} = 0b0;
5315  let Inst{12} = subtract;
5316  let Inst{8} = 0b1;
5317  let Inst{5} = 0b0;
5318  let validForTailPredication = 1;
5319}
5320
5321multiclass MVE_VHADDSUB_qr_m<string iname, MVEVectorVTInfo VTI, bit subtract,
5322                             Intrinsic unpred_int, Intrinsic pred_int> {
5323  def "" : MVE_VxADDSUB_qr<iname, VTI.Suffix, VTI.Unsigned, VTI.Size, subtract>;
5324  defm : MVE_vec_scalar_int_pat_m<!cast<Instruction>(NAME),
5325                                  VTI, unpred_int, pred_int, 1, 1>;
5326}
5327
5328multiclass MVE_VHADD_qr_m<MVEVectorVTInfo VTI> :
5329  MVE_VHADDSUB_qr_m<"vhadd", VTI, 0b0, int_arm_mve_vhadd,
5330                                       int_arm_mve_hadd_predicated>;
5331
5332multiclass MVE_VHSUB_qr_m<MVEVectorVTInfo VTI> :
5333  MVE_VHADDSUB_qr_m<"vhsub", VTI, 0b1, int_arm_mve_vhsub,
5334                                       int_arm_mve_hsub_predicated>;
5335
5336defm MVE_VHADD_qr_s8  : MVE_VHADD_qr_m<MVE_v16s8>;
5337defm MVE_VHADD_qr_s16 : MVE_VHADD_qr_m<MVE_v8s16>;
5338defm MVE_VHADD_qr_s32 : MVE_VHADD_qr_m<MVE_v4s32>;
5339defm MVE_VHADD_qr_u8  : MVE_VHADD_qr_m<MVE_v16u8>;
5340defm MVE_VHADD_qr_u16 : MVE_VHADD_qr_m<MVE_v8u16>;
5341defm MVE_VHADD_qr_u32 : MVE_VHADD_qr_m<MVE_v4u32>;
5342
5343defm MVE_VHSUB_qr_s8  : MVE_VHSUB_qr_m<MVE_v16s8>;
5344defm MVE_VHSUB_qr_s16 : MVE_VHSUB_qr_m<MVE_v8s16>;
5345defm MVE_VHSUB_qr_s32 : MVE_VHSUB_qr_m<MVE_v4s32>;
5346defm MVE_VHSUB_qr_u8  : MVE_VHSUB_qr_m<MVE_v16u8>;
5347defm MVE_VHSUB_qr_u16 : MVE_VHSUB_qr_m<MVE_v8u16>;
5348defm MVE_VHSUB_qr_u32 : MVE_VHSUB_qr_m<MVE_v4u32>;
5349
5350multiclass MVE_VADDSUB_qr_f<string iname, MVEVectorVTInfo VTI, bit subtract,
5351                            SDNode Op, Intrinsic PredInt> {
5352  def "" : MVE_VxADDSUB_qr<iname, VTI.Suffix, VTI.Size{0}, 0b11, subtract>;
5353  defm : MVE_TwoOpPatternDup<VTI, Op, PredInt, (? ),
5354                              !cast<Instruction>(NAME)>;
5355}
5356
5357let Predicates = [HasMVEFloat] in {
5358  defm MVE_VADD_qr_f32 : MVE_VADDSUB_qr_f<"vadd", MVE_v4f32, 0b0, fadd,
5359                                          int_arm_mve_add_predicated>;
5360  defm MVE_VADD_qr_f16 : MVE_VADDSUB_qr_f<"vadd", MVE_v8f16, 0b0, fadd,
5361                                          int_arm_mve_add_predicated>;
5362
5363  defm MVE_VSUB_qr_f32 : MVE_VADDSUB_qr_f<"vsub", MVE_v4f32, 0b1, fsub,
5364                                          int_arm_mve_sub_predicated>;
5365  defm MVE_VSUB_qr_f16 : MVE_VADDSUB_qr_f<"vsub", MVE_v8f16, 0b1, fsub,
5366                                          int_arm_mve_sub_predicated>;
5367}
5368
5369class MVE_VxSHL_qr<string iname, string suffix, bit U, bits<2> size,
5370                   bit bit_7, bit bit_17, list<dag> pattern=[]>
5371  : MVE_qDest_single_rSrc<iname, suffix, pattern> {
5372
5373  let Inst{28} = U;
5374  let Inst{25-23} = 0b100;
5375  let Inst{21-20} = 0b11;
5376  let Inst{19-18} = size;
5377  let Inst{17} = bit_17;
5378  let Inst{16} = 0b1;
5379  let Inst{12-8} = 0b11110;
5380  let Inst{7} = bit_7;
5381  let Inst{6-4} = 0b110;
5382  let validForTailPredication = 1;
5383}
5384
5385multiclass MVE_VxSHL_qr_p<string iname, MVEVectorVTInfo VTI, bit q, bit r> {
5386  def "" : MVE_VxSHL_qr<iname, VTI.Suffix, VTI.Unsigned, VTI.Size, q, r>;
5387  defvar Inst = !cast<Instruction>(NAME);
5388
5389  def : Pat<(VTI.Vec (int_arm_mve_vshl_scalar
5390                         (VTI.Vec MQPR:$in), (i32 rGPR:$sh),
5391                         (i32 q), (i32 r), (i32 VTI.Unsigned))),
5392            (VTI.Vec (Inst (VTI.Vec MQPR:$in), (i32 rGPR:$sh)))>;
5393
5394  def : Pat<(VTI.Vec (int_arm_mve_vshl_scalar_predicated
5395                         (VTI.Vec MQPR:$in), (i32 rGPR:$sh),
5396                         (i32 q), (i32 r), (i32 VTI.Unsigned),
5397                         (VTI.Pred VCCR:$mask))),
5398            (VTI.Vec (Inst (VTI.Vec MQPR:$in), (i32 rGPR:$sh),
5399                           ARMVCCThen, (VTI.Pred VCCR:$mask)))>;
5400}
5401
5402multiclass MVE_VxSHL_qr_types<string iname, bit bit_7, bit bit_17> {
5403  defm s8  : MVE_VxSHL_qr_p<iname, MVE_v16s8, bit_7, bit_17>;
5404  defm s16 : MVE_VxSHL_qr_p<iname, MVE_v8s16, bit_7, bit_17>;
5405  defm s32 : MVE_VxSHL_qr_p<iname, MVE_v4s32, bit_7, bit_17>;
5406  defm u8  : MVE_VxSHL_qr_p<iname, MVE_v16u8, bit_7, bit_17>;
5407  defm u16 : MVE_VxSHL_qr_p<iname, MVE_v8u16, bit_7, bit_17>;
5408  defm u32 : MVE_VxSHL_qr_p<iname, MVE_v4u32, bit_7, bit_17>;
5409}
5410
5411defm MVE_VSHL_qr   : MVE_VxSHL_qr_types<"vshl",   0b0, 0b0>;
5412defm MVE_VRSHL_qr  : MVE_VxSHL_qr_types<"vrshl",  0b0, 0b1>;
5413defm MVE_VQSHL_qr  : MVE_VxSHL_qr_types<"vqshl",  0b1, 0b0>;
5414defm MVE_VQRSHL_qr : MVE_VxSHL_qr_types<"vqrshl", 0b1, 0b1>;
5415
5416let Predicates = [HasMVEInt] in {
5417  def : Pat<(v4i32 (ARMvshlu (v4i32 MQPR:$Qm), (v4i32 (ARMvdup rGPR:$Rm)))),
5418            (v4i32 (MVE_VSHL_qru32 (v4i32 MQPR:$Qm), rGPR:$Rm))>;
5419  def : Pat<(v8i16 (ARMvshlu (v8i16 MQPR:$Qm), (v8i16 (ARMvdup rGPR:$Rm)))),
5420            (v8i16 (MVE_VSHL_qru16 (v8i16 MQPR:$Qm), rGPR:$Rm))>;
5421  def : Pat<(v16i8 (ARMvshlu (v16i8 MQPR:$Qm), (v16i8 (ARMvdup rGPR:$Rm)))),
5422            (v16i8 (MVE_VSHL_qru8 (v16i8 MQPR:$Qm), rGPR:$Rm))>;
5423
5424  def : Pat<(v4i32 (ARMvshls (v4i32 MQPR:$Qm), (v4i32 (ARMvdup rGPR:$Rm)))),
5425            (v4i32 (MVE_VSHL_qrs32 (v4i32 MQPR:$Qm), rGPR:$Rm))>;
5426  def : Pat<(v8i16 (ARMvshls (v8i16 MQPR:$Qm), (v8i16 (ARMvdup rGPR:$Rm)))),
5427            (v8i16 (MVE_VSHL_qrs16 (v8i16 MQPR:$Qm), rGPR:$Rm))>;
5428  def : Pat<(v16i8 (ARMvshls (v16i8 MQPR:$Qm), (v16i8 (ARMvdup rGPR:$Rm)))),
5429            (v16i8 (MVE_VSHL_qrs8 (v16i8 MQPR:$Qm), rGPR:$Rm))>;
5430}
5431
5432class MVE_VBRSR<string iname, string suffix, bits<2> size, list<dag> pattern=[]>
5433  : MVE_qDest_rSrc<iname, suffix, "", pattern> {
5434
5435  let Inst{28} = 0b1;
5436  let Inst{21-20} = size;
5437  let Inst{16} = 0b1;
5438  let Inst{12} = 0b1;
5439  let Inst{8} = 0b0;
5440  let Inst{5} = 0b1;
5441  let validForTailPredication = 1;
5442}
5443
5444def MVE_VBRSR8  : MVE_VBRSR<"vbrsr", "8", 0b00>;
5445def MVE_VBRSR16 : MVE_VBRSR<"vbrsr", "16", 0b01>;
5446def MVE_VBRSR32 : MVE_VBRSR<"vbrsr", "32", 0b10>;
5447
5448multiclass MVE_VBRSR_pat_m<MVEVectorVTInfo VTI, Instruction Inst> {
5449  // Unpredicated
5450  def : Pat<(VTI.Vec (int_arm_mve_vbrsr (VTI.Vec MQPR:$Qn), (i32 rGPR:$Rm))),
5451            (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (i32 rGPR:$Rm)))>;
5452  // Predicated
5453  def : Pat<(VTI.Vec (int_arm_mve_vbrsr_predicated
5454                          (VTI.Vec MQPR:$inactive),
5455                          (VTI.Vec MQPR:$Qn), (i32 rGPR:$Rm),
5456                          (VTI.Pred VCCR:$mask))),
5457            (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (i32 rGPR:$Rm),
5458                          ARMVCCThen, (VTI.Pred VCCR:$mask),
5459                          (VTI.Vec MQPR:$inactive)))>;
5460}
5461
5462let Predicates = [HasMVEInt] in {
5463  def : Pat<(v16i8 ( bitreverse (v16i8 MQPR:$val1))),
5464            (v16i8 ( MVE_VBRSR8 (v16i8 MQPR:$val1), (t2MOVi (i32 8)) ))>;
5465
5466  def : Pat<(v4i32 ( bitreverse (v4i32 MQPR:$val1))),
5467            (v4i32 ( MVE_VBRSR32 (v4i32 MQPR:$val1), (t2MOVi (i32 32)) ))>;
5468
5469  def : Pat<(v8i16 ( bitreverse (v8i16 MQPR:$val1))),
5470            (v8i16 ( MVE_VBRSR16 (v8i16 MQPR:$val1), (t2MOVi (i32 16)) ))>;
5471
5472  defm : MVE_VBRSR_pat_m<MVE_v16i8, MVE_VBRSR8>;
5473  defm : MVE_VBRSR_pat_m<MVE_v8i16, MVE_VBRSR16>;
5474  defm : MVE_VBRSR_pat_m<MVE_v4i32, MVE_VBRSR32>;
5475}
5476
5477let Predicates = [HasMVEFloat] in {
5478  defm : MVE_VBRSR_pat_m<MVE_v8f16, MVE_VBRSR16>;
5479  defm : MVE_VBRSR_pat_m<MVE_v4f32, MVE_VBRSR32>;
5480}
5481
5482class MVE_VMUL_qr_int<string iname, string suffix, bits<2> size>
5483  : MVE_qDest_rSrc<iname, suffix, ""> {
5484
5485  let Inst{28} = 0b0;
5486  let Inst{21-20} = size;
5487  let Inst{16} = 0b1;
5488  let Inst{12} = 0b1;
5489  let Inst{8} = 0b0;
5490  let Inst{5} = 0b1;
5491  let validForTailPredication = 1;
5492}
5493
5494multiclass MVE_VMUL_qr_int_m<MVEVectorVTInfo VTI> {
5495  def "" : MVE_VMUL_qr_int<"vmul", VTI.Suffix, VTI.Size>;
5496  let Predicates = [HasMVEInt] in {
5497    defm : MVE_TwoOpPatternDup<VTI, mul, int_arm_mve_mul_predicated, (? ),
5498                               !cast<Instruction>(NAME), ARMimmOneV>;
5499  }
5500}
5501
5502defm MVE_VMUL_qr_i8  : MVE_VMUL_qr_int_m<MVE_v16i8>;
5503defm MVE_VMUL_qr_i16 : MVE_VMUL_qr_int_m<MVE_v8i16>;
5504defm MVE_VMUL_qr_i32 : MVE_VMUL_qr_int_m<MVE_v4i32>;
5505
5506class MVE_VxxMUL_qr<string iname, string suffix,
5507                    bit bit_28, bits<2> bits_21_20, list<dag> pattern=[]>
5508  : MVE_qDest_rSrc<iname, suffix, "", pattern> {
5509
5510  let Inst{28} = bit_28;
5511  let Inst{21-20} = bits_21_20;
5512  let Inst{16} = 0b1;
5513  let Inst{12} = 0b0;
5514  let Inst{8} = 0b0;
5515  let Inst{5} = 0b1;
5516  let validForTailPredication = 1;
5517}
5518
5519multiclass MVE_VxxMUL_qr_m<string iname, MVEVectorVTInfo VTI, bit bit_28,
5520                           PatFrag Op, Intrinsic int_unpred, Intrinsic int_pred> {
5521  def "" : MVE_VxxMUL_qr<iname, VTI.Suffix, bit_28, VTI.Size>;
5522
5523  let Predicates = [HasMVEInt] in {
5524    defm : MVE_TwoOpPatternDup<VTI, Op, int_pred, (? ), !cast<Instruction>(NAME)>;
5525  }
5526  defm : MVE_vec_scalar_int_pat_m<!cast<Instruction>(NAME), VTI, int_unpred, int_pred>;
5527}
5528
5529multiclass MVE_VQDMULH_qr_m<MVEVectorVTInfo VTI> :
5530  MVE_VxxMUL_qr_m<"vqdmulh", VTI, 0b0, MVEvqdmulh,
5531                  int_arm_mve_vqdmulh, int_arm_mve_qdmulh_predicated>;
5532
5533multiclass MVE_VQRDMULH_qr_m<MVEVectorVTInfo VTI> :
5534  MVE_VxxMUL_qr_m<"vqrdmulh", VTI, 0b1, null_frag,
5535                  int_arm_mve_vqrdmulh, int_arm_mve_qrdmulh_predicated>;
5536
5537defm MVE_VQDMULH_qr_s8    : MVE_VQDMULH_qr_m<MVE_v16s8>;
5538defm MVE_VQDMULH_qr_s16   : MVE_VQDMULH_qr_m<MVE_v8s16>;
5539defm MVE_VQDMULH_qr_s32   : MVE_VQDMULH_qr_m<MVE_v4s32>;
5540
5541defm MVE_VQRDMULH_qr_s8   : MVE_VQRDMULH_qr_m<MVE_v16s8>;
5542defm MVE_VQRDMULH_qr_s16  : MVE_VQRDMULH_qr_m<MVE_v8s16>;
5543defm MVE_VQRDMULH_qr_s32  : MVE_VQRDMULH_qr_m<MVE_v4s32>;
5544
5545multiclass MVE_VxxMUL_qr_f_m<MVEVectorVTInfo VTI> {
5546  let validForTailPredication = 1 in
5547  def "" : MVE_VxxMUL_qr<"vmul", VTI.Suffix, VTI.Size{0}, 0b11>;
5548  defm : MVE_TwoOpPatternDup<VTI, fmul, int_arm_mve_mul_predicated, (? ),
5549                             !cast<Instruction>(NAME)>;
5550}
5551
5552let Predicates = [HasMVEFloat] in {
5553  defm MVE_VMUL_qr_f16   : MVE_VxxMUL_qr_f_m<MVE_v8f16>;
5554  defm MVE_VMUL_qr_f32   : MVE_VxxMUL_qr_f_m<MVE_v4f32>;
5555}
5556
5557class MVE_VFMAMLA_qr<string iname, string suffix,
5558                     bit bit_28, bits<2> bits_21_20, bit S,
5559                     list<dag> pattern=[]>
5560  : MVE_qDestSrc_rSrc<iname, suffix, pattern> {
5561
5562  let Inst{28} = bit_28;
5563  let Inst{21-20} = bits_21_20;
5564  let Inst{16} = 0b1;
5565  let Inst{12} = S;
5566  let Inst{8} = 0b0;
5567  let Inst{5} = 0b0;
5568  let validForTailPredication = 1;
5569  let hasSideEffects = 0;
5570}
5571
5572multiclass MVE_VMLA_qr_multi<string iname, MVEVectorVTInfo VTI,
5573                             bit scalar_addend> {
5574  def "": MVE_VFMAMLA_qr<iname, VTI.Suffix, VTI.Unsigned, VTI.Size,
5575                         scalar_addend>;
5576  defvar Inst = !cast<Instruction>(NAME);
5577  defvar pred_int = !cast<Intrinsic>("int_arm_mve_" # iname # "_n_predicated");
5578  defvar v1   = (VTI.Vec MQPR:$v1);
5579  defvar v2   = (VTI.Vec MQPR:$v2);
5580  defvar vs   = (VTI.Vec (ARMvdup rGPR:$s));
5581  defvar s    = (i32 rGPR:$s);
5582  defvar pred = (VTI.Pred VCCR:$pred);
5583
5584  // The signed and unsigned variants of this instruction have different
5585  // encodings, but they're functionally identical. For the sake of
5586  // determinism, we generate only the unsigned variant.
5587  if VTI.Unsigned then let Predicates = [HasMVEInt] in {
5588    if scalar_addend then {
5589      def : Pat<(VTI.Vec (add (mul v1, v2), vs)),
5590                (VTI.Vec (Inst v1, v2, s))>;
5591    } else {
5592      def : Pat<(VTI.Vec (add (mul v2, vs), v1)),
5593                (VTI.Vec (Inst v1, v2, s))>;
5594    }
5595
5596    def : Pat<(VTI.Vec (pred_int v1, v2, s, pred)),
5597              (VTI.Vec (Inst v1, v2, s, ARMVCCThen, pred))>;
5598  }
5599}
5600
5601defm MVE_VMLA_qr_s8   : MVE_VMLA_qr_multi<"vmla", MVE_v16s8, 0b0>;
5602defm MVE_VMLA_qr_s16  : MVE_VMLA_qr_multi<"vmla", MVE_v8s16, 0b0>;
5603defm MVE_VMLA_qr_s32  : MVE_VMLA_qr_multi<"vmla", MVE_v4s32, 0b0>;
5604defm MVE_VMLA_qr_u8   : MVE_VMLA_qr_multi<"vmla", MVE_v16u8, 0b0>;
5605defm MVE_VMLA_qr_u16  : MVE_VMLA_qr_multi<"vmla", MVE_v8u16, 0b0>;
5606defm MVE_VMLA_qr_u32  : MVE_VMLA_qr_multi<"vmla", MVE_v4u32, 0b0>;
5607
5608defm MVE_VMLAS_qr_s8  : MVE_VMLA_qr_multi<"vmlas", MVE_v16s8, 0b1>;
5609defm MVE_VMLAS_qr_s16 : MVE_VMLA_qr_multi<"vmlas", MVE_v8s16, 0b1>;
5610defm MVE_VMLAS_qr_s32 : MVE_VMLA_qr_multi<"vmlas", MVE_v4s32, 0b1>;
5611defm MVE_VMLAS_qr_u8  : MVE_VMLA_qr_multi<"vmlas", MVE_v16u8, 0b1>;
5612defm MVE_VMLAS_qr_u16 : MVE_VMLA_qr_multi<"vmlas", MVE_v8u16, 0b1>;
5613defm MVE_VMLAS_qr_u32 : MVE_VMLA_qr_multi<"vmlas", MVE_v4u32, 0b1>;
5614
5615multiclass MVE_VFMA_qr_multi<string iname, MVEVectorVTInfo VTI,
5616                             bit scalar_addend> {
5617  def "": MVE_VFMAMLA_qr<iname, VTI.Suffix, VTI.Size{0}, 0b11, scalar_addend>;
5618  defvar Inst = !cast<Instruction>(NAME);
5619  defvar pred_int = int_arm_mve_fma_predicated;
5620  defvar v1   = (VTI.Vec MQPR:$v1);
5621  defvar v2   = (VTI.Vec MQPR:$v2);
5622  defvar vs   = (VTI.Vec (ARMvdup (i32 rGPR:$s)));
5623  defvar is   = (i32 rGPR:$s);
5624  defvar pred = (VTI.Pred VCCR:$pred);
5625
5626  let Predicates = [HasMVEFloat] in {
5627    if scalar_addend then {
5628      def : Pat<(VTI.Vec (fma v1, v2, vs)),
5629                (VTI.Vec (Inst v1, v2, is))>;
5630      def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
5631                                  (VTI.Vec (fma v1, v2, vs)),
5632                                  v1)),
5633                (VTI.Vec (Inst v1, v2, is, ARMVCCThen, $pred))>;
5634      def : Pat<(VTI.Vec (pred_int v1, v2, vs, pred)),
5635                (VTI.Vec (Inst v1, v2, is, ARMVCCThen, pred))>;
5636    } else {
5637      def : Pat<(VTI.Vec (fma v1, vs, v2)),
5638                (VTI.Vec (Inst v2, v1, is))>;
5639      def : Pat<(VTI.Vec (fma vs, v1, v2)),
5640                (VTI.Vec (Inst v2, v1, is))>;
5641      def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
5642                                  (VTI.Vec (fma vs, v2, v1)),
5643                                  v1)),
5644                (VTI.Vec (Inst v1, v2, is, ARMVCCThen, $pred))>;
5645      def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
5646                                  (VTI.Vec (fma v2, vs, v1)),
5647                                  v1)),
5648                (VTI.Vec (Inst v1, v2, is, ARMVCCThen, $pred))>;
5649      def : Pat<(VTI.Vec (pred_int v1, vs, v2, pred)),
5650                (VTI.Vec (Inst v2, v1, is, ARMVCCThen, pred))>;
5651      def : Pat<(VTI.Vec (pred_int vs, v1, v2, pred)),
5652                (VTI.Vec (Inst v2, v1, is, ARMVCCThen, pred))>;
5653    }
5654  }
5655}
5656
5657let Predicates = [HasMVEFloat] in {
5658  defm MVE_VFMA_qr_f16  : MVE_VFMA_qr_multi<"vfma",  MVE_v8f16, 0>;
5659  defm MVE_VFMA_qr_f32  : MVE_VFMA_qr_multi<"vfma",  MVE_v4f32, 0>;
5660  defm MVE_VFMA_qr_Sf16 : MVE_VFMA_qr_multi<"vfmas", MVE_v8f16, 1>;
5661  defm MVE_VFMA_qr_Sf32 : MVE_VFMA_qr_multi<"vfmas", MVE_v4f32, 1>;
5662}
5663
5664class MVE_VQDMLAH_qr<string iname, string suffix, bit U, bits<2> size,
5665                     bit bit_5, bit bit_12, list<dag> pattern=[]>
5666  : MVE_qDestSrc_rSrc<iname, suffix, pattern> {
5667
5668  let Inst{28} = U;
5669  let Inst{21-20} = size;
5670  let Inst{16} = 0b0;
5671  let Inst{12} = bit_12;
5672  let Inst{8} = 0b0;
5673  let Inst{5} = bit_5;
5674}
5675
5676multiclass MVE_VQDMLAH_qr_multi<string iname, MVEVectorVTInfo VTI,
5677                                bit bit_5, bit bit_12> {
5678  def "": MVE_VQDMLAH_qr<iname, VTI.Suffix, 0b0, VTI.Size, bit_5, bit_12>;
5679  defvar Inst = !cast<Instruction>(NAME);
5680  defvar unpred_int = !cast<Intrinsic>("int_arm_mve_" # iname);
5681  defvar pred_int = !cast<Intrinsic>("int_arm_mve_" # iname # "_predicated");
5682
5683  let Predicates = [HasMVEInt] in {
5684    def : Pat<(VTI.Vec (unpred_int (VTI.Vec MQPR:$v1), (VTI.Vec MQPR:$v2),
5685                                   (i32 rGPR:$s))),
5686              (VTI.Vec (Inst       (VTI.Vec MQPR:$v1), (VTI.Vec MQPR:$v2),
5687                                   (i32 rGPR:$s)))>;
5688    def : Pat<(VTI.Vec (pred_int   (VTI.Vec MQPR:$v1), (VTI.Vec MQPR:$v2),
5689                                   (i32 rGPR:$s), (VTI.Pred VCCR:$pred))),
5690              (VTI.Vec (Inst       (VTI.Vec MQPR:$v1), (VTI.Vec MQPR:$v2),
5691                                   (i32 rGPR:$s), ARMVCCThen,
5692                                   (VTI.Pred VCCR:$pred)))>;
5693  }
5694}
5695
5696multiclass MVE_VQDMLAH_qr_types<string iname, bit bit_5, bit bit_12> {
5697  defm s8  : MVE_VQDMLAH_qr_multi<iname, MVE_v16s8, bit_5, bit_12>;
5698  defm s16 : MVE_VQDMLAH_qr_multi<iname, MVE_v8s16, bit_5, bit_12>;
5699  defm s32 : MVE_VQDMLAH_qr_multi<iname, MVE_v4s32, bit_5, bit_12>;
5700}
5701
5702defm MVE_VQDMLAH_qr   : MVE_VQDMLAH_qr_types<"vqdmlah",   0b1, 0b0>;
5703defm MVE_VQRDMLAH_qr  : MVE_VQDMLAH_qr_types<"vqrdmlah",  0b0, 0b0>;
5704defm MVE_VQDMLASH_qr  : MVE_VQDMLAH_qr_types<"vqdmlash",  0b1, 0b1>;
5705defm MVE_VQRDMLASH_qr : MVE_VQDMLAH_qr_types<"vqrdmlash", 0b0, 0b1>;
5706
5707class MVE_VxDUP<string iname, string suffix, bits<2> size, bit bit_12,
5708                ValueType VT, SDPatternOperator vxdup>
5709  : MVE_p<(outs MQPR:$Qd, tGPREven:$Rn),
5710          (ins tGPREven:$Rn_src, MVE_VIDUP_imm:$imm), NoItinerary,
5711          iname, suffix, "$Qd, $Rn, $imm", vpred_r, "$Rn = $Rn_src",
5712          [(set (VT MQPR:$Qd), (i32 tGPREven:$Rn),
5713              (vxdup (i32 tGPREven:$Rn_src), (i32 imm:$imm)))]> {
5714  bits<4> Qd;
5715  bits<4> Rn;
5716  bits<2> imm;
5717
5718  let Inst{28} = 0b0;
5719  let Inst{25-23} = 0b100;
5720  let Inst{22} = Qd{3};
5721  let Inst{21-20} = size;
5722  let Inst{19-17} = Rn{3-1};
5723  let Inst{16} = 0b1;
5724  let Inst{15-13} = Qd{2-0};
5725  let Inst{12} = bit_12;
5726  let Inst{11-8} = 0b1111;
5727  let Inst{7} = imm{1};
5728  let Inst{6-1} = 0b110111;
5729  let Inst{0} = imm{0};
5730  let validForTailPredication = 1;
5731  let hasSideEffects = 0;
5732}
5733
5734def MVE_VIDUPu8  : MVE_VxDUP<"vidup", "u8",  0b00, 0b0, v16i8, ARMvidup>;
5735def MVE_VIDUPu16 : MVE_VxDUP<"vidup", "u16", 0b01, 0b0, v8i16, ARMvidup>;
5736def MVE_VIDUPu32 : MVE_VxDUP<"vidup", "u32", 0b10, 0b0, v4i32, ARMvidup>;
5737
5738def MVE_VDDUPu8  : MVE_VxDUP<"vddup", "u8",  0b00, 0b1, v16i8, null_frag>;
5739def MVE_VDDUPu16 : MVE_VxDUP<"vddup", "u16", 0b01, 0b1, v8i16, null_frag>;
5740def MVE_VDDUPu32 : MVE_VxDUP<"vddup", "u32", 0b10, 0b1, v4i32, null_frag>;
5741
5742class MVE_VxWDUP<string iname, string suffix, bits<2> size, bit bit_12,
5743                 list<dag> pattern=[]>
5744  : MVE_p<(outs MQPR:$Qd, tGPREven:$Rn),
5745          (ins tGPREven:$Rn_src, tGPROdd:$Rm, MVE_VIDUP_imm:$imm), NoItinerary,
5746          iname, suffix, "$Qd, $Rn, $Rm, $imm", vpred_r, "$Rn = $Rn_src",
5747          pattern> {
5748  bits<4> Qd;
5749  bits<4> Rm;
5750  bits<4> Rn;
5751  bits<2> imm;
5752
5753  let Inst{28} = 0b0;
5754  let Inst{25-23} = 0b100;
5755  let Inst{22} = Qd{3};
5756  let Inst{21-20} = size;
5757  let Inst{19-17} = Rn{3-1};
5758  let Inst{16} = 0b1;
5759  let Inst{15-13} = Qd{2-0};
5760  let Inst{12} = bit_12;
5761  let Inst{11-8} = 0b1111;
5762  let Inst{7} = imm{1};
5763  let Inst{6-4} = 0b110;
5764  let Inst{3-1} = Rm{3-1};
5765  let Inst{0} = imm{0};
5766  let validForTailPredication = 1;
5767  let hasSideEffects = 0;
5768}
5769
5770def MVE_VIWDUPu8  : MVE_VxWDUP<"viwdup", "u8",  0b00, 0b0>;
5771def MVE_VIWDUPu16 : MVE_VxWDUP<"viwdup", "u16", 0b01, 0b0>;
5772def MVE_VIWDUPu32 : MVE_VxWDUP<"viwdup", "u32", 0b10, 0b0>;
5773
5774def MVE_VDWDUPu8  : MVE_VxWDUP<"vdwdup", "u8",  0b00, 0b1>;
5775def MVE_VDWDUPu16 : MVE_VxWDUP<"vdwdup", "u16", 0b01, 0b1>;
5776def MVE_VDWDUPu32 : MVE_VxWDUP<"vdwdup", "u32", 0b10, 0b1>;
5777
5778let isReMaterializable = 1 in
5779class MVE_VCTPInst<string suffix, bits<2> size, list<dag> pattern=[]>
5780  : MVE_p<(outs VCCR:$P0), (ins rGPR:$Rn), NoItinerary, "vctp", suffix,
5781          "$Rn", vpred_n, "", pattern> {
5782  bits<4> Rn;
5783
5784  let Inst{28-27} = 0b10;
5785  let Inst{26-22} = 0b00000;
5786  let Inst{21-20} = size;
5787  let Inst{19-16} = Rn{3-0};
5788  let Inst{15-11} = 0b11101;
5789  let Inst{10-0}  = 0b00000000001;
5790  let Unpredictable{10-0} = 0b11111111111;
5791
5792  let Constraints = "";
5793  let DecoderMethod = "DecodeMveVCTP";
5794  let validForTailPredication = 1;
5795}
5796
5797multiclass MVE_VCTP<MVEVectorVTInfo VTI, Intrinsic intr> {
5798  def "": MVE_VCTPInst<VTI.BitsSuffix, VTI.Size>;
5799  defvar Inst = !cast<Instruction>(NAME);
5800
5801  let Predicates = [HasMVEInt] in {
5802    def : Pat<(intr rGPR:$Rn),
5803              (VTI.Pred (Inst rGPR:$Rn))>;
5804    def : Pat<(and (intr rGPR:$Rn), (VTI.Pred VCCR:$mask)),
5805              (VTI.Pred (Inst rGPR:$Rn, ARMVCCThen, VCCR:$mask))>;
5806  }
5807}
5808
5809defm MVE_VCTP8  : MVE_VCTP<MVE_v16i8, int_arm_mve_vctp8>;
5810defm MVE_VCTP16 : MVE_VCTP<MVE_v8i16, int_arm_mve_vctp16>;
5811defm MVE_VCTP32 : MVE_VCTP<MVE_v4i32, int_arm_mve_vctp32>;
5812defm MVE_VCTP64 : MVE_VCTP<MVE_v2i64, int_arm_mve_vctp64>;
5813
5814// end of mve_qDest_rSrc
5815
5816// start of coproc mov
5817
5818class MVE_VMOV_64bit<dag oops, dag iops, bit to_qreg, string ops, string cstr>
5819  : MVE_VMOV_lane_base<oops, !con(iops, (ins MVEPairVectorIndex2:$idx,
5820                                             MVEPairVectorIndex0:$idx2)),
5821                       NoItinerary, "vmov", "", ops, cstr, []> {
5822  bits<5> Rt;
5823  bits<5> Rt2;
5824  bits<4> Qd;
5825  bit idx;
5826  bit idx2;
5827
5828  let Inst{31-23} = 0b111011000;
5829  let Inst{22} = Qd{3};
5830  let Inst{21} = 0b0;
5831  let Inst{20} = to_qreg;
5832  let Inst{19-16} = Rt2{3-0};
5833  let Inst{15-13} = Qd{2-0};
5834  let Inst{12-5} = 0b01111000;
5835  let Inst{4} = idx2;
5836  let Inst{3-0} = Rt{3-0};
5837
5838  let hasSideEffects = 0;
5839}
5840
5841// The assembly syntax for these instructions mentions the vector
5842// register name twice, e.g.
5843//
5844//    vmov q2[2], q2[0], r0, r1
5845//    vmov r0, r1, q2[2], q2[0]
5846//
5847// which needs a bit of juggling with MC operand handling.
5848//
5849// For the move _into_ a vector register, the MC operand list also has
5850// to mention the register name twice: once as the output, and once as
5851// an extra input to represent where the unchanged half of the output
5852// register comes from (when this instruction is used in code
5853// generation). So we arrange that the first mention of the vector reg
5854// in the instruction is considered by the AsmMatcher to be the output
5855// ($Qd), and the second one is the input ($QdSrc). Binding them
5856// together with the existing 'tie' constraint is enough to enforce at
5857// register allocation time that they have to be the same register.
5858//
5859// For the move _from_ a vector register, there's no way to get round
5860// the fact that both instances of that register name have to be
5861// inputs. They have to be the same register again, but this time, we
5862// can't use a tie constraint, because that has to be between an
5863// output and an input operand. So this time, we have to arrange that
5864// the q-reg appears just once in the MC operand list, in spite of
5865// being mentioned twice in the asm syntax - which needs a custom
5866// AsmMatchConverter.
5867
5868def MVE_VMOV_q_rr : MVE_VMOV_64bit<(outs MQPR:$Qd),
5869                                   (ins MQPR:$QdSrc, rGPR:$Rt, rGPR:$Rt2),
5870                                   0b1, "$Qd$idx, $QdSrc$idx2, $Rt, $Rt2",
5871                                   "$Qd = $QdSrc"> {
5872  let DecoderMethod = "DecodeMVEVMOVDRegtoQ";
5873}
5874
5875def MVE_VMOV_rr_q : MVE_VMOV_64bit<(outs rGPR:$Rt, rGPR:$Rt2), (ins MQPR:$Qd),
5876                                   0b0, "$Rt, $Rt2, $Qd$idx, $Qd$idx2", ""> {
5877  let DecoderMethod = "DecodeMVEVMOVQtoDReg";
5878  let AsmMatchConverter = "cvtMVEVMOVQtoDReg";
5879}
5880
5881let Predicates = [HasMVEInt] in {
5882  // Double lane moves. There are a number of patterns here. We know that the
5883  // insertelt's will be in descending order by index, and need to match the 5
5884  // patterns that might contain 2-0 or 3-1 pairs. These are:
5885  // 3 2 1 0    -> vmovqrr 31; vmovqrr 20
5886  // 3 2 1      -> vmovqrr 31; vmov 2
5887  // 3 1        -> vmovqrr 31
5888  // 2 1 0      -> vmovqrr 20; vmov 1
5889  // 2 0        -> vmovqrr 20
5890  // The other potential patterns will be handled by single lane inserts.
5891  def : Pat<(insertelt (insertelt (insertelt (insertelt (v4i32 MQPR:$src1),
5892                                                        rGPR:$srcA, (i32 0)),
5893                                             rGPR:$srcB, (i32 1)),
5894                                  rGPR:$srcC, (i32 2)),
5895                       rGPR:$srcD, (i32 3)),
5896            (MVE_VMOV_q_rr (MVE_VMOV_q_rr MQPR:$src1, rGPR:$srcA, rGPR:$srcC, (i32 2), (i32 0)),
5897                           rGPR:$srcB, rGPR:$srcD, (i32 3), (i32 1))>;
5898  def : Pat<(insertelt (insertelt (insertelt (v4i32 MQPR:$src1),
5899                                             rGPR:$srcB, (i32 1)),
5900                                  rGPR:$srcC, (i32 2)),
5901                       rGPR:$srcD, (i32 3)),
5902            (MVE_VMOV_q_rr (MVE_VMOV_to_lane_32 MQPR:$src1, rGPR:$srcC, (i32 2)),
5903                           rGPR:$srcB, rGPR:$srcD, (i32 3), (i32 1))>;
5904  def : Pat<(insertelt (insertelt (v4i32 MQPR:$src1), rGPR:$srcA, (i32 1)), rGPR:$srcB, (i32 3)),
5905            (MVE_VMOV_q_rr MQPR:$src1, rGPR:$srcA, rGPR:$srcB, (i32 3), (i32 1))>;
5906  def : Pat<(insertelt (insertelt (insertelt (v4i32 MQPR:$src1),
5907                                             rGPR:$srcB, (i32 0)),
5908                                  rGPR:$srcC, (i32 1)),
5909                       rGPR:$srcD, (i32 2)),
5910            (MVE_VMOV_q_rr (MVE_VMOV_to_lane_32 MQPR:$src1, rGPR:$srcC, (i32 1)),
5911                           rGPR:$srcB, rGPR:$srcD, (i32 2), (i32 0))>;
5912  def : Pat<(insertelt (insertelt (v4i32 MQPR:$src1), rGPR:$srcA, (i32 0)), rGPR:$srcB, (i32 2)),
5913            (MVE_VMOV_q_rr MQPR:$src1, rGPR:$srcA, rGPR:$srcB, (i32 2), (i32 0))>;
5914}
5915
5916// end of coproc mov
5917
5918// start of MVE interleaving load/store
5919
5920// Base class for the family of interleaving/deinterleaving
5921// load/stores with names like VLD20.8 and VST43.32.
5922class MVE_vldst24_base<bit writeback, bit fourregs, bits<2> stage, bits<2> size,
5923                       bit load, dag Oops, dag loadIops, dag wbIops,
5924                       string iname, string ops,
5925                       string cstr, list<dag> pattern=[]>
5926  : MVE_MI<Oops, !con(loadIops, wbIops), NoItinerary, iname, ops, cstr, pattern> {
5927  bits<4> VQd;
5928  bits<4> Rn;
5929
5930  let Inst{31-22} = 0b1111110010;
5931  let Inst{21} = writeback;
5932  let Inst{20} = load;
5933  let Inst{19-16} = Rn;
5934  let Inst{15-13} = VQd{2-0};
5935  let Inst{12-9} = 0b1111;
5936  let Inst{8-7} = size;
5937  let Inst{6-5} = stage;
5938  let Inst{4-1} = 0b0000;
5939  let Inst{0} = fourregs;
5940
5941  let mayLoad = load;
5942  let mayStore = !eq(load,0);
5943  let hasSideEffects = 0;
5944  let validForTailPredication = load;
5945}
5946
5947// A parameter class used to encapsulate all the ways the writeback
5948// variants of VLD20 and friends differ from the non-writeback ones.
5949class MVE_vldst24_writeback<bit b, dag Oo, dag Io,
5950                            string sy="", string c="", string n=""> {
5951  bit writeback = b;
5952  dag Oops = Oo;
5953  dag Iops = Io;
5954  string syntax = sy;
5955  string cstr = c;
5956  string id_suffix = n;
5957}
5958
5959// Another parameter class that encapsulates the differences between VLD2x
5960// and VLD4x.
5961class MVE_vldst24_nvecs<int n, list<int> s, bit b, RegisterOperand vl> {
5962  int nvecs = n;
5963  list<int> stages = s;
5964  bit bit0 = b;
5965  RegisterOperand VecList = vl;
5966}
5967
5968// A third parameter class that distinguishes VLDnn.8 from .16 from .32.
5969class MVE_vldst24_lanesize<int i, bits<2> b> {
5970  int lanesize = i;
5971  bits<2> sizebits = b;
5972}
5973
5974// A base class for each direction of transfer: one for load, one for
5975// store. I can't make these a fourth independent parametric tuple
5976// class, because they have to take the nvecs tuple class as a
5977// parameter, in order to find the right VecList operand type.
5978
5979class MVE_vld24_base<MVE_vldst24_nvecs n, bits<2> pat, bits<2> size,
5980                     MVE_vldst24_writeback wb, string iname,
5981                     list<dag> pattern=[]>
5982  : MVE_vldst24_base<wb.writeback, n.bit0, pat, size, 1,
5983                     !con((outs n.VecList:$VQd), wb.Oops),
5984                     (ins n.VecList:$VQdSrc), wb.Iops,
5985                     iname, "$VQd, $Rn" # wb.syntax,
5986                     wb.cstr # ",$VQdSrc = $VQd", pattern>;
5987
5988class MVE_vst24_base<MVE_vldst24_nvecs n, bits<2> pat, bits<2> size,
5989                     MVE_vldst24_writeback wb, string iname,
5990                     list<dag> pattern=[]>
5991  : MVE_vldst24_base<wb.writeback, n.bit0, pat, size, 0,
5992                     wb.Oops, (ins n.VecList:$VQd), wb.Iops,
5993                     iname, "$VQd, $Rn" # wb.syntax,
5994                     wb.cstr, pattern>;
5995
5996// Actually define all the interleaving loads and stores, by a series
5997// of nested foreaches over number of vectors (VLD2/VLD4); stage
5998// within one of those series (VLDx0/VLDx1/VLDx2/VLDx3); size of
5999// vector lane; writeback or no writeback.
6000foreach n = [MVE_vldst24_nvecs<2, [0,1],     0, VecList2Q>,
6001             MVE_vldst24_nvecs<4, [0,1,2,3], 1, VecList4Q>] in
6002foreach stage = n.stages in
6003foreach s = [MVE_vldst24_lanesize< 8, 0b00>,
6004             MVE_vldst24_lanesize<16, 0b01>,
6005             MVE_vldst24_lanesize<32, 0b10>] in
6006foreach wb = [MVE_vldst24_writeback<
6007                1, (outs rGPR:$wb), (ins t2_nosp_addr_offset_none:$Rn),
6008                "!", "$Rn.base = $wb", "_wb">,
6009              MVE_vldst24_writeback<0, (outs), (ins t2_addr_offset_none:$Rn)>] in {
6010
6011  // For each case within all of those foreaches, define the actual
6012  // instructions. The def names are made by gluing together pieces
6013  // from all the parameter classes, and will end up being things like
6014  // MVE_VLD20_8 and MVE_VST43_16_wb.
6015
6016  def "MVE_VLD" # n.nvecs # stage # "_" # s.lanesize # wb.id_suffix
6017    : MVE_vld24_base<n, stage, s.sizebits, wb,
6018                     "vld" # n.nvecs # stage # "." # s.lanesize>;
6019
6020  def "MVE_VST" # n.nvecs # stage # "_" # s.lanesize # wb.id_suffix
6021    : MVE_vst24_base<n, stage, s.sizebits, wb,
6022                     "vst" # n.nvecs # stage # "." # s.lanesize>;
6023}
6024
6025def SDTARMVST2    : SDTypeProfile<1, 5, [SDTCisPtrTy<0>, SDTCisPtrTy<1>, SDTCisVT<2, i32>, SDTCisVec<3>,
6026                                         SDTCisSameAs<3, 4>, SDTCisVT<5, i32>]>;
6027def SDTARMVST4    : SDTypeProfile<1, 7, [SDTCisPtrTy<0>, SDTCisPtrTy<1>, SDTCisVT<2, i32>, SDTCisVec<3>,
6028                                         SDTCisSameAs<3, 4>, SDTCisSameAs<3, 5>,
6029                                         SDTCisSameAs<3, 6>, SDTCisVT<7, i32>]>;
6030def MVEVST2UPD       : SDNode<"ARMISD::VST2_UPD", SDTARMVST2, [SDNPHasChain, SDNPMemOperand]>;
6031def MVEVST4UPD       : SDNode<"ARMISD::VST4_UPD", SDTARMVST4, [SDNPHasChain, SDNPMemOperand]>;
6032
6033multiclass MVE_vst24_patterns<int lanesize, ValueType VT> {
6034  foreach stage = [0,1] in
6035    def : Pat<(int_arm_mve_vst2q i32:$addr,
6036                (VT MQPR:$v0), (VT MQPR:$v1), (i32 stage)),
6037              (!cast<Instruction>("MVE_VST2"#stage#"_"#lanesize)
6038                (REG_SEQUENCE QQPR, VT:$v0, qsub_0, VT:$v1, qsub_1),
6039                t2_addr_offset_none:$addr)>;
6040  foreach stage = [0,1] in
6041    def : Pat<(i32 (MVEVST2UPD i32:$addr, (i32 32),
6042                (VT MQPR:$v0), (VT MQPR:$v1), (i32 stage))),
6043              (i32 (!cast<Instruction>("MVE_VST2"#stage#"_"#lanesize#_wb)
6044                (REG_SEQUENCE QQPR, VT:$v0, qsub_0, VT:$v1, qsub_1),
6045                t2_addr_offset_none:$addr))>;
6046
6047  foreach stage = [0,1,2,3] in
6048    def : Pat<(int_arm_mve_vst4q i32:$addr,
6049                (VT MQPR:$v0), (VT MQPR:$v1),
6050                (VT MQPR:$v2), (VT MQPR:$v3), (i32 stage)),
6051              (!cast<Instruction>("MVE_VST4"#stage#"_"#lanesize)
6052                (REG_SEQUENCE QQQQPR, VT:$v0, qsub_0, VT:$v1, qsub_1,
6053                                      VT:$v2, qsub_2, VT:$v3, qsub_3),
6054                t2_addr_offset_none:$addr)>;
6055  foreach stage = [0,1,2,3] in
6056    def : Pat<(i32 (MVEVST4UPD i32:$addr, (i32 64),
6057                (VT MQPR:$v0), (VT MQPR:$v1),
6058                (VT MQPR:$v2), (VT MQPR:$v3), (i32 stage))),
6059              (i32 (!cast<Instruction>("MVE_VST4"#stage#"_"#lanesize#_wb)
6060                (REG_SEQUENCE QQQQPR, VT:$v0, qsub_0, VT:$v1, qsub_1,
6061                                      VT:$v2, qsub_2, VT:$v3, qsub_3),
6062                t2_addr_offset_none:$addr))>;
6063}
6064defm : MVE_vst24_patterns<8, v16i8>;
6065defm : MVE_vst24_patterns<16, v8i16>;
6066defm : MVE_vst24_patterns<32, v4i32>;
6067defm : MVE_vst24_patterns<16, v8f16>;
6068defm : MVE_vst24_patterns<32, v4f32>;
6069
6070// end of MVE interleaving load/store
6071
6072// start of MVE predicable load/store
6073
6074// A parameter class for the direction of transfer.
6075class MVE_ldst_direction<bit b, dag Oo, dag Io, string c=""> {
6076  bit load = b;
6077  dag Oops = Oo;
6078  dag Iops = Io;
6079  string cstr = c;
6080}
6081def MVE_ld: MVE_ldst_direction<1, (outs MQPR:$Qd), (ins), ",@earlyclobber $Qd">;
6082def MVE_st: MVE_ldst_direction<0, (outs), (ins MQPR:$Qd)>;
6083
6084// A parameter class for the size of memory access in a load.
6085class MVE_memsz<bits<2> e, int s, AddrMode m, string mn, list<string> types> {
6086  bits<2> encoding = e;         // opcode bit(s) for encoding
6087  int shift = s;                // shift applied to immediate load offset
6088  AddrMode AM = m;
6089
6090  // For instruction aliases: define the complete list of type
6091  // suffixes at this size, and the canonical ones for loads and
6092  // stores.
6093  string MnemonicLetter = mn;
6094  int TypeBits = !shl(8, s);
6095  string CanonLoadSuffix = ".u" # TypeBits;
6096  string CanonStoreSuffix = "." # TypeBits;
6097  list<string> suffixes = !foreach(letter, types, "." # letter # TypeBits);
6098}
6099
6100// Instances of MVE_memsz.
6101//
6102// (memD doesn't need an AddrMode, because those are only for
6103// contiguous loads, and memD is only used by gather/scatters.)
6104def MVE_memB: MVE_memsz<0b00, 0, AddrModeT2_i7,   "b", ["", "u", "s"]>;
6105def MVE_memH: MVE_memsz<0b01, 1, AddrModeT2_i7s2, "h", ["", "u", "s", "f"]>;
6106def MVE_memW: MVE_memsz<0b10, 2, AddrModeT2_i7s4, "w", ["", "u", "s", "f"]>;
6107def MVE_memD: MVE_memsz<0b11, 3, ?,               "d", ["", "u", "s", "f"]>;
6108
6109// This is the base class for all the MVE loads and stores other than
6110// the interleaving ones. All the non-interleaving loads/stores share
6111// the characteristic that they operate on just one vector register,
6112// so they are VPT-predicable.
6113//
6114// The predication operand is vpred_n, for both loads and stores. For
6115// store instructions, the reason is obvious: if there is no output
6116// register, there can't be a need for an input parameter giving the
6117// output register's previous value. Load instructions also don't need
6118// that input parameter, because unlike MVE data processing
6119// instructions, predicated loads are defined to set the inactive
6120// lanes of the output register to zero, instead of preserving their
6121// input values.
6122class MVE_VLDRSTR_base<MVE_ldst_direction dir, bit U, bit P, bit W, bit opc,
6123                       dag oops, dag iops, string asm, string suffix,
6124                       string ops, string cstr, list<dag> pattern=[]>
6125 : MVE_p<oops, iops, NoItinerary, asm, suffix, ops, vpred_n, cstr, pattern> {
6126  bits<3> Qd;
6127
6128  let Inst{28} = U;
6129  let Inst{25} = 0b0;
6130  let Inst{24} = P;
6131  let Inst{22} = 0b0;
6132  let Inst{21} = W;
6133  let Inst{20} = dir.load;
6134  let Inst{15-13} = Qd{2-0};
6135  let Inst{12} = opc;
6136  let Inst{11-9} = 0b111;
6137
6138  let mayLoad = dir.load;
6139  let mayStore = !eq(dir.load,0);
6140  let hasSideEffects = 0;
6141  let validForTailPredication = 1;
6142}
6143
6144// Contiguous load and store instructions. These come in two main
6145// categories: same-size loads/stores in which 128 bits of vector
6146// register is transferred to or from 128 bits of memory in the most
6147// obvious way, and widening loads / narrowing stores, in which the
6148// size of memory accessed is less than the size of a vector register,
6149// so the load instructions sign- or zero-extend each memory value
6150// into a wider vector lane, and the store instructions truncate
6151// correspondingly.
6152//
6153// The instruction mnemonics for these two classes look reasonably
6154// similar, but the actual encodings are different enough to need two
6155// separate base classes.
6156
6157// Contiguous, same size
6158class MVE_VLDRSTR_cs<MVE_ldst_direction dir, MVE_memsz memsz, bit P, bit W,
6159                     dag oops, dag iops, string asm, string suffix,
6160                     IndexMode im, string ops, string cstr>
6161  : MVE_VLDRSTR_base<dir, 0, P, W, 1, oops, iops, asm, suffix, ops, cstr> {
6162  bits<12> addr;
6163  let Inst{23} = addr{7};
6164  let Inst{19-16} = addr{11-8};
6165  let Inst{8-7} = memsz.encoding;
6166  let Inst{6-0} = addr{6-0};
6167}
6168
6169// Contiguous, widening/narrowing
6170class MVE_VLDRSTR_cw<MVE_ldst_direction dir, MVE_memsz memsz, bit U,
6171                     bit P, bit W, bits<2> size, dag oops, dag iops,
6172                     string asm, string suffix, IndexMode im,
6173                     string ops, string cstr>
6174  : MVE_VLDRSTR_base<dir, U, P, W, 0, oops, iops, asm, suffix, ops, cstr> {
6175  bits<11> addr;
6176  let Inst{23} = addr{7};
6177  let Inst{19} = memsz.encoding{0}; // enough to tell 16- from 32-bit
6178  let Inst{18-16} = addr{10-8};
6179  let Inst{8-7} = size;
6180  let Inst{6-0} = addr{6-0};
6181
6182  let IM = im;
6183}
6184
6185// Multiclass wrapper on each of the _cw and _cs base classes, to
6186// generate three writeback modes (none, preindex, postindex).
6187
6188multiclass MVE_VLDRSTR_cw_m<MVE_ldst_direction dir, MVE_memsz memsz,
6189                            string asm, string suffix, bit U, bits<2> size> {
6190  let AM = memsz.AM in {
6191    def "" : MVE_VLDRSTR_cw<
6192        dir, memsz, U, 1, 0, size,
6193        dir.Oops, !con(dir.Iops, (ins taddrmode_imm7<memsz.shift>:$addr)),
6194        asm, suffix, IndexModeNone, "$Qd, $addr", "">;
6195
6196    def _pre : MVE_VLDRSTR_cw<
6197        dir, memsz, U, 1, 1, size,
6198        !con((outs tGPR:$wb), dir.Oops),
6199        !con(dir.Iops, (ins taddrmode_imm7<memsz.shift>:$addr)),
6200        asm, suffix, IndexModePre, "$Qd, $addr!", "$addr.base = $wb"> {
6201      let DecoderMethod = "DecodeMVE_MEM_1_pre<"#memsz.shift#">";
6202    }
6203
6204    def _post : MVE_VLDRSTR_cw<
6205        dir, memsz, U, 0, 1, size,
6206        !con((outs tGPR:$wb), dir.Oops),
6207        !con(dir.Iops, (ins t_addr_offset_none:$Rn,
6208                            t2am_imm7_offset<memsz.shift>:$addr)),
6209        asm, suffix, IndexModePost, "$Qd, $Rn$addr", "$Rn.base = $wb"> {
6210      bits<4> Rn;
6211      let Inst{18-16} = Rn{2-0};
6212    }
6213  }
6214}
6215
6216multiclass MVE_VLDRSTR_cs_m<MVE_ldst_direction dir, MVE_memsz memsz,
6217                            string asm, string suffix> {
6218  let AM = memsz.AM in {
6219    def "" : MVE_VLDRSTR_cs<
6220        dir, memsz, 1, 0,
6221        dir.Oops, !con(dir.Iops, (ins t2addrmode_imm7<memsz.shift>:$addr)),
6222        asm, suffix, IndexModeNone, "$Qd, $addr", "">;
6223
6224    def _pre : MVE_VLDRSTR_cs<
6225        dir, memsz, 1, 1,
6226        !con((outs rGPR:$wb), dir.Oops),
6227        !con(dir.Iops, (ins t2addrmode_imm7_pre<memsz.shift>:$addr)),
6228        asm, suffix, IndexModePre, "$Qd, $addr!", "$addr.base = $wb"> {
6229      let DecoderMethod = "DecodeMVE_MEM_2_pre<"#memsz.shift#">";
6230    }
6231
6232    def _post : MVE_VLDRSTR_cs<
6233        dir, memsz, 0, 1,
6234        !con((outs rGPR:$wb), dir.Oops),
6235        !con(dir.Iops, (ins t2_nosp_addr_offset_none:$Rn,
6236                            t2am_imm7_offset<memsz.shift>:$addr)),
6237        asm, suffix, IndexModePost, "$Qd, $Rn$addr", "$Rn.base = $wb"> {
6238      bits<4> Rn;
6239      let Inst{19-16} = Rn{3-0};
6240    }
6241  }
6242}
6243
6244// Now actually declare all the contiguous load/stores, via those
6245// multiclasses. The instruction ids coming out of this are the bare
6246// names shown in the defm, with _pre or _post appended for writeback,
6247// e.g. MVE_VLDRBS16, MVE_VSTRB16_pre, MVE_VSTRHU16_post.
6248
6249defm MVE_VLDRBS16: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memB, "vldrb", "s16", 0, 0b01>;
6250defm MVE_VLDRBS32: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memB, "vldrb", "s32", 0, 0b10>;
6251defm MVE_VLDRBU16: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memB, "vldrb", "u16", 1, 0b01>;
6252defm MVE_VLDRBU32: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memB, "vldrb", "u32", 1, 0b10>;
6253defm MVE_VLDRHS32: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memH, "vldrh", "s32", 0, 0b10>;
6254defm MVE_VLDRHU32: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memH, "vldrh", "u32", 1, 0b10>;
6255
6256defm MVE_VLDRBU8:  MVE_VLDRSTR_cs_m<MVE_ld, MVE_memB, "vldrb", "u8">;
6257defm MVE_VLDRHU16: MVE_VLDRSTR_cs_m<MVE_ld, MVE_memH, "vldrh", "u16">;
6258defm MVE_VLDRWU32: MVE_VLDRSTR_cs_m<MVE_ld, MVE_memW, "vldrw", "u32">;
6259
6260defm MVE_VSTRB16:  MVE_VLDRSTR_cw_m<MVE_st, MVE_memB, "vstrb", "16",  0, 0b01>;
6261defm MVE_VSTRB32:  MVE_VLDRSTR_cw_m<MVE_st, MVE_memB, "vstrb", "32",  0, 0b10>;
6262defm MVE_VSTRH32:  MVE_VLDRSTR_cw_m<MVE_st, MVE_memH, "vstrh", "32",  0, 0b10>;
6263
6264defm MVE_VSTRBU8 : MVE_VLDRSTR_cs_m<MVE_st, MVE_memB, "vstrb", "8">;
6265defm MVE_VSTRHU16: MVE_VLDRSTR_cs_m<MVE_st, MVE_memH, "vstrh", "16">;
6266defm MVE_VSTRWU32: MVE_VLDRSTR_cs_m<MVE_st, MVE_memW, "vstrw", "32">;
6267
6268// Gather loads / scatter stores whose address operand is of the form
6269// [Rn,Qm], i.e. a single GPR as the common base address, plus a
6270// vector of offset from it. ('Load/store this sequence of elements of
6271// the same array.')
6272//
6273// Like the contiguous family, these loads and stores can widen the
6274// loaded values / truncate the stored ones, or they can just
6275// load/store the same size of memory and vector lane. But unlike the
6276// contiguous family, there's no particular difference in encoding
6277// between those two cases.
6278//
6279// This family also comes with the option to scale the offset values
6280// in Qm by the size of the loaded memory (i.e. to treat them as array
6281// indices), or not to scale them (to treat them as plain byte offsets
6282// in memory, so that perhaps the loaded values are unaligned). The
6283// scaled instructions' address operand in assembly looks like
6284// [Rn,Qm,UXTW #2] or similar.
6285
6286// Base class.
6287class MVE_VLDRSTR_rq<MVE_ldst_direction dir, MVE_memsz memsz, bit U,
6288                     bits<2> size, bit os, string asm, string suffix, int shift>
6289  : MVE_VLDRSTR_base<dir, U, 0b0, 0b0, 0, dir.Oops,
6290                     !con(dir.Iops, (ins mve_addr_rq_shift<shift>:$addr)),
6291                     asm, suffix, "$Qd, $addr", dir.cstr> {
6292  bits<7> addr;
6293  let Inst{23} = 0b1;
6294  let Inst{19-16} = addr{6-3};
6295  let Inst{8-7} = size;
6296  let Inst{6} = memsz.encoding{1};
6297  let Inst{5} = 0;
6298  let Inst{4} = memsz.encoding{0};
6299  let Inst{3-1} = addr{2-0};
6300  let Inst{0} = os;
6301}
6302
6303// Multiclass that defines the scaled and unscaled versions of an
6304// instruction, when the memory size is wider than a byte. The scaled
6305// version gets the default name like MVE_VLDRBU16_rq; the unscaled /
6306// potentially unaligned version gets a "_u" suffix, e.g.
6307// MVE_VLDRBU16_rq_u.
6308multiclass MVE_VLDRSTR_rq_w<MVE_ldst_direction dir, MVE_memsz memsz,
6309                            string asm, string suffix, bit U, bits<2> size> {
6310  def _u : MVE_VLDRSTR_rq<dir, memsz, U, size, 0, asm, suffix, 0>;
6311  def "" : MVE_VLDRSTR_rq<dir, memsz, U, size, 1, asm, suffix, memsz.shift>;
6312}
6313
6314// Subclass of MVE_VLDRSTR_rq with the same API as that multiclass,
6315// for use when the memory size is one byte, so there's no 'scaled'
6316// version of the instruction at all. (This is encoded as if it were
6317// unscaled, but named in the default way with no _u suffix.)
6318class MVE_VLDRSTR_rq_b<MVE_ldst_direction dir, MVE_memsz memsz,
6319                       string asm, string suffix, bit U, bits<2> size>
6320  : MVE_VLDRSTR_rq<dir, memsz, U, size, 0, asm, suffix, 0>;
6321
6322// Multiclasses wrapping that to add ISel patterns for intrinsics.
6323multiclass MVE_VLDR_rq_w<MVE_memsz memsz, list<MVEVectorVTInfo> VTIs> {
6324  defm "": MVE_VLDRSTR_rq_w<MVE_ld, memsz, "vldr" # memsz.MnemonicLetter,
6325                            VTIs[0].Suffix, VTIs[0].Unsigned, VTIs[0].Size>;
6326  defvar Inst = !cast<Instruction>(NAME);
6327  defvar InstU = !cast<Instruction>(NAME # "_u");
6328
6329  foreach VTI = VTIs in
6330  foreach UnsignedFlag = !if(!eq(VTI.Size, memsz.encoding),
6331                             [0,1], [VTI.Unsigned]) in {
6332    def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), memsz.TypeBits, 0, UnsignedFlag)),
6333              (VTI.Vec (InstU GPR:$base, MQPR:$offsets))>;
6334    def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), memsz.TypeBits, memsz.shift, UnsignedFlag)),
6335              (VTI.Vec (Inst GPR:$base, MQPR:$offsets))>;
6336    def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), memsz.TypeBits, 0, UnsignedFlag, (VTI.Pred VCCR:$pred))),
6337              (VTI.Vec (InstU GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred))>;
6338    def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), memsz.TypeBits, memsz.shift, UnsignedFlag, (VTI.Pred VCCR:$pred))),
6339              (VTI.Vec (Inst GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred))>;
6340  }
6341}
6342multiclass MVE_VLDR_rq_b<list<MVEVectorVTInfo> VTIs> {
6343  def "": MVE_VLDRSTR_rq_b<MVE_ld, MVE_memB, "vldrb",
6344                           VTIs[0].Suffix, VTIs[0].Unsigned, VTIs[0].Size>;
6345  defvar Inst = !cast<Instruction>(NAME);
6346
6347  foreach VTI = VTIs in {
6348    def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), 8, 0, VTI.Unsigned)),
6349              (VTI.Vec (Inst GPR:$base, MQPR:$offsets))>;
6350    def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), 8, 0, VTI.Unsigned, (VTI.Pred VCCR:$pred))),
6351              (VTI.Vec (Inst GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred))>;
6352  }
6353}
6354multiclass MVE_VSTR_rq_w<MVE_memsz memsz, list<MVEVectorVTInfo> VTIs> {
6355  defm "": MVE_VLDRSTR_rq_w<MVE_st, memsz, "vstr" # memsz.MnemonicLetter,
6356                            VTIs[0].BitsSuffix, 0, VTIs[0].Size>;
6357  defvar Inst = !cast<Instruction>(NAME);
6358  defvar InstU = !cast<Instruction>(NAME # "_u");
6359
6360  foreach VTI = VTIs in {
6361    def : Pat<(int_arm_mve_vstr_scatter_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), memsz.TypeBits, 0),
6362              (InstU MQPR:$data, GPR:$base, MQPR:$offsets)>;
6363    def : Pat<(int_arm_mve_vstr_scatter_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), memsz.TypeBits, memsz.shift),
6364              (Inst MQPR:$data, GPR:$base, MQPR:$offsets)>;
6365    def : Pat<(int_arm_mve_vstr_scatter_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), memsz.TypeBits, 0, (VTI.Pred VCCR:$pred)),
6366              (InstU MQPR:$data, GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred)>;
6367    def : Pat<(int_arm_mve_vstr_scatter_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), memsz.TypeBits, memsz.shift, (VTI.Pred VCCR:$pred)),
6368              (Inst MQPR:$data, GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred)>;
6369  }
6370}
6371multiclass MVE_VSTR_rq_b<list<MVEVectorVTInfo> VTIs> {
6372  def "": MVE_VLDRSTR_rq_b<MVE_st, MVE_memB, "vstrb",
6373                           VTIs[0].BitsSuffix, 0, VTIs[0].Size>;
6374  defvar Inst = !cast<Instruction>(NAME);
6375
6376  foreach VTI = VTIs in {
6377    def : Pat<(int_arm_mve_vstr_scatter_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), 8, 0),
6378              (Inst MQPR:$data, GPR:$base, MQPR:$offsets)>;
6379    def : Pat<(int_arm_mve_vstr_scatter_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), 8, 0, (VTI.Pred VCCR:$pred)),
6380              (Inst MQPR:$data, GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred)>;
6381  }
6382}
6383
6384// Actually define all the loads and stores in this family.
6385
6386defm MVE_VLDRBU8_rq : MVE_VLDR_rq_b<[MVE_v16u8,MVE_v16s8]>;
6387defm MVE_VLDRBU16_rq: MVE_VLDR_rq_b<[MVE_v8u16]>;
6388defm MVE_VLDRBS16_rq: MVE_VLDR_rq_b<[MVE_v8s16]>;
6389defm MVE_VLDRBU32_rq: MVE_VLDR_rq_b<[MVE_v4u32]>;
6390defm MVE_VLDRBS32_rq: MVE_VLDR_rq_b<[MVE_v4s32]>;
6391
6392defm MVE_VLDRHU16_rq: MVE_VLDR_rq_w<MVE_memH, [MVE_v8u16,MVE_v8s16,MVE_v8f16]>;
6393defm MVE_VLDRHU32_rq: MVE_VLDR_rq_w<MVE_memH, [MVE_v4u32]>;
6394defm MVE_VLDRHS32_rq: MVE_VLDR_rq_w<MVE_memH, [MVE_v4s32]>;
6395defm MVE_VLDRWU32_rq: MVE_VLDR_rq_w<MVE_memW, [MVE_v4u32,MVE_v4s32,MVE_v4f32]>;
6396defm MVE_VLDRDU64_rq: MVE_VLDR_rq_w<MVE_memD, [MVE_v2u64,MVE_v2s64]>;
6397
6398defm MVE_VSTRB8_rq  : MVE_VSTR_rq_b<[MVE_v16i8]>;
6399defm MVE_VSTRB16_rq : MVE_VSTR_rq_b<[MVE_v8i16]>;
6400defm MVE_VSTRB32_rq : MVE_VSTR_rq_b<[MVE_v4i32]>;
6401
6402defm MVE_VSTRH16_rq : MVE_VSTR_rq_w<MVE_memH, [MVE_v8i16,MVE_v8f16]>;
6403defm MVE_VSTRH32_rq : MVE_VSTR_rq_w<MVE_memH, [MVE_v4i32]>;
6404defm MVE_VSTRW32_rq : MVE_VSTR_rq_w<MVE_memW, [MVE_v4i32,MVE_v4f32]>;
6405defm MVE_VSTRD64_rq : MVE_VSTR_rq_w<MVE_memD, [MVE_v2i64]>;
6406
6407// Gather loads / scatter stores whose address operand is of the form
6408// [Qm,#imm], i.e. a vector containing a full base address for each
6409// loaded item, plus an immediate offset applied consistently to all
6410// of them. ('Load/store the same field from this vector of pointers
6411// to a structure type.')
6412//
6413// This family requires the vector lane size to be at least 32 bits
6414// (so there's room for an address in each lane at all). It has no
6415// widening/narrowing variants. But it does support preindex
6416// writeback, in which the address vector is updated to hold the
6417// addresses actually loaded from.
6418
6419// Base class.
6420class MVE_VLDRSTR_qi<MVE_ldst_direction dir, MVE_memsz memsz, bit W, dag wbops,
6421                     string asm, string wbAsm, string suffix, string cstr = "">
6422  : MVE_VLDRSTR_base<dir, 1, 1, W, 1, !con(wbops, dir.Oops),
6423                     !con(dir.Iops, (ins mve_addr_q_shift<memsz.shift>:$addr)),
6424                     asm, suffix, "$Qd, $addr" # wbAsm, cstr # dir.cstr> {
6425  bits<11> addr;
6426  let Inst{23} = addr{7};
6427  let Inst{19-17} = addr{10-8};
6428  let Inst{16} = 0;
6429  let Inst{8} = memsz.encoding{0}; // enough to distinguish 32- from 64-bit
6430  let Inst{7} = 0;
6431  let Inst{6-0} = addr{6-0};
6432}
6433
6434// Multiclass that generates the non-writeback and writeback variants.
6435multiclass MVE_VLDRSTR_qi_m<MVE_ldst_direction dir, MVE_memsz memsz,
6436                            string asm, string suffix> {
6437  def ""   : MVE_VLDRSTR_qi<dir, memsz, 0, (outs),          asm, "",  suffix>;
6438  def _pre : MVE_VLDRSTR_qi<dir, memsz, 1, (outs MQPR:$wb), asm, "!", suffix,
6439                            "$addr.base = $wb"> {
6440    let DecoderMethod="DecodeMVE_MEM_3_pre<"#memsz.shift#">";
6441  }
6442}
6443
6444// Multiclasses wrapping that one, adding selection patterns for the
6445// non-writeback loads and all the stores. (The writeback loads must
6446// deliver multiple output values, so they have to be selected by C++
6447// code.)
6448multiclass MVE_VLDR_qi<MVE_memsz memsz, MVEVectorVTInfo AVTI,
6449                       list<MVEVectorVTInfo> DVTIs> {
6450  defm "" : MVE_VLDRSTR_qi_m<MVE_ld, memsz, "vldr" # memsz.MnemonicLetter,
6451                             "u" # memsz.TypeBits>;
6452  defvar Inst = !cast<Instruction>(NAME);
6453
6454  foreach DVTI = DVTIs in {
6455    def : Pat<(DVTI.Vec (int_arm_mve_vldr_gather_base
6456                 (AVTI.Vec MQPR:$addr), (i32 imm:$offset))),
6457              (DVTI.Vec (Inst (AVTI.Vec MQPR:$addr), (i32 imm:$offset)))>;
6458    def : Pat<(DVTI.Vec (int_arm_mve_vldr_gather_base_predicated
6459                 (AVTI.Vec MQPR:$addr), (i32 imm:$offset), (AVTI.Pred VCCR:$pred))),
6460              (DVTI.Vec (Inst (AVTI.Vec MQPR:$addr), (i32 imm:$offset),
6461                        ARMVCCThen, VCCR:$pred))>;
6462  }
6463}
6464multiclass MVE_VSTR_qi<MVE_memsz memsz, MVEVectorVTInfo AVTI,
6465                       list<MVEVectorVTInfo> DVTIs> {
6466  defm "" : MVE_VLDRSTR_qi_m<MVE_st, memsz, "vstr" # memsz.MnemonicLetter,
6467                             !cast<string>(memsz.TypeBits)>;
6468  defvar Inst = !cast<Instruction>(NAME);
6469  defvar InstPre = !cast<Instruction>(NAME # "_pre");
6470
6471  foreach DVTI = DVTIs in {
6472    def : Pat<(int_arm_mve_vstr_scatter_base
6473                (AVTI.Vec MQPR:$addr), (i32 imm:$offset), (DVTI.Vec MQPR:$data)),
6474              (Inst (DVTI.Vec MQPR:$data), (AVTI.Vec MQPR:$addr),
6475                    (i32 imm:$offset))>;
6476    def : Pat<(int_arm_mve_vstr_scatter_base_predicated
6477                (AVTI.Vec MQPR:$addr), (i32 imm:$offset), (DVTI.Vec MQPR:$data), (AVTI.Pred VCCR:$pred)),
6478              (Inst (DVTI.Vec MQPR:$data), (AVTI.Vec MQPR:$addr),
6479                    (i32 imm:$offset), ARMVCCThen, VCCR:$pred)>;
6480    def : Pat<(AVTI.Vec (int_arm_mve_vstr_scatter_base_wb
6481                (AVTI.Vec MQPR:$addr), (i32 imm:$offset), (DVTI.Vec MQPR:$data))),
6482              (AVTI.Vec (InstPre (DVTI.Vec MQPR:$data), (AVTI.Vec MQPR:$addr),
6483                                 (i32 imm:$offset)))>;
6484    def : Pat<(AVTI.Vec (int_arm_mve_vstr_scatter_base_wb_predicated
6485                (AVTI.Vec MQPR:$addr), (i32 imm:$offset), (DVTI.Vec MQPR:$data), (AVTI.Pred VCCR:$pred))),
6486              (AVTI.Vec (InstPre (DVTI.Vec MQPR:$data), (AVTI.Vec MQPR:$addr),
6487                                 (i32 imm:$offset), ARMVCCThen, VCCR:$pred))>;
6488  }
6489}
6490
6491// Actual instruction definitions.
6492defm MVE_VLDRWU32_qi: MVE_VLDR_qi<MVE_memW, MVE_v4i32, [MVE_v4i32,MVE_v4f32]>;
6493defm MVE_VLDRDU64_qi: MVE_VLDR_qi<MVE_memD, MVE_v2i64, [MVE_v2i64,MVE_v2f64]>;
6494defm MVE_VSTRW32_qi:  MVE_VSTR_qi<MVE_memW, MVE_v4i32, [MVE_v4i32,MVE_v4f32]>;
6495defm MVE_VSTRD64_qi:  MVE_VSTR_qi<MVE_memD, MVE_v2i64, [MVE_v2i64,MVE_v2f64]>;
6496
6497// Define aliases for all the instructions where memory size and
6498// vector lane size are the same. These are mnemonic aliases, so they
6499// apply consistently across all of the above families - contiguous
6500// loads, and both the rq and qi types of gather/scatter.
6501//
6502// Rationale: As long as you're loading (for example) 16-bit memory
6503// values into 16-bit vector lanes, you can think of them as signed or
6504// unsigned integers, fp16 or just raw 16-bit blobs and it makes no
6505// difference. So we permit all of vldrh.16, vldrh.u16, vldrh.s16,
6506// vldrh.f16 and treat them all as equivalent to the canonical
6507// spelling (which happens to be .u16 for loads, and just .16 for
6508// stores).
6509
6510foreach vpt_cond = ["", "t", "e"] in
6511foreach memsz = [MVE_memB, MVE_memH, MVE_memW, MVE_memD] in
6512foreach suffix = memsz.suffixes in {
6513  // Define an alias with every suffix in the list, except for the one
6514  // used by the real Instruction record (i.e. the one that all the
6515  // rest are aliases *for*).
6516
6517  if !ne(suffix, memsz.CanonLoadSuffix) then {
6518    def : MnemonicAlias<
6519      "vldr" # memsz.MnemonicLetter # vpt_cond # suffix,
6520      "vldr" # memsz.MnemonicLetter # vpt_cond # memsz.CanonLoadSuffix>;
6521  }
6522
6523  if !ne(suffix, memsz.CanonStoreSuffix) then {
6524    def : MnemonicAlias<
6525      "vstr" # memsz.MnemonicLetter # vpt_cond # suffix,
6526      "vstr" # memsz.MnemonicLetter # vpt_cond # memsz.CanonStoreSuffix>;
6527  }
6528}
6529
6530// end of MVE predicable load/store
6531
6532class MVE_VPT<string suffix, bits<2> size, dag iops, string asm, list<dag> pattern=[]>
6533  : MVE_MI<(outs ), iops, NoItinerary, !strconcat("vpt", "${Mk}", ".", suffix), asm, "", pattern> {
6534  bits<3> fc;
6535  bits<4> Mk;
6536  bits<3> Qn;
6537
6538  let Inst{31-23} = 0b111111100;
6539  let Inst{22} = Mk{3};
6540  let Inst{21-20} = size;
6541  let Inst{19-17} = Qn{2-0};
6542  let Inst{16} = 0b1;
6543  let Inst{15-13} = Mk{2-0};
6544  let Inst{12} = fc{2};
6545  let Inst{11-8} = 0b1111;
6546  let Inst{7} = fc{0};
6547  let Inst{4} = 0b0;
6548
6549  let Defs = [VPR];
6550  let validForTailPredication=1;
6551}
6552
6553class MVE_VPTt1<string suffix, bits<2> size, dag iops>
6554  : MVE_VPT<suffix, size, iops, "$fc, $Qn, $Qm"> {
6555  bits<4> Qm;
6556  bits<4> Mk;
6557
6558  let Inst{6} = 0b0;
6559  let Inst{5} = Qm{3};
6560  let Inst{3-1} = Qm{2-0};
6561  let Inst{0} = fc{1};
6562}
6563
6564class MVE_VPTt1i<string suffix, bits<2> size>
6565 : MVE_VPTt1<suffix, size,
6566           (ins vpt_mask:$Mk, MQPR:$Qn, MQPR:$Qm, pred_basic_i:$fc)> {
6567  let Inst{12} = 0b0;
6568  let Inst{0} = 0b0;
6569}
6570
6571def MVE_VPTv4i32 : MVE_VPTt1i<"i32", 0b10>;
6572def MVE_VPTv8i16 : MVE_VPTt1i<"i16", 0b01>;
6573def MVE_VPTv16i8 : MVE_VPTt1i<"i8", 0b00>;
6574
6575class MVE_VPTt1u<string suffix, bits<2> size>
6576 : MVE_VPTt1<suffix, size,
6577           (ins vpt_mask:$Mk, MQPR:$Qn, MQPR:$Qm, pred_basic_u:$fc)> {
6578  let Inst{12} = 0b0;
6579  let Inst{0} = 0b1;
6580}
6581
6582def MVE_VPTv4u32 : MVE_VPTt1u<"u32", 0b10>;
6583def MVE_VPTv8u16 : MVE_VPTt1u<"u16", 0b01>;
6584def MVE_VPTv16u8 : MVE_VPTt1u<"u8", 0b00>;
6585
6586class MVE_VPTt1s<string suffix, bits<2> size>
6587 : MVE_VPTt1<suffix, size,
6588           (ins vpt_mask:$Mk, MQPR:$Qn, MQPR:$Qm, pred_basic_s:$fc)> {
6589  let Inst{12} = 0b1;
6590}
6591
6592def MVE_VPTv4s32 : MVE_VPTt1s<"s32", 0b10>;
6593def MVE_VPTv8s16 : MVE_VPTt1s<"s16", 0b01>;
6594def MVE_VPTv16s8 : MVE_VPTt1s<"s8", 0b00>;
6595
6596class MVE_VPTt2<string suffix, bits<2> size, dag iops>
6597  : MVE_VPT<suffix, size, iops,
6598          "$fc, $Qn, $Rm"> {
6599  bits<4> Rm;
6600  bits<3> fc;
6601  bits<4> Mk;
6602
6603  let Inst{6} = 0b1;
6604  let Inst{5} = fc{1};
6605  let Inst{3-0} = Rm{3-0};
6606}
6607
6608class MVE_VPTt2i<string suffix, bits<2> size>
6609  : MVE_VPTt2<suffix, size,
6610            (ins vpt_mask:$Mk, MQPR:$Qn, GPRwithZR:$Rm, pred_basic_i:$fc)> {
6611  let Inst{12} = 0b0;
6612  let Inst{5} = 0b0;
6613}
6614
6615def MVE_VPTv4i32r : MVE_VPTt2i<"i32", 0b10>;
6616def MVE_VPTv8i16r : MVE_VPTt2i<"i16", 0b01>;
6617def MVE_VPTv16i8r : MVE_VPTt2i<"i8", 0b00>;
6618
6619class MVE_VPTt2u<string suffix, bits<2> size>
6620  : MVE_VPTt2<suffix, size,
6621            (ins vpt_mask:$Mk, MQPR:$Qn, GPRwithZR:$Rm, pred_basic_u:$fc)> {
6622  let Inst{12} = 0b0;
6623  let Inst{5} = 0b1;
6624}
6625
6626def MVE_VPTv4u32r : MVE_VPTt2u<"u32", 0b10>;
6627def MVE_VPTv8u16r : MVE_VPTt2u<"u16", 0b01>;
6628def MVE_VPTv16u8r : MVE_VPTt2u<"u8", 0b00>;
6629
6630class MVE_VPTt2s<string suffix, bits<2> size>
6631  : MVE_VPTt2<suffix, size,
6632            (ins vpt_mask:$Mk, MQPR:$Qn, GPRwithZR:$Rm, pred_basic_s:$fc)> {
6633  let Inst{12} = 0b1;
6634}
6635
6636def MVE_VPTv4s32r : MVE_VPTt2s<"s32", 0b10>;
6637def MVE_VPTv8s16r : MVE_VPTt2s<"s16", 0b01>;
6638def MVE_VPTv16s8r : MVE_VPTt2s<"s8", 0b00>;
6639
6640
6641class MVE_VPTf<string suffix, bit size, dag iops, string asm, list<dag> pattern=[]>
6642  : MVE_MI<(outs ), iops, NoItinerary, !strconcat("vpt", "${Mk}", ".", suffix), asm,
6643            "", pattern> {
6644  bits<3> fc;
6645  bits<4> Mk;
6646  bits<3> Qn;
6647
6648  let Inst{31-29} = 0b111;
6649  let Inst{28} = size;
6650  let Inst{27-23} = 0b11100;
6651  let Inst{22} = Mk{3};
6652  let Inst{21-20} = 0b11;
6653  let Inst{19-17} = Qn{2-0};
6654  let Inst{16} = 0b1;
6655  let Inst{15-13} = Mk{2-0};
6656  let Inst{12} = fc{2};
6657  let Inst{11-8} = 0b1111;
6658  let Inst{7} = fc{0};
6659  let Inst{4} = 0b0;
6660
6661  let Defs = [VPR];
6662  let Predicates = [HasMVEFloat];
6663  let validForTailPredication=1;
6664}
6665
6666class MVE_VPTft1<string suffix, bit size>
6667  : MVE_VPTf<suffix, size, (ins vpt_mask:$Mk, MQPR:$Qn, MQPR:$Qm, pred_basic_fp:$fc),
6668          "$fc, $Qn, $Qm"> {
6669  bits<3> fc;
6670  bits<4> Qm;
6671
6672  let Inst{6} = 0b0;
6673  let Inst{5} = Qm{3};
6674  let Inst{3-1} = Qm{2-0};
6675  let Inst{0} = fc{1};
6676}
6677
6678def MVE_VPTv4f32         : MVE_VPTft1<"f32", 0b0>;
6679def MVE_VPTv8f16         : MVE_VPTft1<"f16", 0b1>;
6680
6681class MVE_VPTft2<string suffix, bit size>
6682  : MVE_VPTf<suffix, size, (ins vpt_mask:$Mk, MQPR:$Qn, GPRwithZR:$Rm, pred_basic_fp:$fc),
6683          "$fc, $Qn, $Rm"> {
6684  bits<3> fc;
6685  bits<4> Rm;
6686
6687  let Inst{6} = 0b1;
6688  let Inst{5} = fc{1};
6689  let Inst{3-0} = Rm{3-0};
6690}
6691
6692def MVE_VPTv4f32r        : MVE_VPTft2<"f32", 0b0>;
6693def MVE_VPTv8f16r        : MVE_VPTft2<"f16", 0b1>;
6694
6695def MVE_VPST : MVE_MI<(outs ), (ins vpt_mask:$Mk), NoItinerary,
6696       !strconcat("vpst", "${Mk}"), "", "", []> {
6697  bits<4> Mk;
6698
6699  let Inst{31-23} = 0b111111100;
6700  let Inst{22} = Mk{3};
6701  let Inst{21-16} = 0b110001;
6702  let Inst{15-13} = Mk{2-0};
6703  let Inst{12-0} = 0b0111101001101;
6704  let Unpredictable{12} = 0b1;
6705  let Unpredictable{7} = 0b1;
6706  let Unpredictable{5} = 0b1;
6707
6708  let Uses = [VPR];
6709  let validForTailPredication = 1;
6710}
6711
6712def MVE_VPSEL : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm), NoItinerary,
6713                      "vpsel", "", "$Qd, $Qn, $Qm", vpred_n, "", []> {
6714  bits<4> Qn;
6715  bits<4> Qd;
6716  bits<4> Qm;
6717
6718  let Inst{28} = 0b1;
6719  let Inst{25-23} = 0b100;
6720  let Inst{22} = Qd{3};
6721  let Inst{21-20} = 0b11;
6722  let Inst{19-17} = Qn{2-0};
6723  let Inst{16} = 0b1;
6724  let Inst{15-13} = Qd{2-0};
6725  let Inst{12-9} = 0b0111;
6726  let Inst{8} = 0b1;
6727  let Inst{7} = Qn{3};
6728  let Inst{6} = 0b0;
6729  let Inst{5} = Qm{3};
6730  let Inst{4} = 0b0;
6731  let Inst{3-1} = Qm{2-0};
6732  let Inst{0} = 0b1;
6733}
6734
6735foreach suffix = ["s8", "s16", "s32", "u8", "u16", "u32",
6736                  "i8", "i16", "i32",       "f16", "f32"] in
6737def : MVEInstAlias<"vpsel${vp}." # suffix # "\t$Qd, $Qn, $Qm",
6738                   (MVE_VPSEL MQPR:$Qd, MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
6739
6740let Predicates = [HasMVEInt] in {
6741  def : Pat<(v16i8 (vselect (v16i1 VCCR:$pred), (v16i8 MQPR:$v1), (v16i8 MQPR:$v2))),
6742            (v16i8 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone, VCCR:$pred))>;
6743  def : Pat<(v8i16 (vselect (v8i1 VCCR:$pred), (v8i16 MQPR:$v1), (v8i16 MQPR:$v2))),
6744            (v8i16 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone, VCCR:$pred))>;
6745  def : Pat<(v4i32 (vselect (v4i1 VCCR:$pred), (v4i32 MQPR:$v1), (v4i32 MQPR:$v2))),
6746            (v4i32 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone, VCCR:$pred))>;
6747
6748  def : Pat<(v8f16 (vselect (v8i1 VCCR:$pred), (v8f16 MQPR:$v1), (v8f16 MQPR:$v2))),
6749            (v8f16 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone, VCCR:$pred))>;
6750  def : Pat<(v4f32 (vselect (v4i1 VCCR:$pred), (v4f32 MQPR:$v1), (v4f32 MQPR:$v2))),
6751            (v4f32 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone, VCCR:$pred))>;
6752
6753  def : Pat<(v16i8 (vselect (v16i8 MQPR:$pred), (v16i8 MQPR:$v1), (v16i8 MQPR:$v2))),
6754            (v16i8 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone,
6755                              (MVE_VCMPi8 (v16i8 MQPR:$pred), (MVE_VMOVimmi8 0), ARMCCne)))>;
6756  def : Pat<(v8i16 (vselect (v8i16 MQPR:$pred), (v8i16 MQPR:$v1), (v8i16 MQPR:$v2))),
6757            (v8i16 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone,
6758                              (MVE_VCMPi16 (v8i16 MQPR:$pred), (MVE_VMOVimmi16 0), ARMCCne)))>;
6759  def : Pat<(v4i32 (vselect (v4i32 MQPR:$pred), (v4i32 MQPR:$v1), (v4i32 MQPR:$v2))),
6760            (v4i32 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone,
6761                              (MVE_VCMPi32 (v4i32 MQPR:$pred), (MVE_VMOVimmi32 0), ARMCCne)))>;
6762
6763  def : Pat<(v8f16 (vselect (v8i16 MQPR:$pred), (v8f16 MQPR:$v1), (v8f16 MQPR:$v2))),
6764            (v8f16 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone,
6765                              (MVE_VCMPi16 (v8i16 MQPR:$pred), (MVE_VMOVimmi16 0), ARMCCne)))>;
6766  def : Pat<(v4f32 (vselect (v4i32 MQPR:$pred), (v4f32 MQPR:$v1), (v4f32 MQPR:$v2))),
6767            (v4f32 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone,
6768                              (MVE_VCMPi32 (v4i32 MQPR:$pred), (MVE_VMOVimmi32 0), ARMCCne)))>;
6769
6770  // Pred <-> Int
6771  def : Pat<(v16i8 (zext  (v16i1 VCCR:$pred))),
6772            (v16i8 (MVE_VPSEL (MVE_VMOVimmi8 1), (MVE_VMOVimmi8 0), ARMVCCNone, VCCR:$pred))>;
6773  def : Pat<(v8i16 (zext  (v8i1  VCCR:$pred))),
6774            (v8i16 (MVE_VPSEL (MVE_VMOVimmi16 1), (MVE_VMOVimmi16 0), ARMVCCNone, VCCR:$pred))>;
6775  def : Pat<(v4i32 (zext  (v4i1  VCCR:$pred))),
6776            (v4i32 (MVE_VPSEL (MVE_VMOVimmi32 1), (MVE_VMOVimmi32 0), ARMVCCNone, VCCR:$pred))>;
6777
6778  def : Pat<(v16i8 (sext  (v16i1 VCCR:$pred))),
6779            (v16i8 (MVE_VPSEL (MVE_VMOVimmi8 255), (MVE_VMOVimmi8 0), ARMVCCNone, VCCR:$pred))>;
6780  def : Pat<(v8i16 (sext  (v8i1  VCCR:$pred))),
6781            (v8i16 (MVE_VPSEL (MVE_VMOVimmi8 255), (MVE_VMOVimmi16 0), ARMVCCNone, VCCR:$pred))>;
6782  def : Pat<(v4i32 (sext  (v4i1  VCCR:$pred))),
6783            (v4i32 (MVE_VPSEL (MVE_VMOVimmi8 255), (MVE_VMOVimmi32 0), ARMVCCNone, VCCR:$pred))>;
6784
6785  def : Pat<(v16i8 (anyext  (v16i1 VCCR:$pred))),
6786            (v16i8 (MVE_VPSEL (MVE_VMOVimmi8 1), (MVE_VMOVimmi8 0), ARMVCCNone, VCCR:$pred))>;
6787  def : Pat<(v8i16 (anyext  (v8i1  VCCR:$pred))),
6788            (v8i16 (MVE_VPSEL (MVE_VMOVimmi16 1), (MVE_VMOVimmi16 0), ARMVCCNone, VCCR:$pred))>;
6789  def : Pat<(v4i32 (anyext  (v4i1  VCCR:$pred))),
6790            (v4i32 (MVE_VPSEL (MVE_VMOVimmi32 1), (MVE_VMOVimmi32 0), ARMVCCNone, VCCR:$pred))>;
6791}
6792
6793let Predicates = [HasMVEFloat] in {
6794  // Pred <-> Float
6795  // 112 is 1.0 in float
6796  def : Pat<(v4f32 (uint_to_fp (v4i1 VCCR:$pred))),
6797            (v4f32 (MVE_VPSEL (v4f32 (MVE_VMOVimmf32 112)), (v4f32 (MVE_VMOVimmi32 0)), ARMVCCNone, VCCR:$pred))>;
6798  // 2620 in 1.0 in half
6799  def : Pat<(v8f16 (uint_to_fp (v8i1 VCCR:$pred))),
6800            (v8f16 (MVE_VPSEL (v8f16 (MVE_VMOVimmi16 2620)), (v8f16 (MVE_VMOVimmi16 0)), ARMVCCNone, VCCR:$pred))>;
6801  // 240 is -1.0 in float
6802  def : Pat<(v4f32 (sint_to_fp (v4i1 VCCR:$pred))),
6803            (v4f32 (MVE_VPSEL (v4f32 (MVE_VMOVimmf32 240)), (v4f32 (MVE_VMOVimmi32 0)), ARMVCCNone, VCCR:$pred))>;
6804  // 2748 is -1.0 in half
6805  def : Pat<(v8f16 (sint_to_fp (v8i1 VCCR:$pred))),
6806            (v8f16 (MVE_VPSEL (v8f16 (MVE_VMOVimmi16 2748)), (v8f16 (MVE_VMOVimmi16 0)), ARMVCCNone, VCCR:$pred))>;
6807
6808  def : Pat<(v4i1 (fp_to_uint (v4f32 MQPR:$v1))),
6809            (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), ZR, ARMCCne))>;
6810  def : Pat<(v8i1 (fp_to_uint (v8f16 MQPR:$v1))),
6811            (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), ZR, ARMCCne))>;
6812  def : Pat<(v4i1 (fp_to_sint (v4f32 MQPR:$v1))),
6813            (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), ZR, ARMCCne))>;
6814  def : Pat<(v8i1 (fp_to_sint (v8f16 MQPR:$v1))),
6815            (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), ZR, ARMCCne))>;
6816}
6817
6818def MVE_VPNOT : MVE_p<(outs VCCR:$P0), (ins VCCR:$P0_in), NoItinerary,
6819                      "vpnot", "", "", vpred_n, "", []> {
6820  let Inst{31-0} = 0b11111110001100010000111101001101;
6821  let Unpredictable{19-17} = 0b111;
6822  let Unpredictable{12} = 0b1;
6823  let Unpredictable{7} = 0b1;
6824  let Unpredictable{5} = 0b1;
6825
6826  let Constraints = "";
6827  let DecoderMethod = "DecodeMVEVPNOT";
6828}
6829
6830let Predicates = [HasMVEInt] in {
6831  def : Pat<(v4i1 (xor (v4i1 VCCR:$pred), (v4i1 (predicate_cast (i32 65535))))),
6832            (v4i1 (MVE_VPNOT (v4i1 VCCR:$pred)))>;
6833  def : Pat<(v8i1 (xor (v8i1 VCCR:$pred), (v8i1 (predicate_cast (i32 65535))))),
6834            (v8i1 (MVE_VPNOT (v8i1 VCCR:$pred)))>;
6835  def : Pat<(v16i1 (xor (v16i1 VCCR:$pred), (v16i1 (predicate_cast (i32 65535))))),
6836            (v16i1 (MVE_VPNOT (v16i1 VCCR:$pred)))>;
6837}
6838
6839
6840class MVE_loltp_start<dag iops, string asm, string ops, bits<2> size>
6841  : t2LOL<(outs GPRlr:$LR), iops, asm, ops> {
6842  bits<4> Rn;
6843  let Predicates = [HasMVEInt];
6844  let Inst{22} = 0b0;
6845  let Inst{21-20} = size;
6846  let Inst{19-16} = Rn{3-0};
6847  let Inst{12} = 0b0;
6848}
6849
6850class MVE_DLSTP<string asm, bits<2> size>
6851  : MVE_loltp_start<(ins rGPR:$Rn), asm, "$LR, $Rn", size> {
6852  let Inst{13} = 0b1;
6853  let Inst{11-1} = 0b00000000000;
6854  let Unpredictable{10-1} = 0b1111111111;
6855}
6856
6857class MVE_WLSTP<string asm, bits<2> size>
6858  : MVE_loltp_start<(ins rGPR:$Rn, wlslabel_u11:$label),
6859                    asm, "$LR, $Rn, $label", size> {
6860  bits<11> label;
6861  let Inst{13} = 0b0;
6862  let Inst{11} = label{0};
6863  let Inst{10-1} = label{10-1};
6864  let isBranch = 1;
6865  let isTerminator = 1;
6866}
6867
6868def SDT_MVEMEMCPYLOOPNODE
6869    : SDTypeProfile<0, 3, [SDTCisPtrTy<0>, SDTCisPtrTy<1>, SDTCisVT<2, i32>]>;
6870def MVE_MEMCPYLOOPNODE : SDNode<"ARMISD::MEMCPYLOOP", SDT_MVEMEMCPYLOOPNODE,
6871                                [SDNPHasChain, SDNPMayStore, SDNPMayLoad]>;
6872
6873let usesCustomInserter = 1, hasNoSchedulingInfo = 1, Defs = [CPSR] in {
6874  def MVE_MEMCPYLOOPINST : PseudoInst<(outs),
6875        (ins rGPR:$dst, rGPR:$src, rGPR:$sz),
6876        NoItinerary,
6877        [(MVE_MEMCPYLOOPNODE rGPR:$dst, rGPR:$src, rGPR:$sz)]>;
6878}
6879
6880def SDT_MVEMEMSETLOOPNODE
6881    : SDTypeProfile<0, 3, [SDTCisPtrTy<0>, SDTCisVT<1, v16i8>, SDTCisVT<2, i32>]>;
6882def MVE_MEMSETLOOPNODE : SDNode<"ARMISD::MEMSETLOOP", SDT_MVEMEMSETLOOPNODE,
6883                                [SDNPHasChain, SDNPMayStore, SDNPMayLoad]>;
6884
6885let usesCustomInserter = 1, hasNoSchedulingInfo = 1, Defs = [CPSR] in {
6886  def MVE_MEMSETLOOPINST : PseudoInst<(outs),
6887        (ins rGPR:$dst, MQPR:$src, rGPR:$sz),
6888        NoItinerary,
6889        [(MVE_MEMSETLOOPNODE rGPR:$dst, MQPR:$src, rGPR:$sz)]>;
6890}
6891
6892def MVE_DLSTP_8  : MVE_DLSTP<"dlstp.8",  0b00>;
6893def MVE_DLSTP_16 : MVE_DLSTP<"dlstp.16", 0b01>;
6894def MVE_DLSTP_32 : MVE_DLSTP<"dlstp.32", 0b10>;
6895def MVE_DLSTP_64 : MVE_DLSTP<"dlstp.64", 0b11>;
6896
6897def MVE_WLSTP_8  : MVE_WLSTP<"wlstp.8",  0b00>;
6898def MVE_WLSTP_16 : MVE_WLSTP<"wlstp.16", 0b01>;
6899def MVE_WLSTP_32 : MVE_WLSTP<"wlstp.32", 0b10>;
6900def MVE_WLSTP_64 : MVE_WLSTP<"wlstp.64", 0b11>;
6901
6902class MVE_loltp_end<dag oops, dag iops, string asm, string ops>
6903  : t2LOL<oops, iops, asm, ops> {
6904  let Predicates = [HasMVEInt];
6905  let Inst{22-21} = 0b00;
6906  let Inst{19-16} = 0b1111;
6907  let Inst{12} = 0b0;
6908}
6909
6910def MVE_LETP : MVE_loltp_end<(outs GPRlr:$LRout),
6911                             (ins GPRlr:$LRin, lelabel_u11:$label),
6912                             "letp", "$LRin, $label"> {
6913  bits<11> label;
6914  let Inst{20} = 0b1;
6915  let Inst{13} = 0b0;
6916  let Inst{11} = label{0};
6917  let Inst{10-1} = label{10-1};
6918  let isBranch = 1;
6919  let isTerminator = 1;
6920}
6921
6922def MVE_LCTP : MVE_loltp_end<(outs), (ins pred:$p), "lctp${p}", ""> {
6923  let Inst{20} = 0b0;
6924  let Inst{13} = 0b1;
6925  let Inst{11-1} = 0b00000000000;
6926  let Unpredictable{21-20} = 0b11;
6927  let Unpredictable{11-1} = 0b11111111111;
6928}
6929
6930
6931//===----------------------------------------------------------------------===//
6932// Patterns
6933//===----------------------------------------------------------------------===//
6934
6935// PatFrags for loads and stores. Often trying to keep semi-consistent names.
6936
6937def aligned32_pre_store : PatFrag<(ops node:$val, node:$ptr, node:$offset),
6938                                  (pre_store node:$val, node:$ptr, node:$offset), [{
6939  return cast<StoreSDNode>(N)->getAlignment() >= 4;
6940}]>;
6941def aligned32_post_store : PatFrag<(ops node:$val, node:$ptr, node:$offset),
6942                                   (post_store node:$val, node:$ptr, node:$offset), [{
6943  return cast<StoreSDNode>(N)->getAlignment() >= 4;
6944}]>;
6945def aligned16_pre_store : PatFrag<(ops node:$val, node:$ptr, node:$offset),
6946                                  (pre_store node:$val, node:$ptr, node:$offset), [{
6947  return cast<StoreSDNode>(N)->getAlignment() >= 2;
6948}]>;
6949def aligned16_post_store : PatFrag<(ops node:$val, node:$ptr, node:$offset),
6950                                   (post_store node:$val, node:$ptr, node:$offset), [{
6951  return cast<StoreSDNode>(N)->getAlignment() >= 2;
6952}]>;
6953
6954
6955def aligned_maskedloadvi8 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
6956                                    (masked_ld node:$ptr, undef, node:$pred, node:$passthru), [{
6957  auto *Ld = cast<MaskedLoadSDNode>(N);
6958  return Ld->getMemoryVT().getScalarType() == MVT::i8;
6959}]>;
6960def aligned_sextmaskedloadvi8 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
6961                                        (aligned_maskedloadvi8 node:$ptr, node:$pred, node:$passthru), [{
6962  return cast<MaskedLoadSDNode>(N)->getExtensionType() == ISD::SEXTLOAD;
6963}]>;
6964def aligned_zextmaskedloadvi8 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
6965                                        (aligned_maskedloadvi8 node:$ptr, node:$pred, node:$passthru), [{
6966  return cast<MaskedLoadSDNode>(N)->getExtensionType() == ISD::ZEXTLOAD;
6967}]>;
6968def aligned_extmaskedloadvi8 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
6969                                       (aligned_maskedloadvi8 node:$ptr, node:$pred, node:$passthru), [{
6970  auto *Ld = cast<MaskedLoadSDNode>(N);
6971  EVT ScalarVT = Ld->getMemoryVT().getScalarType();
6972  return ScalarVT.isInteger() && Ld->getExtensionType() == ISD::EXTLOAD;
6973}]>;
6974def aligned_maskedloadvi16: PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
6975                                    (masked_ld node:$ptr, undef, node:$pred, node:$passthru), [{
6976  auto *Ld = cast<MaskedLoadSDNode>(N);
6977  EVT ScalarVT = Ld->getMemoryVT().getScalarType();
6978  return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && Ld->getAlignment() >= 2;
6979}]>;
6980def aligned_sextmaskedloadvi16 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
6981                                         (aligned_maskedloadvi16 node:$ptr, node:$pred, node:$passthru), [{
6982  return cast<MaskedLoadSDNode>(N)->getExtensionType() == ISD::SEXTLOAD;
6983}]>;
6984def aligned_zextmaskedloadvi16 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
6985                                         (aligned_maskedloadvi16 node:$ptr, node:$pred, node:$passthru), [{
6986  return cast<MaskedLoadSDNode>(N)->getExtensionType() == ISD::ZEXTLOAD;
6987}]>;
6988def aligned_extmaskedloadvi16 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
6989                                        (aligned_maskedloadvi16 node:$ptr, node:$pred, node:$passthru), [{
6990  auto *Ld = cast<MaskedLoadSDNode>(N);
6991  EVT ScalarVT = Ld->getMemoryVT().getScalarType();
6992  return ScalarVT.isInteger() && Ld->getExtensionType() == ISD::EXTLOAD;
6993}]>;
6994def aligned_maskedloadvi32: PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
6995                                    (masked_ld node:$ptr, undef, node:$pred, node:$passthru), [{
6996  auto *Ld = cast<MaskedLoadSDNode>(N);
6997  EVT ScalarVT = Ld->getMemoryVT().getScalarType();
6998  return (ScalarVT == MVT::i32 || ScalarVT == MVT::f32) && Ld->getAlignment() >= 4;
6999}]>;
7000
7001def aligned_maskedstvi8 : PatFrag<(ops node:$val, node:$ptr, node:$pred),
7002                                  (masked_st node:$val, node:$ptr, undef, node:$pred), [{
7003  return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
7004}]>;
7005def aligned_maskedstvi16 : PatFrag<(ops node:$val, node:$ptr, node:$pred),
7006                                   (masked_st node:$val, node:$ptr, undef, node:$pred), [{
7007  auto *St = cast<MaskedStoreSDNode>(N);
7008  EVT ScalarVT = St->getMemoryVT().getScalarType();
7009  return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlignment() >= 2;
7010}]>;
7011def aligned_maskedstvi32 : PatFrag<(ops node:$val, node:$ptr, node:$pred),
7012                                   (masked_st node:$val, node:$ptr, undef, node:$pred), [{
7013  auto *St = cast<MaskedStoreSDNode>(N);
7014  EVT ScalarVT = St->getMemoryVT().getScalarType();
7015  return (ScalarVT == MVT::i32 || ScalarVT == MVT::f32) && St->getAlignment() >= 4;
7016}]>;
7017
7018def pre_maskedstore : PatFrag<(ops node:$val, node:$base, node:$offset, node:$mask),
7019                              (masked_st node:$val, node:$base, node:$offset, node:$mask), [{
7020  ISD::MemIndexedMode AM = cast<MaskedStoreSDNode>(N)->getAddressingMode();
7021  return AM == ISD::PRE_INC || AM == ISD::PRE_DEC;
7022}]>;
7023def post_maskedstore : PatFrag<(ops node:$val, node:$base, node:$offset, node:$mask),
7024                               (masked_st node:$val, node:$base, node:$offset, node:$mask), [{
7025  ISD::MemIndexedMode AM = cast<MaskedStoreSDNode>(N)->getAddressingMode();
7026  return AM == ISD::POST_INC || AM == ISD::POST_DEC;
7027}]>;
7028def aligned_pre_maskedstorevi8 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
7029                                         (pre_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
7030  return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
7031}]>;
7032def aligned_post_maskedstorevi8 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
7033                                          (post_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
7034  return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
7035}]>;
7036def aligned_pre_maskedstorevi16 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
7037                                          (pre_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
7038  auto *St = cast<MaskedStoreSDNode>(N);
7039  EVT ScalarVT = St->getMemoryVT().getScalarType();
7040  return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlignment() >= 2;
7041}]>;
7042def aligned_post_maskedstorevi16 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
7043                                           (post_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
7044  auto *St = cast<MaskedStoreSDNode>(N);
7045  EVT ScalarVT = St->getMemoryVT().getScalarType();
7046  return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlignment() >= 2;
7047}]>;
7048def aligned_pre_maskedstorevi32 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
7049                                          (pre_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
7050  auto *St = cast<MaskedStoreSDNode>(N);
7051  EVT ScalarVT = St->getMemoryVT().getScalarType();
7052  return (ScalarVT == MVT::i32 || ScalarVT == MVT::f32) && St->getAlignment() >= 4;
7053}]>;
7054def aligned_post_maskedstorevi32 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
7055                                           (post_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
7056  auto *St = cast<MaskedStoreSDNode>(N);
7057  EVT ScalarVT = St->getMemoryVT().getScalarType();
7058  return (ScalarVT == MVT::i32 || ScalarVT == MVT::f32) && St->getAlignment() >= 4;
7059}]>;
7060
7061
7062// PatFrags for "Aligned" extending / truncating
7063
7064def aligned_extloadvi8  : PatFrag<(ops node:$ptr), (extloadvi8 node:$ptr)>;
7065def aligned_sextloadvi8 : PatFrag<(ops node:$ptr), (sextloadvi8 node:$ptr)>;
7066def aligned_zextloadvi8 : PatFrag<(ops node:$ptr), (zextloadvi8 node:$ptr)>;
7067
7068def aligned_truncstvi8 : PatFrag<(ops node:$val, node:$ptr),
7069                                 (truncstorevi8 node:$val, node:$ptr)>;
7070def aligned_post_truncstvi8 : PatFrag<(ops node:$val, node:$base, node:$offset),
7071                                      (post_truncstvi8 node:$val, node:$base, node:$offset)>;
7072def aligned_pre_truncstvi8 : PatFrag<(ops node:$val, node:$base, node:$offset),
7073                                     (pre_truncstvi8 node:$val, node:$base, node:$offset)>;
7074
7075let MinAlignment = 2 in {
7076  def aligned_extloadvi16  : PatFrag<(ops node:$ptr), (extloadvi16 node:$ptr)>;
7077  def aligned_sextloadvi16 : PatFrag<(ops node:$ptr), (sextloadvi16 node:$ptr)>;
7078  def aligned_zextloadvi16 : PatFrag<(ops node:$ptr), (zextloadvi16 node:$ptr)>;
7079
7080  def aligned_truncstvi16 : PatFrag<(ops node:$val, node:$ptr),
7081                                    (truncstorevi16 node:$val, node:$ptr)>;
7082  def aligned_post_truncstvi16 : PatFrag<(ops node:$val, node:$base, node:$offset),
7083                                         (post_truncstvi16 node:$val, node:$base, node:$offset)>;
7084  def aligned_pre_truncstvi16 : PatFrag<(ops node:$val, node:$base, node:$offset),
7085                                        (pre_truncstvi16 node:$val, node:$base, node:$offset)>;
7086}
7087
7088def truncmaskedst : PatFrag<(ops node:$val, node:$base, node:$pred),
7089                            (masked_st node:$val, node:$base, undef, node:$pred), [{
7090  return cast<MaskedStoreSDNode>(N)->isTruncatingStore();
7091}]>;
7092def aligned_truncmaskedstvi8 : PatFrag<(ops node:$val, node:$base, node:$pred),
7093                                       (truncmaskedst node:$val, node:$base, node:$pred), [{
7094  return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
7095}]>;
7096def aligned_truncmaskedstvi16 : PatFrag<(ops node:$val, node:$base, node:$pred),
7097                                        (truncmaskedst node:$val, node:$base, node:$pred), [{
7098  auto *St = cast<MaskedStoreSDNode>(N);
7099  EVT ScalarVT = St->getMemoryVT().getScalarType();
7100  return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlignment() >= 2;
7101}]>;
7102def pre_truncmaskedst : PatFrag<(ops node:$val, node:$base, node:$offset, node:$pred),
7103                                (masked_st node:$val, node:$base, node:$offset, node:$pred), [{
7104  ISD::MemIndexedMode AM = cast<MaskedStoreSDNode>(N)->getAddressingMode();
7105  return cast<MaskedStoreSDNode>(N)->isTruncatingStore() && (AM == ISD::PRE_INC || AM == ISD::PRE_DEC);
7106}]>;
7107def aligned_pre_truncmaskedstvi8 : PatFrag<(ops node:$val, node:$base, node:$offset, node:$pred),
7108                                           (pre_truncmaskedst node:$val, node:$base, node:$offset, node:$pred), [{
7109  return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
7110}]>;
7111def aligned_pre_truncmaskedstvi16 : PatFrag<(ops node:$val, node:$base, node:$offset, node:$pred),
7112                                            (pre_truncmaskedst node:$val, node:$base, node:$offset, node:$pred), [{
7113  auto *St = cast<MaskedStoreSDNode>(N);
7114  EVT ScalarVT = St->getMemoryVT().getScalarType();
7115  return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlignment() >= 2;
7116}]>;
7117def post_truncmaskedst : PatFrag<(ops node:$val, node:$base, node:$offset, node:$postd),
7118                                 (masked_st node:$val, node:$base, node:$offset, node:$postd), [{
7119  ISD::MemIndexedMode AM = cast<MaskedStoreSDNode>(N)->getAddressingMode();
7120  return cast<MaskedStoreSDNode>(N)->isTruncatingStore() && (AM == ISD::POST_INC || AM == ISD::POST_DEC);
7121}]>;
7122def aligned_post_truncmaskedstvi8 : PatFrag<(ops node:$val, node:$base, node:$offset, node:$postd),
7123                                            (post_truncmaskedst node:$val, node:$base, node:$offset, node:$postd), [{
7124  return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
7125}]>;
7126def aligned_post_truncmaskedstvi16 : PatFrag<(ops node:$val, node:$base, node:$offset, node:$postd),
7127                                             (post_truncmaskedst node:$val, node:$base, node:$offset, node:$postd), [{
7128  auto *St = cast<MaskedStoreSDNode>(N);
7129  EVT ScalarVT = St->getMemoryVT().getScalarType();
7130  return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlignment() >= 2;
7131}]>;
7132
7133// Load/store patterns
7134
7135class MVE_vector_store_typed<ValueType Ty, Instruction RegImmInst,
7136                             PatFrag StoreKind, int shift>
7137  : Pat<(StoreKind (Ty MQPR:$val), t2addrmode_imm7<shift>:$addr),
7138        (RegImmInst (Ty MQPR:$val), t2addrmode_imm7<shift>:$addr)>;
7139
7140class MVE_vector_maskedstore_typed<ValueType Ty, Instruction RegImmInst,
7141                                   PatFrag StoreKind, int shift>
7142  : Pat<(StoreKind (Ty MQPR:$val), t2addrmode_imm7<shift>:$addr, VCCR:$pred),
7143        (RegImmInst (Ty MQPR:$val), t2addrmode_imm7<shift>:$addr, ARMVCCThen, VCCR:$pred)>;
7144
7145multiclass MVE_vector_store<Instruction RegImmInst, PatFrag StoreKind,
7146                            int shift> {
7147  def : MVE_vector_store_typed<v16i8, RegImmInst, StoreKind, shift>;
7148  def : MVE_vector_store_typed<v8i16, RegImmInst, StoreKind, shift>;
7149  def : MVE_vector_store_typed<v8f16, RegImmInst, StoreKind, shift>;
7150  def : MVE_vector_store_typed<v4i32, RegImmInst, StoreKind, shift>;
7151  def : MVE_vector_store_typed<v4f32, RegImmInst, StoreKind, shift>;
7152  def : MVE_vector_store_typed<v2i64, RegImmInst, StoreKind, shift>;
7153  def : MVE_vector_store_typed<v2f64, RegImmInst, StoreKind, shift>;
7154}
7155
7156class MVE_vector_load_typed<ValueType Ty, Instruction RegImmInst,
7157                            PatFrag LoadKind, int shift>
7158  : Pat<(Ty (LoadKind t2addrmode_imm7<shift>:$addr)),
7159        (Ty (RegImmInst t2addrmode_imm7<shift>:$addr))>;
7160
7161class MVE_vector_maskedload_typed<ValueType Ty, Instruction RegImmInst,
7162                                  PatFrag LoadKind, int shift>
7163  : Pat<(Ty (LoadKind t2addrmode_imm7<shift>:$addr, VCCR:$pred, (Ty (ARMvmovImm (i32 0))))),
7164        (Ty (RegImmInst t2addrmode_imm7<shift>:$addr, ARMVCCThen, VCCR:$pred))>;
7165
7166multiclass MVE_vector_load<Instruction RegImmInst, PatFrag LoadKind,
7167                           int shift> {
7168  def : MVE_vector_load_typed<v16i8, RegImmInst, LoadKind, shift>;
7169  def : MVE_vector_load_typed<v8i16, RegImmInst, LoadKind, shift>;
7170  def : MVE_vector_load_typed<v8f16, RegImmInst, LoadKind, shift>;
7171  def : MVE_vector_load_typed<v4i32, RegImmInst, LoadKind, shift>;
7172  def : MVE_vector_load_typed<v4f32, RegImmInst, LoadKind, shift>;
7173  def : MVE_vector_load_typed<v2i64, RegImmInst, LoadKind, shift>;
7174  def : MVE_vector_load_typed<v2f64, RegImmInst, LoadKind, shift>;
7175}
7176
7177class MVE_vector_offset_store_typed<ValueType Ty, Instruction Opcode,
7178                                    PatFrag StoreKind, int shift>
7179  : Pat<(StoreKind (Ty MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<shift>:$addr),
7180        (Opcode MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<shift>:$addr)>;
7181
7182class MVE_vector_offset_maskedstore_typed<ValueType Ty, Instruction Opcode,
7183                                          PatFrag StoreKind, int shift>
7184  : Pat<(StoreKind (Ty MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<shift>:$addr, VCCR:$pred),
7185        (Opcode MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<shift>:$addr, ARMVCCThen, VCCR:$pred)>;
7186
7187multiclass MVE_vector_offset_store<Instruction RegImmInst, PatFrag StoreKind,
7188                                   int shift> {
7189  def : MVE_vector_offset_store_typed<v16i8, RegImmInst, StoreKind, shift>;
7190  def : MVE_vector_offset_store_typed<v8i16, RegImmInst, StoreKind, shift>;
7191  def : MVE_vector_offset_store_typed<v8f16, RegImmInst, StoreKind, shift>;
7192  def : MVE_vector_offset_store_typed<v4i32, RegImmInst, StoreKind, shift>;
7193  def : MVE_vector_offset_store_typed<v4f32, RegImmInst, StoreKind, shift>;
7194  def : MVE_vector_offset_store_typed<v2i64, RegImmInst, StoreKind, shift>;
7195  def : MVE_vector_offset_store_typed<v2f64, RegImmInst, StoreKind, shift>;
7196}
7197
7198
7199let Predicates = [HasMVEInt, IsLE] in {
7200  // Stores
7201  defm : MVE_vector_store<MVE_VSTRBU8, byte_alignedstore, 0>;
7202  defm : MVE_vector_store<MVE_VSTRHU16, hword_alignedstore, 1>;
7203  defm : MVE_vector_store<MVE_VSTRWU32, alignedstore32, 2>;
7204
7205  // Loads
7206  defm : MVE_vector_load<MVE_VLDRBU8, byte_alignedload, 0>;
7207  defm : MVE_vector_load<MVE_VLDRHU16, hword_alignedload, 1>;
7208  defm : MVE_vector_load<MVE_VLDRWU32, alignedload32, 2>;
7209
7210  // Pre/post inc stores
7211  defm : MVE_vector_offset_store<MVE_VSTRBU8_pre, pre_store, 0>;
7212  defm : MVE_vector_offset_store<MVE_VSTRBU8_post, post_store, 0>;
7213  defm : MVE_vector_offset_store<MVE_VSTRHU16_pre, aligned16_pre_store, 1>;
7214  defm : MVE_vector_offset_store<MVE_VSTRHU16_post, aligned16_post_store, 1>;
7215  defm : MVE_vector_offset_store<MVE_VSTRWU32_pre, aligned32_pre_store, 2>;
7216  defm : MVE_vector_offset_store<MVE_VSTRWU32_post, aligned32_post_store, 2>;
7217}
7218
7219let Predicates = [HasMVEInt, IsBE] in {
7220  // Aligned Stores
7221  def : MVE_vector_store_typed<v16i8, MVE_VSTRBU8, store, 0>;
7222  def : MVE_vector_store_typed<v8i16, MVE_VSTRHU16, alignedstore16, 1>;
7223  def : MVE_vector_store_typed<v8f16, MVE_VSTRHU16, alignedstore16, 1>;
7224  def : MVE_vector_store_typed<v4i32, MVE_VSTRWU32, alignedstore32, 2>;
7225  def : MVE_vector_store_typed<v4f32, MVE_VSTRWU32, alignedstore32, 2>;
7226
7227  // Aligned Loads
7228  def : MVE_vector_load_typed<v16i8, MVE_VLDRBU8, load, 0>;
7229  def : MVE_vector_load_typed<v8i16, MVE_VLDRHU16, alignedload16, 1>;
7230  def : MVE_vector_load_typed<v8f16, MVE_VLDRHU16, alignedload16, 1>;
7231  def : MVE_vector_load_typed<v4i32, MVE_VLDRWU32, alignedload32, 2>;
7232  def : MVE_vector_load_typed<v4f32, MVE_VLDRWU32, alignedload32, 2>;
7233
7234  // Other unaligned loads/stores need to go though a VREV
7235  def : Pat<(v2f64 (load t2addrmode_imm7<0>:$addr)),
7236            (v2f64 (MVE_VREV64_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
7237  def : Pat<(v2i64 (load t2addrmode_imm7<0>:$addr)),
7238            (v2i64 (MVE_VREV64_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
7239  def : Pat<(v4i32 (load t2addrmode_imm7<0>:$addr)),
7240            (v4i32 (MVE_VREV32_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
7241  def : Pat<(v4f32 (load t2addrmode_imm7<0>:$addr)),
7242            (v4f32 (MVE_VREV32_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
7243  def : Pat<(v8i16 (load t2addrmode_imm7<0>:$addr)),
7244            (v8i16 (MVE_VREV16_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
7245  def : Pat<(v8f16 (load t2addrmode_imm7<0>:$addr)),
7246            (v8f16 (MVE_VREV16_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
7247  def : Pat<(store (v2f64 MQPR:$val), t2addrmode_imm7<0>:$addr),
7248            (MVE_VSTRBU8 (MVE_VREV64_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
7249  def : Pat<(store (v2i64 MQPR:$val), t2addrmode_imm7<0>:$addr),
7250            (MVE_VSTRBU8 (MVE_VREV64_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
7251  def : Pat<(store (v4i32 MQPR:$val), t2addrmode_imm7<0>:$addr),
7252            (MVE_VSTRBU8 (MVE_VREV32_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
7253  def : Pat<(store (v4f32 MQPR:$val), t2addrmode_imm7<0>:$addr),
7254            (MVE_VSTRBU8 (MVE_VREV32_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
7255  def : Pat<(store (v8i16 MQPR:$val), t2addrmode_imm7<0>:$addr),
7256            (MVE_VSTRBU8 (MVE_VREV16_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
7257  def : Pat<(store (v8f16 MQPR:$val), t2addrmode_imm7<0>:$addr),
7258            (MVE_VSTRBU8 (MVE_VREV16_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
7259
7260  // Pre/Post inc stores
7261  def : MVE_vector_offset_store_typed<v16i8, MVE_VSTRBU8_pre, pre_store, 0>;
7262  def : MVE_vector_offset_store_typed<v16i8, MVE_VSTRBU8_post, post_store, 0>;
7263  def : MVE_vector_offset_store_typed<v8i16, MVE_VSTRHU16_pre, aligned16_pre_store, 1>;
7264  def : MVE_vector_offset_store_typed<v8i16, MVE_VSTRHU16_post, aligned16_post_store, 1>;
7265  def : MVE_vector_offset_store_typed<v8f16, MVE_VSTRHU16_pre, aligned16_pre_store, 1>;
7266  def : MVE_vector_offset_store_typed<v8f16, MVE_VSTRHU16_post, aligned16_post_store, 1>;
7267  def : MVE_vector_offset_store_typed<v4i32, MVE_VSTRWU32_pre, aligned32_pre_store, 2>;
7268  def : MVE_vector_offset_store_typed<v4i32, MVE_VSTRWU32_post, aligned32_post_store, 2>;
7269  def : MVE_vector_offset_store_typed<v4f32, MVE_VSTRWU32_pre, aligned32_pre_store, 2>;
7270  def : MVE_vector_offset_store_typed<v4f32, MVE_VSTRWU32_post, aligned32_post_store, 2>;
7271}
7272
7273let Predicates = [HasMVEInt] in {
7274  // Aligned masked store, shared between LE and BE
7275  def : MVE_vector_maskedstore_typed<v16i8, MVE_VSTRBU8, aligned_maskedstvi8, 0>;
7276  def : MVE_vector_maskedstore_typed<v8i16, MVE_VSTRHU16, aligned_maskedstvi16, 1>;
7277  def : MVE_vector_maskedstore_typed<v8f16, MVE_VSTRHU16, aligned_maskedstvi16, 1>;
7278  def : MVE_vector_maskedstore_typed<v4i32, MVE_VSTRWU32, aligned_maskedstvi32, 2>;
7279  def : MVE_vector_maskedstore_typed<v4f32, MVE_VSTRWU32, aligned_maskedstvi32, 2>;
7280
7281  // Pre/Post inc masked stores
7282  def : MVE_vector_offset_maskedstore_typed<v16i8, MVE_VSTRBU8_pre, aligned_pre_maskedstorevi8, 0>;
7283  def : MVE_vector_offset_maskedstore_typed<v16i8, MVE_VSTRBU8_post, aligned_post_maskedstorevi8, 0>;
7284  def : MVE_vector_offset_maskedstore_typed<v8i16, MVE_VSTRHU16_pre, aligned_pre_maskedstorevi16, 1>;
7285  def : MVE_vector_offset_maskedstore_typed<v8i16, MVE_VSTRHU16_post, aligned_post_maskedstorevi16, 1>;
7286  def : MVE_vector_offset_maskedstore_typed<v8f16, MVE_VSTRHU16_pre, aligned_pre_maskedstorevi16, 1>;
7287  def : MVE_vector_offset_maskedstore_typed<v8f16, MVE_VSTRHU16_post, aligned_post_maskedstorevi16, 1>;
7288  def : MVE_vector_offset_maskedstore_typed<v4i32, MVE_VSTRWU32_pre, aligned_pre_maskedstorevi32, 2>;
7289  def : MVE_vector_offset_maskedstore_typed<v4i32, MVE_VSTRWU32_post, aligned_post_maskedstorevi32, 2>;
7290  def : MVE_vector_offset_maskedstore_typed<v4f32, MVE_VSTRWU32_pre, aligned_pre_maskedstorevi32, 2>;
7291  def : MVE_vector_offset_maskedstore_typed<v4f32, MVE_VSTRWU32_post, aligned_post_maskedstorevi32, 2>;
7292
7293  // Aligned masked loads
7294  def : MVE_vector_maskedload_typed<v16i8, MVE_VLDRBU8, aligned_maskedloadvi8, 0>;
7295  def : MVE_vector_maskedload_typed<v8i16, MVE_VLDRHU16, aligned_maskedloadvi16, 1>;
7296  def : MVE_vector_maskedload_typed<v8f16, MVE_VLDRHU16, aligned_maskedloadvi16, 1>;
7297  def : MVE_vector_maskedload_typed<v4i32, MVE_VLDRWU32, aligned_maskedloadvi32, 2>;
7298  def : MVE_vector_maskedload_typed<v4f32, MVE_VLDRWU32, aligned_maskedloadvi32, 2>;
7299}
7300
7301// Widening/Narrowing Loads/Stores
7302
7303multiclass MVEExtLoadStore<Instruction LoadSInst, Instruction LoadUInst, string StoreInst,
7304                         string Amble, ValueType VT, int Shift> {
7305  // Trunc stores
7306  def : Pat<(!cast<PatFrag>("aligned_truncst"#Amble) (VT MQPR:$val), taddrmode_imm7<Shift>:$addr),
7307            (!cast<Instruction>(StoreInst) MQPR:$val, taddrmode_imm7<Shift>:$addr)>;
7308  def : Pat<(!cast<PatFrag>("aligned_post_truncst"#Amble) (VT MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<Shift>:$addr),
7309            (!cast<Instruction>(StoreInst#"_post") MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<Shift>:$addr)>;
7310  def : Pat<(!cast<PatFrag>("aligned_pre_truncst"#Amble) (VT MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<Shift>:$addr),
7311            (!cast<Instruction>(StoreInst#"_pre") MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<Shift>:$addr)>;
7312
7313  // Masked trunc stores
7314  def : Pat<(!cast<PatFrag>("aligned_truncmaskedst"#Amble) (VT MQPR:$val), taddrmode_imm7<Shift>:$addr, VCCR:$pred),
7315            (!cast<Instruction>(StoreInst) MQPR:$val, taddrmode_imm7<Shift>:$addr, ARMVCCThen, VCCR:$pred)>;
7316  def : Pat<(!cast<PatFrag>("aligned_post_truncmaskedst"#Amble) (VT MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<Shift>:$addr, VCCR:$pred),
7317            (!cast<Instruction>(StoreInst#"_post") MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<Shift>:$addr, ARMVCCThen, VCCR:$pred)>;
7318  def : Pat<(!cast<PatFrag>("aligned_pre_truncmaskedst"#Amble) (VT MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<Shift>:$addr, VCCR:$pred),
7319            (!cast<Instruction>(StoreInst#"_pre") MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<Shift>:$addr, ARMVCCThen, VCCR:$pred)>;
7320
7321  // Ext loads
7322  def : Pat<(VT (!cast<PatFrag>("aligned_extload"#Amble) taddrmode_imm7<Shift>:$addr)),
7323            (VT (LoadUInst taddrmode_imm7<Shift>:$addr))>;
7324  def : Pat<(VT (!cast<PatFrag>("aligned_sextload"#Amble) taddrmode_imm7<Shift>:$addr)),
7325            (VT (LoadSInst taddrmode_imm7<Shift>:$addr))>;
7326  def : Pat<(VT (!cast<PatFrag>("aligned_zextload"#Amble) taddrmode_imm7<Shift>:$addr)),
7327            (VT (LoadUInst taddrmode_imm7<Shift>:$addr))>;
7328
7329  // Masked ext loads
7330  def : Pat<(VT (!cast<PatFrag>("aligned_extmaskedload"#Amble) taddrmode_imm7<Shift>:$addr, VCCR:$pred, (VT (ARMvmovImm (i32 0))))),
7331            (VT (LoadUInst taddrmode_imm7<Shift>:$addr, ARMVCCThen, VCCR:$pred))>;
7332  def : Pat<(VT (!cast<PatFrag>("aligned_sextmaskedload"#Amble) taddrmode_imm7<Shift>:$addr, VCCR:$pred, (VT (ARMvmovImm (i32 0))))),
7333            (VT (LoadSInst taddrmode_imm7<Shift>:$addr, ARMVCCThen, VCCR:$pred))>;
7334  def : Pat<(VT (!cast<PatFrag>("aligned_zextmaskedload"#Amble) taddrmode_imm7<Shift>:$addr, VCCR:$pred, (VT (ARMvmovImm (i32 0))))),
7335            (VT (LoadUInst taddrmode_imm7<Shift>:$addr, ARMVCCThen, VCCR:$pred))>;
7336}
7337
7338let Predicates = [HasMVEInt] in {
7339  defm : MVEExtLoadStore<MVE_VLDRBS16, MVE_VLDRBU16, "MVE_VSTRB16", "vi8", v8i16, 0>;
7340  defm : MVEExtLoadStore<MVE_VLDRBS32, MVE_VLDRBU32, "MVE_VSTRB32", "vi8", v4i32, 0>;
7341  defm : MVEExtLoadStore<MVE_VLDRHS32, MVE_VLDRHU32, "MVE_VSTRH32", "vi16", v4i32, 1>;
7342}
7343
7344
7345// Bit convert patterns
7346
7347let Predicates = [HasMVEInt] in {
7348  def : Pat<(v2f64 (bitconvert (v2i64 MQPR:$src))), (v2f64 MQPR:$src)>;
7349  def : Pat<(v2i64 (bitconvert (v2f64 MQPR:$src))), (v2i64 MQPR:$src)>;
7350
7351  def : Pat<(v4i32 (bitconvert (v4f32 MQPR:$src))), (v4i32 MQPR:$src)>;
7352  def : Pat<(v4f32 (bitconvert (v4i32 MQPR:$src))), (v4f32 MQPR:$src)>;
7353
7354  def : Pat<(v8i16 (bitconvert (v8f16 MQPR:$src))), (v8i16  MQPR:$src)>;
7355  def : Pat<(v8f16 (bitconvert (v8i16 MQPR:$src))), (v8f16  MQPR:$src)>;
7356}
7357
7358let Predicates = [IsLE,HasMVEInt] in {
7359  def : Pat<(v2f64 (bitconvert (v4f32 MQPR:$src))), (v2f64 MQPR:$src)>;
7360  def : Pat<(v2f64 (bitconvert (v4i32 MQPR:$src))), (v2f64 MQPR:$src)>;
7361  def : Pat<(v2f64 (bitconvert (v8f16 MQPR:$src))), (v2f64 MQPR:$src)>;
7362  def : Pat<(v2f64 (bitconvert (v8i16 MQPR:$src))), (v2f64 MQPR:$src)>;
7363  def : Pat<(v2f64 (bitconvert (v16i8 MQPR:$src))), (v2f64 MQPR:$src)>;
7364
7365  def : Pat<(v2i64 (bitconvert (v4f32 MQPR:$src))), (v2i64 MQPR:$src)>;
7366  def : Pat<(v2i64 (bitconvert (v4i32 MQPR:$src))), (v2i64 MQPR:$src)>;
7367  def : Pat<(v2i64 (bitconvert (v8f16 MQPR:$src))), (v2i64 MQPR:$src)>;
7368  def : Pat<(v2i64 (bitconvert (v8i16 MQPR:$src))), (v2i64 MQPR:$src)>;
7369  def : Pat<(v2i64 (bitconvert (v16i8 MQPR:$src))), (v2i64 MQPR:$src)>;
7370
7371  def : Pat<(v4f32 (bitconvert (v2f64 MQPR:$src))), (v4f32 MQPR:$src)>;
7372  def : Pat<(v4f32 (bitconvert (v2i64 MQPR:$src))), (v4f32 MQPR:$src)>;
7373  def : Pat<(v4f32 (bitconvert (v8f16 MQPR:$src))), (v4f32 MQPR:$src)>;
7374  def : Pat<(v4f32 (bitconvert (v8i16 MQPR:$src))), (v4f32 MQPR:$src)>;
7375  def : Pat<(v4f32 (bitconvert (v16i8 MQPR:$src))), (v4f32 MQPR:$src)>;
7376
7377  def : Pat<(v4i32 (bitconvert (v2f64 MQPR:$src))), (v4i32 MQPR:$src)>;
7378  def : Pat<(v4i32 (bitconvert (v2i64 MQPR:$src))), (v4i32 MQPR:$src)>;
7379  def : Pat<(v4i32 (bitconvert (v8f16 MQPR:$src))), (v4i32 MQPR:$src)>;
7380  def : Pat<(v4i32 (bitconvert (v8i16 MQPR:$src))), (v4i32 MQPR:$src)>;
7381  def : Pat<(v4i32 (bitconvert (v16i8 MQPR:$src))), (v4i32 MQPR:$src)>;
7382
7383  def : Pat<(v8f16 (bitconvert (v2f64 MQPR:$src))), (v8f16 MQPR:$src)>;
7384  def : Pat<(v8f16 (bitconvert (v2i64 MQPR:$src))), (v8f16 MQPR:$src)>;
7385  def : Pat<(v8f16 (bitconvert (v4f32 MQPR:$src))), (v8f16 MQPR:$src)>;
7386  def : Pat<(v8f16 (bitconvert (v4i32 MQPR:$src))), (v8f16 MQPR:$src)>;
7387  def : Pat<(v8f16 (bitconvert (v16i8 MQPR:$src))), (v8f16 MQPR:$src)>;
7388
7389  def : Pat<(v8i16 (bitconvert (v2f64 MQPR:$src))), (v8i16 MQPR:$src)>;
7390  def : Pat<(v8i16 (bitconvert (v2i64 MQPR:$src))), (v8i16 MQPR:$src)>;
7391  def : Pat<(v8i16 (bitconvert (v4f32 MQPR:$src))), (v8i16 MQPR:$src)>;
7392  def : Pat<(v8i16 (bitconvert (v4i32 MQPR:$src))), (v8i16 MQPR:$src)>;
7393  def : Pat<(v8i16 (bitconvert (v16i8 MQPR:$src))), (v8i16 MQPR:$src)>;
7394
7395  def : Pat<(v16i8 (bitconvert (v2f64 MQPR:$src))), (v16i8 MQPR:$src)>;
7396  def : Pat<(v16i8 (bitconvert (v2i64 MQPR:$src))), (v16i8 MQPR:$src)>;
7397  def : Pat<(v16i8 (bitconvert (v4f32 MQPR:$src))), (v16i8 MQPR:$src)>;
7398  def : Pat<(v16i8 (bitconvert (v4i32 MQPR:$src))), (v16i8 MQPR:$src)>;
7399  def : Pat<(v16i8 (bitconvert (v8f16 MQPR:$src))), (v16i8 MQPR:$src)>;
7400  def : Pat<(v16i8 (bitconvert (v8i16 MQPR:$src))), (v16i8 MQPR:$src)>;
7401}
7402
7403let Predicates = [IsBE,HasMVEInt] in {
7404  def : Pat<(v2f64 (bitconvert (v4f32 MQPR:$src))), (v2f64 (MVE_VREV64_32 MQPR:$src))>;
7405  def : Pat<(v2f64 (bitconvert (v4i32 MQPR:$src))), (v2f64 (MVE_VREV64_32 MQPR:$src))>;
7406  def : Pat<(v2f64 (bitconvert (v8f16 MQPR:$src))), (v2f64 (MVE_VREV64_16 MQPR:$src))>;
7407  def : Pat<(v2f64 (bitconvert (v8i16 MQPR:$src))), (v2f64 (MVE_VREV64_16 MQPR:$src))>;
7408  def : Pat<(v2f64 (bitconvert (v16i8 MQPR:$src))), (v2f64 (MVE_VREV64_8 MQPR:$src))>;
7409
7410  def : Pat<(v2i64 (bitconvert (v4f32 MQPR:$src))), (v2i64 (MVE_VREV64_32 MQPR:$src))>;
7411  def : Pat<(v2i64 (bitconvert (v4i32 MQPR:$src))), (v2i64 (MVE_VREV64_32 MQPR:$src))>;
7412  def : Pat<(v2i64 (bitconvert (v8f16 MQPR:$src))), (v2i64 (MVE_VREV64_16 MQPR:$src))>;
7413  def : Pat<(v2i64 (bitconvert (v8i16 MQPR:$src))), (v2i64 (MVE_VREV64_16 MQPR:$src))>;
7414  def : Pat<(v2i64 (bitconvert (v16i8 MQPR:$src))), (v2i64 (MVE_VREV64_8 MQPR:$src))>;
7415
7416  def : Pat<(v4f32 (bitconvert (v2f64 MQPR:$src))), (v4f32 (MVE_VREV64_32 MQPR:$src))>;
7417  def : Pat<(v4f32 (bitconvert (v2i64 MQPR:$src))), (v4f32 (MVE_VREV64_32 MQPR:$src))>;
7418  def : Pat<(v4f32 (bitconvert (v8f16 MQPR:$src))), (v4f32 (MVE_VREV32_16 MQPR:$src))>;
7419  def : Pat<(v4f32 (bitconvert (v8i16 MQPR:$src))), (v4f32 (MVE_VREV32_16 MQPR:$src))>;
7420  def : Pat<(v4f32 (bitconvert (v16i8 MQPR:$src))), (v4f32 (MVE_VREV32_8 MQPR:$src))>;
7421
7422  def : Pat<(v4i32 (bitconvert (v2f64 MQPR:$src))), (v4i32 (MVE_VREV64_32 MQPR:$src))>;
7423  def : Pat<(v4i32 (bitconvert (v2i64 MQPR:$src))), (v4i32 (MVE_VREV64_32 MQPR:$src))>;
7424  def : Pat<(v4i32 (bitconvert (v8f16 MQPR:$src))), (v4i32 (MVE_VREV32_16 MQPR:$src))>;
7425  def : Pat<(v4i32 (bitconvert (v8i16 MQPR:$src))), (v4i32 (MVE_VREV32_16 MQPR:$src))>;
7426  def : Pat<(v4i32 (bitconvert (v16i8 MQPR:$src))), (v4i32 (MVE_VREV32_8 MQPR:$src))>;
7427
7428  def : Pat<(v8f16 (bitconvert (v2f64 MQPR:$src))), (v8f16 (MVE_VREV64_16 MQPR:$src))>;
7429  def : Pat<(v8f16 (bitconvert (v2i64 MQPR:$src))), (v8f16 (MVE_VREV64_16 MQPR:$src))>;
7430  def : Pat<(v8f16 (bitconvert (v4f32 MQPR:$src))), (v8f16 (MVE_VREV32_16 MQPR:$src))>;
7431  def : Pat<(v8f16 (bitconvert (v4i32 MQPR:$src))), (v8f16 (MVE_VREV32_16 MQPR:$src))>;
7432  def : Pat<(v8f16 (bitconvert (v16i8 MQPR:$src))), (v8f16 (MVE_VREV16_8 MQPR:$src))>;
7433
7434  def : Pat<(v8i16 (bitconvert (v2f64 MQPR:$src))), (v8i16 (MVE_VREV64_16 MQPR:$src))>;
7435  def : Pat<(v8i16 (bitconvert (v2i64 MQPR:$src))), (v8i16 (MVE_VREV64_16 MQPR:$src))>;
7436  def : Pat<(v8i16 (bitconvert (v4f32 MQPR:$src))), (v8i16 (MVE_VREV32_16 MQPR:$src))>;
7437  def : Pat<(v8i16 (bitconvert (v4i32 MQPR:$src))), (v8i16 (MVE_VREV32_16 MQPR:$src))>;
7438  def : Pat<(v8i16 (bitconvert (v16i8 MQPR:$src))), (v8i16 (MVE_VREV16_8 MQPR:$src))>;
7439
7440  def : Pat<(v16i8 (bitconvert (v2f64 MQPR:$src))), (v16i8 (MVE_VREV64_8 MQPR:$src))>;
7441  def : Pat<(v16i8 (bitconvert (v2i64 MQPR:$src))), (v16i8 (MVE_VREV64_8 MQPR:$src))>;
7442  def : Pat<(v16i8 (bitconvert (v4f32 MQPR:$src))), (v16i8 (MVE_VREV32_8 MQPR:$src))>;
7443  def : Pat<(v16i8 (bitconvert (v4i32 MQPR:$src))), (v16i8 (MVE_VREV32_8 MQPR:$src))>;
7444  def : Pat<(v16i8 (bitconvert (v8f16 MQPR:$src))), (v16i8 (MVE_VREV16_8 MQPR:$src))>;
7445  def : Pat<(v16i8 (bitconvert (v8i16 MQPR:$src))), (v16i8 (MVE_VREV16_8 MQPR:$src))>;
7446}
7447