xref: /llvm-project/llvm/lib/Target/ARM/ARMInstrMVE.td (revision 6b2232606d01a029f640b61b4f985d9dea79d4b6)
1//===-- ARMInstrMVE.td - MVE support for ARM ---------------*- tablegen -*-===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This file describes the ARM MVE instruction set.
10//
11//===----------------------------------------------------------------------===//
12
13// VPT condition mask
14def vpt_mask : Operand<i32> {
15  let PrintMethod = "printVPTMask";
16  let ParserMatchClass = it_mask_asmoperand;
17  let EncoderMethod = "getVPTMaskOpValue";
18  let DecoderMethod = "DecodeVPTMaskOperand";
19}
20
21// VPT/VCMP restricted predicate for sign invariant types
22def pred_restricted_i_asmoperand : AsmOperandClass {
23  let Name = "CondCodeRestrictedI";
24  let RenderMethod = "addITCondCodeOperands";
25  let PredicateMethod = "isITCondCodeRestrictedI";
26  let ParserMethod = "parseITCondCode";
27  let DiagnosticString = "condition code for sign-independent integer "#
28                         "comparison must be EQ or NE";
29}
30
31// VPT/VCMP restricted predicate for signed types
32def pred_restricted_s_asmoperand : AsmOperandClass {
33  let Name = "CondCodeRestrictedS";
34  let RenderMethod = "addITCondCodeOperands";
35  let PredicateMethod = "isITCondCodeRestrictedS";
36  let ParserMethod = "parseITCondCode";
37  let DiagnosticString = "condition code for signed integer "#
38                         "comparison must be EQ, NE, LT, GT, LE or GE";
39}
40
41// VPT/VCMP restricted predicate for unsigned types
42def pred_restricted_u_asmoperand : AsmOperandClass {
43  let Name = "CondCodeRestrictedU";
44  let RenderMethod = "addITCondCodeOperands";
45  let PredicateMethod = "isITCondCodeRestrictedU";
46  let ParserMethod = "parseITCondCode";
47  let DiagnosticString = "condition code for unsigned integer "#
48                         "comparison must be EQ, NE, HS or HI";
49}
50
51// VPT/VCMP restricted predicate for floating point
52def pred_restricted_fp_asmoperand : AsmOperandClass {
53  let Name = "CondCodeRestrictedFP";
54  let RenderMethod = "addITCondCodeOperands";
55  let PredicateMethod = "isITCondCodeRestrictedFP";
56  let ParserMethod = "parseITCondCode";
57  let DiagnosticString = "condition code for floating-point "#
58                         "comparison must be EQ, NE, LT, GT, LE or GE";
59}
60
61class VCMPPredicateOperand : Operand<i32>;
62
63def pred_basic_i : VCMPPredicateOperand {
64  let PrintMethod = "printMandatoryRestrictedPredicateOperand";
65  let ParserMatchClass = pred_restricted_i_asmoperand;
66  let DecoderMethod = "DecodeRestrictedIPredicateOperand";
67  let EncoderMethod = "getRestrictedCondCodeOpValue";
68}
69
70def pred_basic_u : VCMPPredicateOperand {
71  let PrintMethod = "printMandatoryRestrictedPredicateOperand";
72  let ParserMatchClass = pred_restricted_u_asmoperand;
73  let DecoderMethod = "DecodeRestrictedUPredicateOperand";
74  let EncoderMethod = "getRestrictedCondCodeOpValue";
75}
76
77def pred_basic_s : VCMPPredicateOperand {
78  let PrintMethod = "printMandatoryRestrictedPredicateOperand";
79  let ParserMatchClass = pred_restricted_s_asmoperand;
80  let DecoderMethod = "DecodeRestrictedSPredicateOperand";
81  let EncoderMethod = "getRestrictedCondCodeOpValue";
82}
83
84def pred_basic_fp : VCMPPredicateOperand {
85  let PrintMethod = "printMandatoryRestrictedPredicateOperand";
86  let ParserMatchClass = pred_restricted_fp_asmoperand;
87  let DecoderMethod = "DecodeRestrictedFPPredicateOperand";
88  let EncoderMethod = "getRestrictedCondCodeOpValue";
89}
90
91// Register list operands for interleaving load/stores
92def VecList2QAsmOperand : AsmOperandClass {
93  let Name = "VecListTwoMQ";
94  let ParserMethod = "parseVectorList";
95  let RenderMethod = "addMVEVecListOperands";
96  let DiagnosticString = "operand must be a list of two consecutive "#
97                         "q-registers in range [q0,q7]";
98}
99
100def VecList2Q : RegisterOperand<MQQPR, "printMVEVectorListTwoQ"> {
101  let ParserMatchClass = VecList2QAsmOperand;
102  let PrintMethod = "printMVEVectorList<2>";
103}
104
105def VecList4QAsmOperand : AsmOperandClass {
106  let Name = "VecListFourMQ";
107  let ParserMethod = "parseVectorList";
108  let RenderMethod = "addMVEVecListOperands";
109  let DiagnosticString = "operand must be a list of four consecutive "#
110                         "q-registers in range [q0,q7]";
111}
112
113def VecList4Q : RegisterOperand<MQQQQPR, "printMVEVectorListFourQ"> {
114  let ParserMatchClass = VecList4QAsmOperand;
115  let PrintMethod = "printMVEVectorList<4>";
116}
117
118// taddrmode_imm7  := reg[r0-r7] +/- (imm7 << shift)
119class TMemImm7ShiftOffsetAsmOperand<int shift> : AsmOperandClass {
120  let Name = "TMemImm7Shift"#shift#"Offset";
121  let PredicateMethod = "isMemImm7ShiftedOffset<"#shift#",ARM::tGPRRegClassID>";
122  let RenderMethod = "addMemImmOffsetOperands";
123}
124
125class taddrmode_imm7<int shift> : MemOperand,
126    ComplexPattern<i32, 2, "SelectTAddrModeImm7<"#shift#">", []>  {
127  let ParserMatchClass = TMemImm7ShiftOffsetAsmOperand<shift>;
128  // They are printed the same way as the T2 imm8 version
129  let PrintMethod = "printT2AddrModeImm8Operand<false>";
130  // This can also be the same as the T2 version.
131  let EncoderMethod = "getT2AddrModeImmOpValue<7,"#shift#">";
132  let DecoderMethod = "DecodeTAddrModeImm7<"#shift#">";
133  let MIOperandInfo = (ops tGPR:$base, i32imm:$offsimm);
134}
135
136// t2addrmode_imm7  := reg +/- (imm7)
137class MemImm7ShiftOffsetAsmOperand<int shift> : AsmOperandClass {
138  let Name = "MemImm7Shift"#shift#"Offset";
139  let PredicateMethod = "isMemImm7ShiftedOffset<" # shift #
140                        ",ARM::GPRnopcRegClassID>";
141  let RenderMethod = "addMemImmOffsetOperands";
142}
143
144def MemImm7Shift0OffsetAsmOperand : MemImm7ShiftOffsetAsmOperand<0>;
145def MemImm7Shift1OffsetAsmOperand : MemImm7ShiftOffsetAsmOperand<1>;
146def MemImm7Shift2OffsetAsmOperand : MemImm7ShiftOffsetAsmOperand<2>;
147class T2AddrMode_Imm7<int shift> : MemOperand,
148      ComplexPattern<i32, 2, "SelectT2AddrModeImm7<"#shift#">", []> {
149  let EncoderMethod = "getT2AddrModeImmOpValue<7,"#shift#">";
150  let DecoderMethod = "DecodeT2AddrModeImm7<"#shift#", 0>";
151  let ParserMatchClass =
152    !cast<AsmOperandClass>("MemImm7Shift"#shift#"OffsetAsmOperand");
153  let MIOperandInfo = (ops GPRnopc:$base, i32imm:$offsimm);
154}
155
156class t2addrmode_imm7<int shift> : T2AddrMode_Imm7<shift> {
157  // They are printed the same way as the imm8 version
158  let PrintMethod = "printT2AddrModeImm8Operand<false>";
159}
160
161class MemImm7ShiftOffsetWBAsmOperand<int shift> : AsmOperandClass {
162  let Name = "MemImm7Shift"#shift#"OffsetWB";
163  let PredicateMethod = "isMemImm7ShiftedOffset<" # shift #
164                        ",ARM::rGPRRegClassID>";
165  let RenderMethod = "addMemImmOffsetOperands";
166}
167
168def MemImm7Shift0OffsetWBAsmOperand : MemImm7ShiftOffsetWBAsmOperand<0>;
169def MemImm7Shift1OffsetWBAsmOperand : MemImm7ShiftOffsetWBAsmOperand<1>;
170def MemImm7Shift2OffsetWBAsmOperand : MemImm7ShiftOffsetWBAsmOperand<2>;
171
172class t2addrmode_imm7_pre<int shift> : T2AddrMode_Imm7<shift> {
173  // They are printed the same way as the imm8 version
174  let PrintMethod = "printT2AddrModeImm8Operand<true>";
175  let ParserMatchClass =
176    !cast<AsmOperandClass>("MemImm7Shift"#shift#"OffsetWBAsmOperand");
177  let DecoderMethod = "DecodeT2AddrModeImm7<"#shift#", 1>";
178  let MIOperandInfo = (ops rGPR:$base, i32imm:$offsim);
179}
180
181class t2am_imm7shiftOffsetAsmOperand<int shift>
182  : AsmOperandClass { let Name = "Imm7Shift"#shift; }
183def t2am_imm7shift0OffsetAsmOperand : t2am_imm7shiftOffsetAsmOperand<0>;
184def t2am_imm7shift1OffsetAsmOperand : t2am_imm7shiftOffsetAsmOperand<1>;
185def t2am_imm7shift2OffsetAsmOperand : t2am_imm7shiftOffsetAsmOperand<2>;
186
187class t2am_imm7_offset<int shift>
188    : MemOperand,
189      ComplexPattern<i32, 1, "SelectT2AddrModeImm7Offset<"#shift#">"> {
190  // They are printed the same way as the imm8 version
191  let PrintMethod = "printT2AddrModeImm8OffsetOperand";
192  let ParserMatchClass =
193    !cast<AsmOperandClass>("t2am_imm7shift"#shift#"OffsetAsmOperand");
194  let EncoderMethod = "getT2ScaledImmOpValue<7,"#shift#">";
195  let DecoderMethod = "DecodeT2Imm7<"#shift#">";
196  let WantsRoot = true;
197}
198
199// Operands for gather/scatter loads of the form [Rbase, Qoffsets]
200class MemRegRQOffsetAsmOperand<int shift> : AsmOperandClass {
201  let Name = "MemRegRQS"#shift#"Offset";
202  let PredicateMethod = "isMemRegRQOffset<"#shift#">";
203  let RenderMethod = "addMemRegRQOffsetOperands";
204}
205
206def MemRegRQS0OffsetAsmOperand : MemRegRQOffsetAsmOperand<0>;
207def MemRegRQS1OffsetAsmOperand : MemRegRQOffsetAsmOperand<1>;
208def MemRegRQS2OffsetAsmOperand : MemRegRQOffsetAsmOperand<2>;
209def MemRegRQS3OffsetAsmOperand : MemRegRQOffsetAsmOperand<3>;
210
211// mve_addr_rq_shift  := reg + vreg{ << UXTW #shift}
212class mve_addr_rq_shift<int shift> : MemOperand {
213  let EncoderMethod = "getMveAddrModeRQOpValue";
214  let PrintMethod = "printMveAddrModeRQOperand<"#shift#">";
215  let ParserMatchClass =
216    !cast<AsmOperandClass>("MemRegRQS"#shift#"OffsetAsmOperand");
217  let DecoderMethod = "DecodeMveAddrModeRQ";
218  let MIOperandInfo = (ops GPRnopc:$base, MQPR:$offsreg);
219}
220
221class MemRegQOffsetAsmOperand<int shift> : AsmOperandClass {
222  let Name = "MemRegQS"#shift#"Offset";
223  let PredicateMethod = "isMemRegQOffset<"#shift#">";
224  let RenderMethod = "addMemImmOffsetOperands";
225}
226
227def MemRegQS2OffsetAsmOperand : MemRegQOffsetAsmOperand<2>;
228def MemRegQS3OffsetAsmOperand : MemRegQOffsetAsmOperand<3>;
229
230// mve_addr_q_shift  := vreg {+ #imm7s2/4}
231class mve_addr_q_shift<int shift> : MemOperand {
232  let EncoderMethod = "getMveAddrModeQOpValue<"#shift#">";
233  // Can be printed same way as other reg + imm operands
234  let PrintMethod = "printT2AddrModeImm8Operand<false>";
235  let ParserMatchClass =
236    !cast<AsmOperandClass>("MemRegQS"#shift#"OffsetAsmOperand");
237  let DecoderMethod = "DecodeMveAddrModeQ<"#shift#">";
238  let MIOperandInfo = (ops MQPR:$base, i32imm:$imm);
239}
240
241// A family of classes wrapping up information about the vector types
242// used by MVE.
243class MVEVectorVTInfo<ValueType vec, ValueType dblvec,
244                      ValueType pred, ValueType dblpred,
245                      bits<2> size, string suffixletter, bit unsigned> {
246  // The LLVM ValueType representing the vector, so we can use it in
247  // ISel patterns.
248  ValueType Vec = vec;
249
250  // The LLVM ValueType representing a vector with elements double the size
251  // of those in Vec, so we can use it in ISel patterns. It is up to the
252  // invoker of this class to ensure that this is a correct choice.
253  ValueType DblVec = dblvec;
254
255  // An LLVM ValueType representing a corresponding vector of
256  // predicate bits, for use in ISel patterns that handle an IR
257  // intrinsic describing the predicated form of the instruction.
258  ValueType Pred = pred;
259
260  // Same as Pred but for DblVec rather than Vec.
261  ValueType DblPred = dblpred;
262
263  // The most common representation of the vector element size in MVE
264  // instruction encodings: a 2-bit value V representing an (8<<V)-bit
265  // vector element.
266  bits<2> Size = size;
267
268  // For vectors explicitly mentioning a signedness of integers: 0 for
269  // signed and 1 for unsigned. For anything else, undefined.
270  bit Unsigned = unsigned;
271
272  // The number of bits in a vector element, in integer form.
273  int LaneBits = !shl(8, Size);
274
275  // The suffix used in assembly language on an instruction operating
276  // on this lane if it only cares about number of bits.
277  string BitsSuffix = !if(!eq(suffixletter, "p"),
278                          !if(!eq(unsigned, 0b0), "8", "16"),
279                          !cast<string>(LaneBits));
280
281  // The suffix used on an instruction that mentions the whole type.
282  string Suffix = suffixletter # BitsSuffix;
283
284  // The letter part of the suffix only.
285  string SuffixLetter = suffixletter;
286}
287
288// Integer vector types that don't treat signed and unsigned differently.
289def MVE_v16i8 : MVEVectorVTInfo<v16i8, v8i16, v16i1, v8i1, 0b00, "i", ?>;
290def MVE_v8i16 : MVEVectorVTInfo<v8i16, v4i32, v8i1,  v4i1, 0b01, "i", ?>;
291def MVE_v4i32 : MVEVectorVTInfo<v4i32, v2i64, v4i1,  v2i1, 0b10, "i", ?>;
292def MVE_v2i64 : MVEVectorVTInfo<v2i64, ?,     v2i1,  ?,    0b11, "i", ?>;
293
294// Explicitly signed and unsigned integer vectors. They map to the
295// same set of LLVM ValueTypes as above, but are represented
296// differently in assembly and instruction encodings.
297def MVE_v16s8 : MVEVectorVTInfo<v16i8, v8i16, v16i1, v8i1, 0b00, "s", 0b0>;
298def MVE_v8s16 : MVEVectorVTInfo<v8i16, v4i32, v8i1,  v4i1, 0b01, "s", 0b0>;
299def MVE_v4s32 : MVEVectorVTInfo<v4i32, v2i64, v4i1,  v2i1, 0b10, "s", 0b0>;
300def MVE_v2s64 : MVEVectorVTInfo<v2i64, ?,     v2i1,  ?,    0b11, "s", 0b0>;
301def MVE_v16u8 : MVEVectorVTInfo<v16i8, v8i16, v16i1, v8i1, 0b00, "u", 0b1>;
302def MVE_v8u16 : MVEVectorVTInfo<v8i16, v4i32, v8i1,  v4i1, 0b01, "u", 0b1>;
303def MVE_v4u32 : MVEVectorVTInfo<v4i32, v2i64, v4i1,  v2i1, 0b10, "u", 0b1>;
304def MVE_v2u64 : MVEVectorVTInfo<v2i64, ?,     v2i1,  ?,    0b11, "u", 0b1>;
305
306// FP vector types.
307def MVE_v8f16 : MVEVectorVTInfo<v8f16, v4f32, v8i1,  v4i1, 0b01, "f", ?>;
308def MVE_v4f32 : MVEVectorVTInfo<v4f32, v2f64, v4i1,  v2i1, 0b10, "f", ?>;
309def MVE_v2f64 : MVEVectorVTInfo<v2f64, ?,     v2i1,  ?,    0b11, "f", ?>;
310
311// Polynomial vector types.
312def MVE_v16p8 : MVEVectorVTInfo<v16i8, v8i16, v16i1, v8i1, 0b11, "p", 0b0>;
313def MVE_v8p16 : MVEVectorVTInfo<v8i16, v4i32, v8i1,  v4i1, 0b11, "p", 0b1>;
314
315multiclass MVE_TwoOpPattern<MVEVectorVTInfo VTI, SDPatternOperator Op, Intrinsic PredInt,
316                            dag PredOperands, Instruction Inst,
317                            SDPatternOperator IdentityVec = null_frag> {
318  // Unpredicated
319  def : Pat<(VTI.Vec (Op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn))),
320            (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
321
322  // Predicated with select
323  if !ne(VTI.Size, 0b11) then {
324    def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$mask),
325                                (VTI.Vec (Op (VTI.Vec MQPR:$Qm),
326                                             (VTI.Vec MQPR:$Qn))),
327                                (VTI.Vec MQPR:$inactive))),
328              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
329                              ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg,
330                              (VTI.Vec MQPR:$inactive)))>;
331
332    // Optionally with the select folded through the op
333    def : Pat<(VTI.Vec (Op (VTI.Vec MQPR:$Qm),
334                           (VTI.Vec (vselect (VTI.Pred VCCR:$mask),
335                                             (VTI.Vec MQPR:$Qn),
336                                             (VTI.Vec IdentityVec))))),
337              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
338                              ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg,
339                              (VTI.Vec MQPR:$Qm)))>;
340  }
341
342  // Predicated with intrinsic
343  def : Pat<(VTI.Vec !con((PredInt (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)),
344                          PredOperands,
345                          (? (VTI.Pred VCCR:$mask), (VTI.Vec MQPR:$inactive)))),
346            (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
347                            ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg,
348                            (VTI.Vec MQPR:$inactive)))>;
349}
350
351multiclass MVE_TwoOpPatternDup<MVEVectorVTInfo VTI, SDPatternOperator Op, Intrinsic PredInt,
352                               dag PredOperands, Instruction Inst,
353                               SDPatternOperator IdentityVec = null_frag> {
354  // Unpredicated
355  def : Pat<(VTI.Vec (Op (VTI.Vec MQPR:$Qm), (VTI.Vec (ARMvdup rGPR:$Rn)))),
356            (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), rGPR:$Rn))>;
357
358  // Predicated with select
359  if !ne(VTI.Size, 0b11) then {
360    def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$mask),
361                                (VTI.Vec (Op (VTI.Vec MQPR:$Qm),
362                                             (VTI.Vec (ARMvdup rGPR:$Rn)))),
363                                (VTI.Vec MQPR:$inactive))),
364              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), rGPR:$Rn,
365                              ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg,
366                              (VTI.Vec MQPR:$inactive)))>;
367
368    // Optionally with the select folded through the op
369    def : Pat<(VTI.Vec (Op (VTI.Vec MQPR:$Qm),
370                           (VTI.Vec (vselect (VTI.Pred VCCR:$mask),
371                                             (ARMvdup rGPR:$Rn),
372                                             (VTI.Vec IdentityVec))))),
373              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), rGPR:$Rn,
374                              ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg,
375                              (VTI.Vec MQPR:$Qm)))>;
376  }
377
378  // Predicated with intrinsic
379  def : Pat<(VTI.Vec !con((PredInt (VTI.Vec MQPR:$Qm), (VTI.Vec (ARMvdup rGPR:$Rn))),
380                          PredOperands,
381                          (? (VTI.Pred VCCR:$mask), (VTI.Vec MQPR:$inactive)))),
382            (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), rGPR:$Rn,
383                            ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg,
384                            (VTI.Vec MQPR:$inactive)))>;
385}
386
387// --------- Start of base classes for the instructions themselves
388
389class MVE_MI<dag oops, dag iops, InstrItinClass itin, string asm,
390             string ops, string cstr, bits<2> vecsize, list<dag> pattern>
391  : Thumb2XI<oops, iops, AddrModeNone, 4, itin, !strconcat(asm, "\t", ops), cstr,
392             pattern>,
393    Requires<[HasMVEInt]> {
394  let D = MVEDomain;
395  let DecoderNamespace = "MVE";
396  let VecSize = vecsize;
397}
398
399// MVE_p is used for most predicated instructions, to add the cluster
400// of input operands that provides the VPT suffix (none, T or E) and
401// the input predicate register.
402class MVE_p<dag oops, dag iops, InstrItinClass itin, string iname,
403            string suffix, string ops, vpred_ops vpred, string cstr,
404            bits<2> vecsize, list<dag> pattern=[]>
405  : MVE_MI<oops, !con(iops, (ins vpred:$vp)), itin,
406           // If the instruction has a suffix, like vadd.f32, then the
407           // VPT predication suffix goes before the dot, so the full
408           // name has to be "vadd${vp}.f32".
409           !strconcat(iname, "${vp}",
410                      !if(!eq(suffix, ""), "", !strconcat(".", suffix))),
411           ops, !strconcat(cstr, vpred.vpred_constraint), vecsize, pattern> {
412  let Inst{31-29} = 0b111;
413  let Inst{27-26} = 0b11;
414}
415
416class MVE_f<dag oops, dag iops, InstrItinClass itin, string iname,
417            string suffix, string ops, vpred_ops vpred, string cstr,
418            bits<2> vecsize, list<dag> pattern=[]>
419  : MVE_p<oops, iops, itin, iname, suffix, ops, vpred, cstr, vecsize, pattern> {
420  let Predicates = [HasMVEFloat];
421}
422
423class MVE_MI_with_pred<dag oops, dag iops, InstrItinClass itin, string asm,
424                       string ops, string cstr, list<dag> pattern>
425  : Thumb2I<oops, iops, AddrModeNone, 4, itin, asm, !strconcat("\t", ops), cstr,
426             pattern>,
427    Requires<[HasV8_1MMainline, HasMVEInt]> {
428  let D = MVEDomain;
429  let DecoderNamespace = "MVE";
430}
431
432class MVE_VMOV_lane_base<dag oops, dag iops, InstrItinClass itin, string asm,
433                         string suffix, string ops, string cstr,
434                         list<dag> pattern>
435  : Thumb2I<oops, iops, AddrModeNone, 4, itin, asm,
436            !if(!eq(suffix, ""), "", "." # suffix) # "\t" # ops,
437            cstr, pattern>,
438    Requires<[HasV8_1MMainline, HasMVEInt]> {
439  let D = MVEDomain;
440  let DecoderNamespace = "MVE";
441}
442
443class MVE_ScalarShift<string iname, dag oops, dag iops, string asm, string cstr,
444            list<dag> pattern=[]>
445  : MVE_MI_with_pred<oops, iops, NoItinerary, iname, asm, cstr, pattern> {
446  let Inst{31-20} = 0b111010100101;
447  let Inst{8} = 0b1;
448  let validForTailPredication=1;
449}
450
451class MVE_ScalarShiftSingleReg<string iname, dag iops, string asm, string cstr,
452                    list<dag> pattern=[]>
453  : MVE_ScalarShift<iname, (outs rGPR:$RdaDest), iops, asm, cstr, pattern> {
454  bits<4> RdaDest;
455
456  let Inst{19-16} = RdaDest{3-0};
457}
458
459class MVE_ScalarShiftSRegImm<string iname, bits<2> op5_4>
460  : MVE_ScalarShiftSingleReg<iname, (ins rGPR:$RdaSrc, long_shift:$imm),
461                     "$RdaSrc, $imm", "$RdaDest = $RdaSrc",
462                     [(set rGPR:$RdaDest,
463                          (i32 (!cast<Intrinsic>("int_arm_mve_" # iname)
464                                    (i32 rGPR:$RdaSrc), (i32 imm:$imm))))]> {
465  bits<5> imm;
466
467  let Inst{15} = 0b0;
468  let Inst{14-12} = imm{4-2};
469  let Inst{11-8} = 0b1111;
470  let Inst{7-6} = imm{1-0};
471  let Inst{5-4} = op5_4{1-0};
472  let Inst{3-0} = 0b1111;
473}
474
475def MVE_SQSHL : MVE_ScalarShiftSRegImm<"sqshl", 0b11>;
476def MVE_SRSHR : MVE_ScalarShiftSRegImm<"srshr", 0b10>;
477def MVE_UQSHL : MVE_ScalarShiftSRegImm<"uqshl", 0b00>;
478def MVE_URSHR : MVE_ScalarShiftSRegImm<"urshr", 0b01>;
479
480class MVE_ScalarShiftSRegReg<string iname, bits<2> op5_4>
481  : MVE_ScalarShiftSingleReg<iname, (ins rGPR:$RdaSrc, rGPR:$Rm),
482                     "$RdaSrc, $Rm", "@earlyclobber $RdaDest,$RdaDest = $RdaSrc",
483                     [(set rGPR:$RdaDest,
484                         (i32 (!cast<Intrinsic>("int_arm_mve_" # iname)
485                                   (i32 rGPR:$RdaSrc), (i32 rGPR:$Rm))))]> {
486  bits<4> Rm;
487
488  let Inst{15-12} = Rm{3-0};
489  let Inst{11-8} = 0b1111;
490  let Inst{7-6} = 0b00;
491  let Inst{5-4} = op5_4{1-0};
492  let Inst{3-0} = 0b1101;
493
494  let Unpredictable{8-6} = 0b111;
495}
496
497def MVE_SQRSHR : MVE_ScalarShiftSRegReg<"sqrshr", 0b10>;
498def MVE_UQRSHL : MVE_ScalarShiftSRegReg<"uqrshl", 0b00>;
499
500class MVE_ScalarShiftDoubleReg<string iname, dag iops, string asm,
501                               string cstr, list<dag> pattern=[]>
502  : MVE_ScalarShift<iname, (outs tGPREven:$RdaLo, tGPROdd:$RdaHi),
503                    iops, asm, cstr, pattern> {
504  bits<4> RdaLo;
505  bits<4> RdaHi;
506
507  let Inst{19-17} = RdaLo{3-1};
508  let Inst{11-9} = RdaHi{3-1};
509
510  let hasSideEffects = 0;
511}
512
513class MVE_ScalarShiftDRegImm<string iname, bits<2> op5_4, bit op16,
514                             list<dag> pattern=[]>
515  : MVE_ScalarShiftDoubleReg<
516      iname, (ins tGPREven:$RdaLo_src, tGPROdd:$RdaHi_src, long_shift:$imm),
517      "$RdaLo, $RdaHi, $imm", "$RdaLo = $RdaLo_src,$RdaHi = $RdaHi_src",
518      pattern> {
519  bits<5> imm;
520
521  let Inst{16} = op16;
522  let Inst{15} = 0b0;
523  let Inst{14-12} = imm{4-2};
524  let Inst{7-6} = imm{1-0};
525  let Inst{5-4} = op5_4{1-0};
526  let Inst{3-0} = 0b1111;
527}
528
529class MVE_ScalarShiftDRegRegBase<string iname, dag iops, string asm,
530                                 bit op5, bit op16, list<dag> pattern=[]>
531  : MVE_ScalarShiftDoubleReg<
532     iname, iops, asm, "@earlyclobber $RdaHi,@earlyclobber $RdaLo,"
533                       "$RdaLo = $RdaLo_src,$RdaHi = $RdaHi_src",
534     pattern> {
535  bits<4> Rm;
536
537  let Inst{16} = op16;
538  let Inst{15-12} = Rm{3-0};
539  let Inst{6} = 0b0;
540  let Inst{5} = op5;
541  let Inst{4} = 0b0;
542  let Inst{3-0} = 0b1101;
543
544  // Custom decoder method because of the following overlapping encodings:
545  // ASRL and SQRSHR
546  // LSLL and UQRSHL
547  // SQRSHRL and SQRSHR
548  // UQRSHLL and UQRSHL
549  let DecoderMethod = "DecodeMVEOverlappingLongShift";
550}
551
552class MVE_ScalarShiftDRegReg<string iname, bit op5, list<dag> pattern=[]>
553  : MVE_ScalarShiftDRegRegBase<
554     iname, (ins tGPREven:$RdaLo_src, tGPROdd:$RdaHi_src, rGPR:$Rm),
555     "$RdaLo, $RdaHi, $Rm", op5, 0b0, pattern> {
556
557  let Inst{7} = 0b0;
558}
559
560class MVE_ScalarShiftDRegRegWithSat<string iname, bit op5, list<dag> pattern=[]>
561  : MVE_ScalarShiftDRegRegBase<
562     iname, (ins tGPREven:$RdaLo_src, tGPROdd:$RdaHi_src, rGPR:$Rm, saturateop:$sat),
563     "$RdaLo, $RdaHi, $sat, $Rm", op5, 0b1, pattern> {
564  bit sat;
565
566  let Inst{7} = sat;
567}
568
569def MVE_ASRLr   : MVE_ScalarShiftDRegReg<"asrl",    0b1,  [(set tGPREven:$RdaLo, tGPROdd:$RdaHi,
570                                        (ARMasrl tGPREven:$RdaLo_src,
571                                        tGPROdd:$RdaHi_src, rGPR:$Rm))]>;
572def MVE_ASRLi   : MVE_ScalarShiftDRegImm<"asrl",    0b10, ?, [(set tGPREven:$RdaLo, tGPROdd:$RdaHi,
573                                        (ARMasrl tGPREven:$RdaLo_src,
574                                        tGPROdd:$RdaHi_src, (i32 long_shift:$imm)))]>;
575def MVE_LSLLr   : MVE_ScalarShiftDRegReg<"lsll",    0b0,  [(set tGPREven:$RdaLo, tGPROdd:$RdaHi,
576                                        (ARMlsll tGPREven:$RdaLo_src,
577                                        tGPROdd:$RdaHi_src, rGPR:$Rm))]>;
578def MVE_LSLLi   : MVE_ScalarShiftDRegImm<"lsll",    0b00, ?, [(set tGPREven:$RdaLo, tGPROdd:$RdaHi,
579                                        (ARMlsll tGPREven:$RdaLo_src,
580                                        tGPROdd:$RdaHi_src, (i32 long_shift:$imm)))]>;
581def MVE_LSRL    : MVE_ScalarShiftDRegImm<"lsrl",    0b01, ?, [(set tGPREven:$RdaLo, tGPROdd:$RdaHi,
582                                        (ARMlsrl tGPREven:$RdaLo_src,
583                                        tGPROdd:$RdaHi_src, (i32 long_shift:$imm)))]>;
584
585def MVE_SQRSHRL : MVE_ScalarShiftDRegRegWithSat<"sqrshrl", 0b1>;
586def MVE_SQSHLL  : MVE_ScalarShiftDRegImm<"sqshll",  0b11, 0b1>;
587def MVE_SRSHRL  : MVE_ScalarShiftDRegImm<"srshrl",  0b10, 0b1>;
588
589def MVE_UQRSHLL : MVE_ScalarShiftDRegRegWithSat<"uqrshll", 0b0>;
590def MVE_UQSHLL  : MVE_ScalarShiftDRegImm<"uqshll",  0b00, 0b1>;
591def MVE_URSHRL  : MVE_ScalarShiftDRegImm<"urshrl",  0b01, 0b1>;
592
593// start of mve_rDest instructions
594
595class MVE_rDest<dag oops, dag iops, InstrItinClass itin,
596                string iname, string suffix,
597                string ops, string cstr, bits<2> vecsize, list<dag> pattern=[]>
598// Always use vpred_n and not vpred_r: with the output register being
599// a GPR and not a vector register, there can't be any question of
600// what to put in its inactive lanes.
601  : MVE_p<oops, iops, itin, iname, suffix, ops, vpred_n, cstr, vecsize, pattern> {
602
603  let Inst{25-23} = 0b101;
604  let Inst{11-9} = 0b111;
605  let Inst{4} = 0b0;
606}
607
608class MVE_VABAV<string suffix, bit U, bits<2> size>
609  : MVE_rDest<(outs rGPR:$Rda), (ins rGPR:$Rda_src, MQPR:$Qn, MQPR:$Qm),
610              NoItinerary, "vabav", suffix, "$Rda, $Qn, $Qm", "$Rda = $Rda_src",
611              size, []> {
612  bits<4> Qm;
613  bits<4> Qn;
614  bits<4> Rda;
615
616  let Inst{28} = U;
617  let Inst{22} = 0b0;
618  let Inst{21-20} = size{1-0};
619  let Inst{19-17} = Qn{2-0};
620  let Inst{16} = 0b0;
621  let Inst{15-12} = Rda{3-0};
622  let Inst{8} = 0b1;
623  let Inst{7} = Qn{3};
624  let Inst{6} = 0b0;
625  let Inst{5} = Qm{3};
626  let Inst{3-1} = Qm{2-0};
627  let Inst{0} = 0b1;
628  let horizontalReduction = 1;
629}
630
631multiclass MVE_VABAV_m<MVEVectorVTInfo VTI> {
632  def "" : MVE_VABAV<VTI.Suffix, VTI.Unsigned, VTI.Size>;
633  defvar Inst = !cast<Instruction>(NAME);
634
635  let Predicates = [HasMVEInt] in {
636    def : Pat<(i32 (int_arm_mve_vabav
637                         (i32 VTI.Unsigned),
638                         (i32 rGPR:$Rda_src),
639                         (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
640              (i32 (Inst (i32 rGPR:$Rda_src),
641                         (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm)))>;
642
643    def : Pat<(i32 (int_arm_mve_vabav_predicated
644                         (i32 VTI.Unsigned),
645                         (i32 rGPR:$Rda_src),
646                         (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
647                         (VTI.Pred VCCR:$mask))),
648              (i32 (Inst (i32 rGPR:$Rda_src),
649                         (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
650                         ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg))>;
651  }
652}
653
654defm MVE_VABAVs8  : MVE_VABAV_m<MVE_v16s8>;
655defm MVE_VABAVs16 : MVE_VABAV_m<MVE_v8s16>;
656defm MVE_VABAVs32 : MVE_VABAV_m<MVE_v4s32>;
657defm MVE_VABAVu8  : MVE_VABAV_m<MVE_v16u8>;
658defm MVE_VABAVu16 : MVE_VABAV_m<MVE_v8u16>;
659defm MVE_VABAVu32 : MVE_VABAV_m<MVE_v4u32>;
660
661class MVE_VADDV<string iname, string suffix, dag iops, string cstr,
662              bit A, bit U, bits<2> size, list<dag> pattern=[]>
663  : MVE_rDest<(outs tGPREven:$Rda), iops, NoItinerary,
664              iname, suffix, "$Rda, $Qm", cstr, size, pattern> {
665  bits<3> Qm;
666  bits<4> Rda;
667
668  let Inst{28} = U;
669  let Inst{22-20} = 0b111;
670  let Inst{19-18} = size{1-0};
671  let Inst{17-16} = 0b01;
672  let Inst{15-13} = Rda{3-1};
673  let Inst{12} = 0b0;
674  let Inst{8-6} = 0b100;
675  let Inst{5} = A;
676  let Inst{3-1} = Qm{2-0};
677  let Inst{0} = 0b0;
678  let horizontalReduction = 1;
679  let validForTailPredication = 1;
680}
681
682def SDTVecReduceP : SDTypeProfile<1, 2, [    // VADDLVp
683  SDTCisInt<0>, SDTCisVec<1>, SDTCisVec<2>
684]>;
685def ARMVADDVs       : SDNode<"ARMISD::VADDVs", SDTVecReduce>;
686def ARMVADDVu       : SDNode<"ARMISD::VADDVu", SDTVecReduce>;
687def ARMVADDVps      : SDNode<"ARMISD::VADDVps", SDTVecReduceP>;
688def ARMVADDVpu      : SDNode<"ARMISD::VADDVpu", SDTVecReduceP>;
689
690multiclass MVE_VADDV_A<MVEVectorVTInfo VTI> {
691  def acc    : MVE_VADDV<"vaddva", VTI.Suffix,
692                         (ins tGPREven:$Rda_src, MQPR:$Qm), "$Rda = $Rda_src",
693                         0b1, VTI.Unsigned, VTI.Size>;
694  def no_acc : MVE_VADDV<"vaddv", VTI.Suffix,
695                         (ins MQPR:$Qm), "",
696                         0b0, VTI.Unsigned, VTI.Size>;
697
698  defvar InstA = !cast<Instruction>(NAME # "acc");
699  defvar InstN = !cast<Instruction>(NAME # "no_acc");
700
701  let Predicates = [HasMVEInt] in {
702    if VTI.Unsigned then {
703      def : Pat<(i32 (vecreduce_add (VTI.Vec MQPR:$vec))),
704                (i32 (InstN $vec))>;
705      def : Pat<(i32 (vecreduce_add (VTI.Vec (vselect (VTI.Pred VCCR:$pred),
706                                                      (VTI.Vec MQPR:$vec),
707                                                      (VTI.Vec ARMimmAllZerosV))))),
708                (i32 (InstN $vec, ARMVCCThen, $pred, zero_reg))>;
709      def : Pat<(i32 (ARMVADDVu (VTI.Vec MQPR:$vec))),
710                (i32 (InstN $vec))>;
711      def : Pat<(i32 (ARMVADDVpu (VTI.Vec MQPR:$vec), (VTI.Pred VCCR:$pred))),
712                (i32 (InstN $vec, ARMVCCThen, $pred, zero_reg))>;
713      def : Pat<(i32 (add (i32 (vecreduce_add (VTI.Vec MQPR:$vec))),
714                          (i32 tGPREven:$acc))),
715                (i32 (InstA $acc, $vec))>;
716      def : Pat<(i32 (add (i32 (vecreduce_add (VTI.Vec (vselect (VTI.Pred VCCR:$pred),
717                                                                (VTI.Vec MQPR:$vec),
718                                                                (VTI.Vec ARMimmAllZerosV))))),
719                          (i32 tGPREven:$acc))),
720                (i32 (InstA $acc, $vec, ARMVCCThen, $pred, zero_reg))>;
721      def : Pat<(i32 (add (i32 (ARMVADDVu (VTI.Vec MQPR:$vec))),
722                          (i32 tGPREven:$acc))),
723                (i32 (InstA $acc, $vec))>;
724      def : Pat<(i32 (add (i32 (ARMVADDVpu (VTI.Vec MQPR:$vec), (VTI.Pred VCCR:$pred))),
725                          (i32 tGPREven:$acc))),
726                (i32 (InstA $acc, $vec, ARMVCCThen, $pred, zero_reg))>;
727    } else {
728      def : Pat<(i32 (ARMVADDVs (VTI.Vec MQPR:$vec))),
729                (i32 (InstN $vec))>;
730      def : Pat<(i32 (add (i32 (ARMVADDVs (VTI.Vec MQPR:$vec))),
731                          (i32 tGPREven:$acc))),
732                (i32 (InstA $acc, $vec))>;
733      def : Pat<(i32 (ARMVADDVps (VTI.Vec MQPR:$vec), (VTI.Pred VCCR:$pred))),
734                (i32 (InstN $vec, ARMVCCThen, $pred, zero_reg))>;
735      def : Pat<(i32 (add (i32 (ARMVADDVps (VTI.Vec MQPR:$vec), (VTI.Pred VCCR:$pred))),
736                          (i32 tGPREven:$acc))),
737                (i32 (InstA $acc, $vec, ARMVCCThen, $pred, zero_reg))>;
738    }
739
740    def : Pat<(i32 (int_arm_mve_addv_predicated (VTI.Vec MQPR:$vec),
741                                                (i32 VTI.Unsigned),
742                                                (VTI.Pred VCCR:$pred))),
743              (i32 (InstN $vec, ARMVCCThen, $pred, zero_reg))>;
744    def : Pat<(i32 (add (int_arm_mve_addv_predicated (VTI.Vec MQPR:$vec),
745                                                     (i32 VTI.Unsigned),
746                                                     (VTI.Pred VCCR:$pred)),
747                        (i32 tGPREven:$acc))),
748              (i32 (InstA $acc, $vec, ARMVCCThen, $pred, zero_reg))>;
749  }
750}
751
752defm MVE_VADDVs8  : MVE_VADDV_A<MVE_v16s8>;
753defm MVE_VADDVs16 : MVE_VADDV_A<MVE_v8s16>;
754defm MVE_VADDVs32 : MVE_VADDV_A<MVE_v4s32>;
755defm MVE_VADDVu8  : MVE_VADDV_A<MVE_v16u8>;
756defm MVE_VADDVu16 : MVE_VADDV_A<MVE_v8u16>;
757defm MVE_VADDVu32 : MVE_VADDV_A<MVE_v4u32>;
758
759class MVE_VADDLV<string iname, string suffix, dag iops, string cstr,
760               bit A, bit U, list<dag> pattern=[]>
761  : MVE_rDest<(outs tGPREven:$RdaLo, tGPROdd:$RdaHi), iops, NoItinerary, iname,
762              suffix, "$RdaLo, $RdaHi, $Qm", cstr, 0b10, pattern> {
763  bits<3> Qm;
764  bits<4> RdaLo;
765  bits<4> RdaHi;
766
767  let Inst{28} = U;
768  let Inst{22-20} = RdaHi{3-1};
769  let Inst{19-18} = 0b10;
770  let Inst{17-16} = 0b01;
771  let Inst{15-13} = RdaLo{3-1};
772  let Inst{12} = 0b0;
773  let Inst{8-6} = 0b100;
774  let Inst{5} = A;
775  let Inst{3-1} = Qm{2-0};
776  let Inst{0} = 0b0;
777  let horizontalReduction = 1;
778}
779
780def SDTVecReduceL : SDTypeProfile<2, 1, [    // VADDLV
781  SDTCisInt<0>, SDTCisInt<1>, SDTCisVec<2>
782]>;
783def SDTVecReduceLA : SDTypeProfile<2, 3, [    // VADDLVA
784  SDTCisInt<0>, SDTCisInt<1>, SDTCisInt<2>, SDTCisInt<3>,
785  SDTCisVec<4>
786]>;
787def SDTVecReduceLP : SDTypeProfile<2, 2, [    // VADDLVp
788  SDTCisInt<0>, SDTCisInt<1>, SDTCisVec<2>, SDTCisVec<2>
789]>;
790def SDTVecReduceLPA : SDTypeProfile<2, 4, [    // VADDLVAp
791  SDTCisInt<0>, SDTCisInt<1>, SDTCisInt<2>, SDTCisInt<3>,
792  SDTCisVec<4>, SDTCisVec<5>
793]>;
794
795multiclass MVE_VADDLV_A<MVEVectorVTInfo VTI> {
796  def acc    : MVE_VADDLV<"vaddlva", VTI.Suffix,
797                        (ins tGPREven:$RdaLo_src, tGPROdd:$RdaHi_src, MQPR:$Qm),
798                        "$RdaLo = $RdaLo_src,$RdaHi = $RdaHi_src",
799                        0b1, VTI.Unsigned>;
800  def no_acc : MVE_VADDLV<"vaddlv", VTI.Suffix,
801                        (ins MQPR:$Qm), "",
802                        0b0, VTI.Unsigned>;
803
804  defvar InstA = !cast<Instruction>(NAME # "acc");
805  defvar InstN = !cast<Instruction>(NAME # "no_acc");
806
807  defvar letter = VTI.SuffixLetter;
808  defvar ARMVADDLV = SDNode<"ARMISD::VADDLV" # letter, SDTVecReduceL>;
809  defvar ARMVADDLVA = SDNode<"ARMISD::VADDLVA" # letter, SDTVecReduceLA>;
810  defvar ARMVADDLVp = SDNode<"ARMISD::VADDLVp" # letter, SDTVecReduceLP>;
811  defvar ARMVADDLVAp = SDNode<"ARMISD::VADDLVAp" # letter, SDTVecReduceLPA>;
812
813  let Predicates = [HasMVEInt] in {
814    def : Pat<(ARMVADDLV (v4i32 MQPR:$vec)),
815              (InstN (v4i32 MQPR:$vec))>;
816    def : Pat<(ARMVADDLVA tGPREven:$acclo, tGPROdd:$acchi, (v4i32 MQPR:$vec)),
817              (InstA tGPREven:$acclo, tGPROdd:$acchi, (v4i32 MQPR:$vec))>;
818    def : Pat<(ARMVADDLVp (v4i32 MQPR:$vec), (VTI.Pred VCCR:$pred)),
819              (InstN (v4i32 MQPR:$vec), ARMVCCThen, (VTI.Pred VCCR:$pred), zero_reg)>;
820    def : Pat<(ARMVADDLVAp tGPREven:$acclo, tGPROdd:$acchi, (v4i32 MQPR:$vec),
821                           (VTI.Pred VCCR:$pred)),
822              (InstA tGPREven:$acclo, tGPROdd:$acchi, (v4i32 MQPR:$vec),
823                     ARMVCCThen, (VTI.Pred VCCR:$pred), zero_reg)>;
824  }
825}
826
827defm MVE_VADDLVs32 : MVE_VADDLV_A<MVE_v4s32>;
828defm MVE_VADDLVu32 : MVE_VADDLV_A<MVE_v4u32>;
829
830class MVE_VMINMAXNMV<string iname, string suffix, bit sz,
831                     bit bit_17, bit bit_7, list<dag> pattern=[]>
832  : MVE_rDest<(outs rGPR:$RdaDest), (ins rGPR:$RdaSrc, MQPR:$Qm),
833              NoItinerary, iname, suffix, "$RdaSrc, $Qm",
834              "$RdaDest = $RdaSrc", !if(sz, 0b01, 0b10), pattern> {
835  bits<3> Qm;
836  bits<4> RdaDest;
837
838  let Inst{28} = sz;
839  let Inst{22-20} = 0b110;
840  let Inst{19-18} = 0b11;
841  let Inst{17} = bit_17;
842  let Inst{16} = 0b0;
843  let Inst{15-12} = RdaDest{3-0};
844  let Inst{8} = 0b1;
845  let Inst{7} = bit_7;
846  let Inst{6-5} = 0b00;
847  let Inst{3-1} = Qm{2-0};
848  let Inst{0} = 0b0;
849  let horizontalReduction = 1;
850
851  let Predicates = [HasMVEFloat];
852  let hasSideEffects = 0;
853}
854
855multiclass MVE_VMINMAXNMV_p<string iname, bit notAbs, bit isMin,
856                            MVEVectorVTInfo VTI, string intrBaseName,
857                            ValueType Scalar, RegisterClass ScalarReg> {
858  def "": MVE_VMINMAXNMV<iname, VTI.Suffix, VTI.Size{0}, notAbs, isMin>;
859  defvar Inst        = !cast<Instruction>(NAME);
860  defvar unpred_intr = !cast<Intrinsic>(intrBaseName);
861  defvar pred_intr   = !cast<Intrinsic>(intrBaseName#"_predicated");
862
863  let Predicates = [HasMVEFloat] in {
864    def : Pat<(Scalar (unpred_intr (Scalar ScalarReg:$prev),
865                                   (VTI.Vec MQPR:$vec))),
866           (COPY_TO_REGCLASS (Inst (COPY_TO_REGCLASS ScalarReg:$prev, rGPR),
867                                   (VTI.Vec MQPR:$vec)),
868                              ScalarReg)>;
869    def : Pat<(Scalar (pred_intr   (Scalar ScalarReg:$prev),
870                                   (VTI.Vec MQPR:$vec),
871                                   (VTI.Pred VCCR:$pred))),
872           (COPY_TO_REGCLASS (Inst (COPY_TO_REGCLASS ScalarReg:$prev, rGPR),
873                                   (VTI.Vec MQPR:$vec),
874                                   ARMVCCThen, (VTI.Pred VCCR:$pred), zero_reg),
875                              ScalarReg)>;
876  }
877}
878
879multiclass MVE_VMINMAXNMV_fty<string iname, bit notAbs, bit isMin,
880                              string intrBase> {
881  defm f32 : MVE_VMINMAXNMV_p<iname, notAbs, isMin, MVE_v4f32, intrBase,
882                              f32, SPR>;
883  defm f16 : MVE_VMINMAXNMV_p<iname, notAbs, isMin, MVE_v8f16, intrBase,
884                              f16, HPR>;
885}
886
887defm MVE_VMINNMV : MVE_VMINMAXNMV_fty<"vminnmv",  1, 1, "int_arm_mve_minnmv">;
888defm MVE_VMAXNMV : MVE_VMINMAXNMV_fty<"vmaxnmv",  1, 0, "int_arm_mve_maxnmv">;
889defm MVE_VMINNMAV: MVE_VMINMAXNMV_fty<"vminnmav", 0, 1, "int_arm_mve_minnmav">;
890defm MVE_VMAXNMAV: MVE_VMINMAXNMV_fty<"vmaxnmav", 0, 0, "int_arm_mve_maxnmav">;
891
892class MVE_VMINMAXV<string iname, string suffix, bit U, bits<2> size,
893                 bit bit_17, bit bit_7, list<dag> pattern=[]>
894  : MVE_rDest<(outs rGPR:$RdaDest), (ins rGPR:$RdaSrc, MQPR:$Qm), NoItinerary,
895              iname, suffix, "$RdaSrc, $Qm", "$RdaDest = $RdaSrc", size, pattern> {
896  bits<3> Qm;
897  bits<4> RdaDest;
898
899  let Inst{28} = U;
900  let Inst{22-20} = 0b110;
901  let Inst{19-18} = size{1-0};
902  let Inst{17} = bit_17;
903  let Inst{16} = 0b0;
904  let Inst{15-12} = RdaDest{3-0};
905  let Inst{8} = 0b1;
906  let Inst{7} = bit_7;
907  let Inst{6-5} = 0b00;
908  let Inst{3-1} = Qm{2-0};
909  let Inst{0} = 0b0;
910  let horizontalReduction = 1;
911}
912
913multiclass MVE_VMINMAXV_p<string iname, bit notAbs, bit isMin,
914                          MVEVectorVTInfo VTI, string intrBaseName> {
915  def "": MVE_VMINMAXV<iname, VTI.Suffix, VTI.Unsigned, VTI.Size,
916                       notAbs, isMin>;
917  defvar Inst        = !cast<Instruction>(NAME);
918  defvar unpred_intr = !cast<Intrinsic>(intrBaseName);
919  defvar pred_intr   = !cast<Intrinsic>(intrBaseName#"_predicated");
920  defvar base_args   = (? (i32 rGPR:$prev), (VTI.Vec MQPR:$vec));
921  defvar args        = !if(notAbs, !con(base_args, (? (i32 VTI.Unsigned))),
922                           base_args);
923
924  let Predicates = [HasMVEInt] in {
925    def : Pat<(i32 !con(args, (unpred_intr))),
926              (i32 (Inst (i32 rGPR:$prev), (VTI.Vec MQPR:$vec)))>;
927    def : Pat<(i32 !con(args, (pred_intr (VTI.Pred VCCR:$pred)))),
928              (i32 (Inst (i32 rGPR:$prev), (VTI.Vec MQPR:$vec),
929                         ARMVCCThen, (VTI.Pred VCCR:$pred), zero_reg))>;
930  }
931}
932
933multiclass MVE_VMINMAXV_ty<string iname, bit isMin, string intrBaseName> {
934  defm s8 : MVE_VMINMAXV_p<iname, 1, isMin, MVE_v16s8, intrBaseName>;
935  defm s16: MVE_VMINMAXV_p<iname, 1, isMin, MVE_v8s16, intrBaseName>;
936  defm s32: MVE_VMINMAXV_p<iname, 1, isMin, MVE_v4s32, intrBaseName>;
937  defm u8 : MVE_VMINMAXV_p<iname, 1, isMin, MVE_v16u8, intrBaseName>;
938  defm u16: MVE_VMINMAXV_p<iname, 1, isMin, MVE_v8u16, intrBaseName>;
939  defm u32: MVE_VMINMAXV_p<iname, 1, isMin, MVE_v4u32, intrBaseName>;
940}
941
942def SDTVecReduceR : SDTypeProfile<1, 2, [   // Reduction of an integer and vector into an integer
943  SDTCisInt<0>, SDTCisInt<1>, SDTCisVec<2>
944]>;
945def ARMVMINVu       : SDNode<"ARMISD::VMINVu", SDTVecReduceR>;
946def ARMVMINVs       : SDNode<"ARMISD::VMINVs", SDTVecReduceR>;
947def ARMVMAXVu       : SDNode<"ARMISD::VMAXVu", SDTVecReduceR>;
948def ARMVMAXVs       : SDNode<"ARMISD::VMAXVs", SDTVecReduceR>;
949
950defm MVE_VMINV : MVE_VMINMAXV_ty<"vminv", 1, "int_arm_mve_minv">;
951defm MVE_VMAXV : MVE_VMINMAXV_ty<"vmaxv", 0, "int_arm_mve_maxv">;
952
953let Predicates = [HasMVEInt] in {
954  def : Pat<(i32 (vecreduce_smax (v16i8 MQPR:$src))),
955            (i32 (MVE_VMAXVs8 (t2MVNi (i32 127)), $src))>;
956  def : Pat<(i32 (vecreduce_smax (v8i16 MQPR:$src))),
957            (i32 (MVE_VMAXVs16 (t2MOVi32imm (i32 -32768)), $src))>;
958  def : Pat<(i32 (vecreduce_smax (v4i32 MQPR:$src))),
959            (i32 (MVE_VMAXVs32 (t2MOVi (i32 -2147483648)), $src))>;
960  def : Pat<(i32 (vecreduce_umax (v16i8 MQPR:$src))),
961            (i32 (MVE_VMAXVu8 (t2MOVi (i32 0)), $src))>;
962  def : Pat<(i32 (vecreduce_umax (v8i16 MQPR:$src))),
963            (i32 (MVE_VMAXVu16 (t2MOVi (i32 0)), $src))>;
964  def : Pat<(i32 (vecreduce_umax (v4i32 MQPR:$src))),
965            (i32 (MVE_VMAXVu32 (t2MOVi (i32 0)), $src))>;
966
967  def : Pat<(i32 (vecreduce_smin (v16i8 MQPR:$src))),
968            (i32 (MVE_VMINVs8 (t2MOVi (i32 127)), $src))>;
969  def : Pat<(i32 (vecreduce_smin (v8i16 MQPR:$src))),
970            (i32 (MVE_VMINVs16 (t2MOVi16 (i32 32767)), $src))>;
971  def : Pat<(i32 (vecreduce_smin (v4i32 MQPR:$src))),
972            (i32 (MVE_VMINVs32 (t2MVNi (i32 -2147483648)), $src))>;
973  def : Pat<(i32 (vecreduce_umin (v16i8 MQPR:$src))),
974            (i32 (MVE_VMINVu8 (t2MOVi (i32 255)), $src))>;
975  def : Pat<(i32 (vecreduce_umin (v8i16 MQPR:$src))),
976            (i32 (MVE_VMINVu16 (t2MOVi16 (i32 65535)), $src))>;
977  def : Pat<(i32 (vecreduce_umin (v4i32 MQPR:$src))),
978            (i32 (MVE_VMINVu32 (t2MOVi (i32 4294967295)), $src))>;
979
980  def : Pat<(i32 (ARMVMINVu (i32 rGPR:$x), (v16i8 MQPR:$src))),
981            (i32 (MVE_VMINVu8 $x, $src))>;
982  def : Pat<(i32 (ARMVMINVu (i32 rGPR:$x), (v8i16 MQPR:$src))),
983            (i32 (MVE_VMINVu16 $x, $src))>;
984  def : Pat<(i32 (ARMVMINVu (i32 rGPR:$x), (v4i32 MQPR:$src))),
985            (i32 (MVE_VMINVu32 $x, $src))>;
986  def : Pat<(i32 (ARMVMINVs (i32 rGPR:$x), (v16i8 MQPR:$src))),
987            (i32 (MVE_VMINVs8 $x, $src))>;
988  def : Pat<(i32 (ARMVMINVs (i32 rGPR:$x), (v8i16 MQPR:$src))),
989            (i32 (MVE_VMINVs16 $x, $src))>;
990  def : Pat<(i32 (ARMVMINVs (i32 rGPR:$x), (v4i32 MQPR:$src))),
991            (i32 (MVE_VMINVs32 $x, $src))>;
992
993  def : Pat<(i32 (ARMVMAXVu (i32 rGPR:$x), (v16i8 MQPR:$src))),
994            (i32 (MVE_VMAXVu8 $x, $src))>;
995  def : Pat<(i32 (ARMVMAXVu (i32 rGPR:$x), (v8i16 MQPR:$src))),
996            (i32 (MVE_VMAXVu16 $x, $src))>;
997  def : Pat<(i32 (ARMVMAXVu (i32 rGPR:$x), (v4i32 MQPR:$src))),
998            (i32 (MVE_VMAXVu32 $x, $src))>;
999  def : Pat<(i32 (ARMVMAXVs (i32 rGPR:$x), (v16i8 MQPR:$src))),
1000            (i32 (MVE_VMAXVs8 $x, $src))>;
1001  def : Pat<(i32 (ARMVMAXVs (i32 rGPR:$x), (v8i16 MQPR:$src))),
1002            (i32 (MVE_VMAXVs16 $x, $src))>;
1003  def : Pat<(i32 (ARMVMAXVs (i32 rGPR:$x), (v4i32 MQPR:$src))),
1004            (i32 (MVE_VMAXVs32 $x, $src))>;
1005
1006}
1007
1008multiclass MVE_VMINMAXAV_ty<string iname, bit isMin, string intrBaseName> {
1009  defm s8 : MVE_VMINMAXV_p<iname, 0, isMin, MVE_v16s8, intrBaseName>;
1010  defm s16: MVE_VMINMAXV_p<iname, 0, isMin, MVE_v8s16, intrBaseName>;
1011  defm s32: MVE_VMINMAXV_p<iname, 0, isMin, MVE_v4s32, intrBaseName>;
1012}
1013
1014defm MVE_VMINAV : MVE_VMINMAXAV_ty<"vminav", 1, "int_arm_mve_minav">;
1015defm MVE_VMAXAV : MVE_VMINMAXAV_ty<"vmaxav", 0, "int_arm_mve_maxav">;
1016
1017class MVE_VMLAMLSDAV<string iname, string suffix, dag iops, string cstr,
1018                   bit sz, bit bit_28, bit A, bit X, bit bit_8, bit bit_0,
1019                   bits<2> vecsize>
1020  : MVE_rDest<(outs tGPREven:$RdaDest), iops, NoItinerary, iname, suffix,
1021              "$RdaDest, $Qn, $Qm", cstr, vecsize, []> {
1022  bits<4> RdaDest;
1023  bits<3> Qm;
1024  bits<3> Qn;
1025
1026  let Inst{28} = bit_28;
1027  let Inst{22-20} = 0b111;
1028  let Inst{19-17} = Qn{2-0};
1029  let Inst{16} = sz;
1030  let Inst{15-13} = RdaDest{3-1};
1031  let Inst{12} = X;
1032  let Inst{8} = bit_8;
1033  let Inst{7-6} = 0b00;
1034  let Inst{5} = A;
1035  let Inst{3-1} = Qm{2-0};
1036  let Inst{0} = bit_0;
1037  let horizontalReduction = 1;
1038  // Allow tail predication for non-exchanging versions. As this is also a
1039  // horizontalReduction, ARMLowOverheadLoops will also have to check that
1040  // the vector operands contain zeros in their false lanes for the instruction
1041  // to be properly valid.
1042  let validForTailPredication = !eq(X, 0);
1043}
1044
1045multiclass MVE_VMLAMLSDAV_A<string iname, string x, MVEVectorVTInfo VTI,
1046                            bit sz, bit bit_28, bit X, bit bit_8, bit bit_0> {
1047  def ""#x#VTI.Suffix : MVE_VMLAMLSDAV<iname # x, VTI.Suffix,
1048                                   (ins MQPR:$Qn, MQPR:$Qm), "",
1049                                   sz, bit_28, 0b0, X, bit_8, bit_0, VTI.Size>;
1050  def "a"#x#VTI.Suffix : MVE_VMLAMLSDAV<iname # "a" # x, VTI.Suffix,
1051                                    (ins tGPREven:$RdaSrc, MQPR:$Qn, MQPR:$Qm),
1052                                    "$RdaDest = $RdaSrc",
1053                                    sz, bit_28, 0b1, X, bit_8, bit_0, VTI.Size>;
1054  let Predicates = [HasMVEInt] in {
1055    def : Pat<(i32 (int_arm_mve_vmldava
1056                            (i32 VTI.Unsigned),
1057                            (i32 bit_0) /* subtract */,
1058                            (i32 X) /* exchange */,
1059                            (i32 0) /* accumulator */,
1060                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
1061              (i32 (!cast<Instruction>(NAME # x # VTI.Suffix)
1062                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm)))>;
1063
1064    def : Pat<(i32 (int_arm_mve_vmldava_predicated
1065                            (i32 VTI.Unsigned),
1066                            (i32 bit_0) /* subtract */,
1067                            (i32 X) /* exchange */,
1068                            (i32 0) /* accumulator */,
1069                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
1070                            (VTI.Pred VCCR:$mask))),
1071              (i32 (!cast<Instruction>(NAME # x # VTI.Suffix)
1072                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
1073                             ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg))>;
1074
1075    def : Pat<(i32 (int_arm_mve_vmldava
1076                            (i32 VTI.Unsigned),
1077                            (i32 bit_0) /* subtract */,
1078                            (i32 X) /* exchange */,
1079                            (i32 tGPREven:$RdaSrc),
1080                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
1081              (i32 (!cast<Instruction>(NAME # "a" # x # VTI.Suffix)
1082                            (i32 tGPREven:$RdaSrc),
1083                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm)))>;
1084
1085    def : Pat<(i32 (int_arm_mve_vmldava_predicated
1086                            (i32 VTI.Unsigned),
1087                            (i32 bit_0) /* subtract */,
1088                            (i32 X) /* exchange */,
1089                            (i32 tGPREven:$RdaSrc),
1090                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
1091                            (VTI.Pred VCCR:$mask))),
1092              (i32 (!cast<Instruction>(NAME # "a" # x # VTI.Suffix)
1093                            (i32 tGPREven:$RdaSrc),
1094                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
1095                             ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg))>;
1096  }
1097}
1098
1099multiclass MVE_VMLAMLSDAV_AX<string iname, MVEVectorVTInfo VTI, bit sz,
1100                             bit bit_28, bit bit_8, bit bit_0> {
1101  defm "" : MVE_VMLAMLSDAV_A<iname, "", VTI, sz, bit_28,
1102                             0b0, bit_8, bit_0>;
1103  defm "" : MVE_VMLAMLSDAV_A<iname, "x", VTI, sz, bit_28,
1104                             0b1, bit_8, bit_0>;
1105}
1106
1107multiclass MVE_VMLADAV_multi<MVEVectorVTInfo SVTI, MVEVectorVTInfo UVTI,
1108                             bit sz, bit bit_8> {
1109  defm "" : MVE_VMLAMLSDAV_AX<"vmladav", SVTI,
1110                              sz, 0b0, bit_8, 0b0>;
1111  defm "" : MVE_VMLAMLSDAV_A<"vmladav", "", UVTI,
1112                             sz, 0b1, 0b0, bit_8, 0b0>;
1113}
1114
1115multiclass MVE_VMLSDAV_multi<MVEVectorVTInfo VTI, bit sz, bit bit_28> {
1116  defm "" : MVE_VMLAMLSDAV_AX<"vmlsdav", VTI,
1117                              sz, bit_28, 0b0, 0b1>;
1118}
1119
1120defm MVE_VMLADAV : MVE_VMLADAV_multi<MVE_v16s8, MVE_v16u8, 0b0, 0b1>;
1121defm MVE_VMLADAV : MVE_VMLADAV_multi<MVE_v8s16, MVE_v8u16, 0b0, 0b0>;
1122defm MVE_VMLADAV : MVE_VMLADAV_multi<MVE_v4s32, MVE_v4u32, 0b1, 0b0>;
1123
1124defm MVE_VMLSDAV : MVE_VMLSDAV_multi<MVE_v16s8, 0b0, 0b1>;
1125defm MVE_VMLSDAV : MVE_VMLSDAV_multi<MVE_v8s16, 0b0, 0b0>;
1126defm MVE_VMLSDAV : MVE_VMLSDAV_multi<MVE_v4s32, 0b1, 0b0>;
1127
1128def SDTVecReduce2 : SDTypeProfile<1, 2, [    // VMLAV
1129  SDTCisInt<0>, SDTCisVec<1>, SDTCisVec<2>
1130]>;
1131def SDTVecReduce2L : SDTypeProfile<2, 2, [    // VMLALV
1132  SDTCisInt<0>, SDTCisInt<1>, SDTCisVec<2>, SDTCisVec<3>
1133]>;
1134def SDTVecReduce2LA : SDTypeProfile<2, 4, [    // VMLALVA
1135  SDTCisInt<0>, SDTCisInt<1>, SDTCisInt<2>, SDTCisInt<3>,
1136  SDTCisVec<4>, SDTCisVec<5>
1137]>;
1138def SDTVecReduce2P : SDTypeProfile<1, 3, [    // VMLAV
1139  SDTCisInt<0>, SDTCisVec<1>, SDTCisVec<2>, SDTCisVec<3>
1140]>;
1141def SDTVecReduce2LP : SDTypeProfile<2, 3, [    // VMLALV
1142  SDTCisInt<0>, SDTCisInt<1>, SDTCisVec<2>, SDTCisVec<3>, SDTCisVec<4>
1143]>;
1144def SDTVecReduce2LAP : SDTypeProfile<2, 5, [    // VMLALVA
1145  SDTCisInt<0>, SDTCisInt<1>, SDTCisInt<2>, SDTCisInt<3>,
1146  SDTCisVec<4>, SDTCisVec<5>, SDTCisVec<6>
1147]>;
1148def ARMVMLAVs       : SDNode<"ARMISD::VMLAVs", SDTVecReduce2>;
1149def ARMVMLAVu       : SDNode<"ARMISD::VMLAVu", SDTVecReduce2>;
1150def ARMVMLALVs      : SDNode<"ARMISD::VMLALVs", SDTVecReduce2L>;
1151def ARMVMLALVu      : SDNode<"ARMISD::VMLALVu", SDTVecReduce2L>;
1152def ARMVMLALVAs     : SDNode<"ARMISD::VMLALVAs", SDTVecReduce2LA>;
1153def ARMVMLALVAu     : SDNode<"ARMISD::VMLALVAu", SDTVecReduce2LA>;
1154def ARMVMLAVps      : SDNode<"ARMISD::VMLAVps", SDTVecReduce2P>;
1155def ARMVMLAVpu      : SDNode<"ARMISD::VMLAVpu", SDTVecReduce2P>;
1156def ARMVMLALVps     : SDNode<"ARMISD::VMLALVps", SDTVecReduce2LP>;
1157def ARMVMLALVpu     : SDNode<"ARMISD::VMLALVpu", SDTVecReduce2LP>;
1158def ARMVMLALVAps    : SDNode<"ARMISD::VMLALVAps", SDTVecReduce2LAP>;
1159def ARMVMLALVApu    : SDNode<"ARMISD::VMLALVApu", SDTVecReduce2LAP>;
1160
1161let Predicates = [HasMVEInt] in {
1162  def : Pat<(i32 (vecreduce_add (mul (v4i32 MQPR:$src1), (v4i32 MQPR:$src2)))),
1163            (i32 (MVE_VMLADAVu32 $src1, $src2))>;
1164  def : Pat<(i32 (vecreduce_add (mul (v8i16 MQPR:$src1), (v8i16 MQPR:$src2)))),
1165            (i32 (MVE_VMLADAVu16 $src1, $src2))>;
1166  def : Pat<(i32 (ARMVMLAVs (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))),
1167            (i32 (MVE_VMLADAVs16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)))>;
1168  def : Pat<(i32 (ARMVMLAVu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))),
1169            (i32 (MVE_VMLADAVu16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)))>;
1170  def : Pat<(i32 (vecreduce_add (mul (v16i8 MQPR:$src1), (v16i8 MQPR:$src2)))),
1171            (i32 (MVE_VMLADAVu8 $src1, $src2))>;
1172  def : Pat<(i32 (ARMVMLAVs (v16i8 MQPR:$val1), (v16i8 MQPR:$val2))),
1173            (i32 (MVE_VMLADAVs8 (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)))>;
1174  def : Pat<(i32 (ARMVMLAVu (v16i8 MQPR:$val1), (v16i8 MQPR:$val2))),
1175            (i32 (MVE_VMLADAVu8 (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)))>;
1176
1177  def : Pat<(i32 (add (i32 (vecreduce_add (mul (v4i32 MQPR:$src1), (v4i32 MQPR:$src2)))),
1178                      (i32 tGPREven:$src3))),
1179            (i32 (MVE_VMLADAVau32 $src3, $src1, $src2))>;
1180  def : Pat<(i32 (add (i32 (vecreduce_add (mul (v8i16 MQPR:$src1), (v8i16 MQPR:$src2)))),
1181                      (i32 tGPREven:$src3))),
1182            (i32 (MVE_VMLADAVau16 $src3, $src1, $src2))>;
1183  def : Pat<(i32 (add (ARMVMLAVs (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)), tGPREven:$Rd)),
1184            (i32 (MVE_VMLADAVas16 tGPREven:$Rd, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)))>;
1185  def : Pat<(i32 (add (ARMVMLAVu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)), tGPREven:$Rd)),
1186            (i32 (MVE_VMLADAVau16 tGPREven:$Rd, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)))>;
1187  def : Pat<(i32 (add (i32 (vecreduce_add (mul (v16i8 MQPR:$src1), (v16i8 MQPR:$src2)))),
1188                      (i32 tGPREven:$src3))),
1189            (i32 (MVE_VMLADAVau8 $src3, $src1, $src2))>;
1190  def : Pat<(i32 (add (ARMVMLAVs (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)), tGPREven:$Rd)),
1191            (i32 (MVE_VMLADAVas8 tGPREven:$Rd, (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)))>;
1192  def : Pat<(i32 (add (ARMVMLAVu (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)), tGPREven:$Rd)),
1193            (i32 (MVE_VMLADAVau8 tGPREven:$Rd, (v16i8 MQPR:$val1), (v16i8 MQPR:$val2)))>;
1194
1195  // Predicated
1196  def : Pat<(i32 (vecreduce_add (vselect (v4i1 VCCR:$pred),
1197                                         (mul (v4i32 MQPR:$src1), (v4i32 MQPR:$src2)),
1198                                         (v4i32 ARMimmAllZerosV)))),
1199            (i32 (MVE_VMLADAVu32 $src1, $src2, ARMVCCThen, $pred, zero_reg))>;
1200  def : Pat<(i32 (vecreduce_add (vselect (v8i1 VCCR:$pred),
1201                                         (mul (v8i16 MQPR:$src1), (v8i16 MQPR:$src2)),
1202                                         (v8i16 ARMimmAllZerosV)))),
1203            (i32 (MVE_VMLADAVu16 $src1, $src2, ARMVCCThen, $pred, zero_reg))>;
1204  def : Pat<(i32 (ARMVMLAVps (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred))),
1205            (i32 (MVE_VMLADAVs16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred, zero_reg))>;
1206  def : Pat<(i32 (ARMVMLAVpu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred))),
1207            (i32 (MVE_VMLADAVu16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred, zero_reg))>;
1208  def : Pat<(i32 (vecreduce_add (vselect (v16i1 VCCR:$pred),
1209                                         (mul (v16i8 MQPR:$src1), (v16i8 MQPR:$src2)),
1210                                         (v16i8 ARMimmAllZerosV)))),
1211            (i32 (MVE_VMLADAVu8 $src1, $src2, ARMVCCThen, $pred, zero_reg))>;
1212  def : Pat<(i32 (ARMVMLAVps (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), (v16i1 VCCR:$pred))),
1213            (i32 (MVE_VMLADAVs8 (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), ARMVCCThen, $pred, zero_reg))>;
1214  def : Pat<(i32 (ARMVMLAVpu (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), (v16i1 VCCR:$pred))),
1215            (i32 (MVE_VMLADAVu8 (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), ARMVCCThen, $pred, zero_reg))>;
1216
1217  def : Pat<(i32 (add (i32 (vecreduce_add (vselect (v4i1 VCCR:$pred),
1218                                                   (mul (v4i32 MQPR:$src1), (v4i32 MQPR:$src2)),
1219                                                   (v4i32 ARMimmAllZerosV)))),
1220                      (i32 tGPREven:$src3))),
1221            (i32 (MVE_VMLADAVau32 $src3, $src1, $src2, ARMVCCThen, $pred, zero_reg))>;
1222  def : Pat<(i32 (add (i32 (vecreduce_add (vselect (v8i1 VCCR:$pred),
1223                                                   (mul (v8i16 MQPR:$src1), (v8i16 MQPR:$src2)),
1224                                                   (v8i16 ARMimmAllZerosV)))),
1225                      (i32 tGPREven:$src3))),
1226            (i32 (MVE_VMLADAVau16 $src3, $src1, $src2, ARMVCCThen, $pred, zero_reg))>;
1227  def : Pat<(i32 (add (ARMVMLAVps (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)), tGPREven:$Rd)),
1228            (i32 (MVE_VMLADAVas16 tGPREven:$Rd, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred, zero_reg))>;
1229  def : Pat<(i32 (add (ARMVMLAVpu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)), tGPREven:$Rd)),
1230            (i32 (MVE_VMLADAVau16 tGPREven:$Rd, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred, zero_reg))>;
1231  def : Pat<(i32 (add (i32 (vecreduce_add (vselect (v16i1 VCCR:$pred),
1232                                                   (mul (v16i8 MQPR:$src1), (v16i8 MQPR:$src2)),
1233                                                   (v16i8 ARMimmAllZerosV)))),
1234                      (i32 tGPREven:$src3))),
1235            (i32 (MVE_VMLADAVau8 $src3, $src1, $src2, ARMVCCThen, $pred, zero_reg))>;
1236  def : Pat<(i32 (add (ARMVMLAVps (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), (v16i1 VCCR:$pred)), tGPREven:$Rd)),
1237            (i32 (MVE_VMLADAVas8 tGPREven:$Rd, (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), ARMVCCThen, $pred, zero_reg))>;
1238  def : Pat<(i32 (add (ARMVMLAVpu (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), (v16i1 VCCR:$pred)), tGPREven:$Rd)),
1239            (i32 (MVE_VMLADAVau8 tGPREven:$Rd, (v16i8 MQPR:$val1), (v16i8 MQPR:$val2), ARMVCCThen, $pred, zero_reg))>;
1240}
1241
1242// vmlav aliases vmladav
1243foreach acc = ["", "a"] in {
1244  foreach suffix = ["s8", "s16", "s32", "u8", "u16", "u32"] in {
1245    def : MVEInstAlias<"vmlav"#acc#"${vp}."#suffix#"\t$RdaDest, $Qn, $Qm",
1246                       (!cast<Instruction>("MVE_VMLADAV"#acc#suffix)
1247                        tGPREven:$RdaDest, MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
1248  }
1249}
1250
1251// Base class for VMLALDAV and VMLSLDAV, VRMLALDAVH, VRMLSLDAVH
1252class MVE_VMLALDAVBase<string iname, string suffix, dag iops, string cstr,
1253                       bit sz, bit bit_28, bit A, bit X, bit bit_8, bit bit_0,
1254                       bits<2> vecsize, list<dag> pattern=[]>
1255  : MVE_rDest<(outs tGPREven:$RdaLoDest, tGPROdd:$RdaHiDest), iops, NoItinerary,
1256              iname, suffix, "$RdaLoDest, $RdaHiDest, $Qn, $Qm", cstr, vecsize, pattern> {
1257  bits<4> RdaLoDest;
1258  bits<4> RdaHiDest;
1259  bits<3> Qm;
1260  bits<3> Qn;
1261
1262  let Inst{28} = bit_28;
1263  let Inst{22-20} = RdaHiDest{3-1};
1264  let Inst{19-17} = Qn{2-0};
1265  let Inst{16} = sz;
1266  let Inst{15-13} = RdaLoDest{3-1};
1267  let Inst{12} = X;
1268  let Inst{8} = bit_8;
1269  let Inst{7-6} = 0b00;
1270  let Inst{5} = A;
1271  let Inst{3-1} = Qm{2-0};
1272  let Inst{0} = bit_0;
1273  let horizontalReduction = 1;
1274  // Allow tail predication for non-exchanging versions. As this is also a
1275  // horizontalReduction, ARMLowOverheadLoops will also have to check that
1276  // the vector operands contain zeros in their false lanes for the instruction
1277  // to be properly valid.
1278  let validForTailPredication = !eq(X, 0);
1279
1280  let hasSideEffects = 0;
1281}
1282
1283multiclass MVE_VMLALDAVBase_A<string iname, string x, string suffix,
1284                              bit sz, bit bit_28, bit X, bit bit_8, bit bit_0,
1285                              bits<2> vecsize, list<dag> pattern=[]> {
1286  def ""#x#suffix : MVE_VMLALDAVBase<
1287     iname # x, suffix, (ins MQPR:$Qn, MQPR:$Qm), "",
1288     sz, bit_28, 0b0, X, bit_8, bit_0, vecsize, pattern>;
1289  def "a"#x#suffix : MVE_VMLALDAVBase<
1290     iname # "a" # x, suffix,
1291     (ins tGPREven:$RdaLoSrc, tGPROdd:$RdaHiSrc, MQPR:$Qn, MQPR:$Qm),
1292     "$RdaLoDest = $RdaLoSrc,$RdaHiDest = $RdaHiSrc",
1293     sz, bit_28, 0b1, X, bit_8, bit_0, vecsize, pattern>;
1294}
1295
1296
1297multiclass MVE_VMLALDAVBase_AX<string iname, string suffix, bit sz, bit bit_28,
1298                               bit bit_8, bit bit_0, bits<2> vecsize, list<dag> pattern=[]> {
1299  defm "" : MVE_VMLALDAVBase_A<iname, "", suffix, sz,
1300                               bit_28, 0b0, bit_8, bit_0, vecsize, pattern>;
1301  defm "" : MVE_VMLALDAVBase_A<iname, "x", suffix, sz,
1302                               bit_28, 0b1, bit_8, bit_0, vecsize, pattern>;
1303}
1304
1305multiclass MVE_VRMLALDAVH_multi<MVEVectorVTInfo VTI, list<dag> pattern=[]> {
1306  defm "" : MVE_VMLALDAVBase_AX<"vrmlaldavh", "s"#VTI.BitsSuffix,
1307                                0b0, 0b0, 0b1, 0b0, VTI.Size, pattern>;
1308  defm "" : MVE_VMLALDAVBase_A<"vrmlaldavh", "", "u"#VTI.BitsSuffix,
1309                               0b0, 0b1, 0b0, 0b1, 0b0, VTI.Size, pattern>;
1310}
1311
1312defm MVE_VRMLALDAVH : MVE_VRMLALDAVH_multi<MVE_v4i32>;
1313
1314// vrmlalvh aliases for vrmlaldavh
1315def : MVEInstAlias<"vrmlalvh${vp}.s32\t$RdaLo, $RdaHi, $Qn, $Qm",
1316                  (MVE_VRMLALDAVHs32
1317                   tGPREven:$RdaLo, tGPROdd:$RdaHi,
1318                   MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
1319def : MVEInstAlias<"vrmlalvha${vp}.s32\t$RdaLo, $RdaHi, $Qn, $Qm",
1320                  (MVE_VRMLALDAVHas32
1321                   tGPREven:$RdaLo, tGPROdd:$RdaHi,
1322                   MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
1323def : MVEInstAlias<"vrmlalvh${vp}.u32\t$RdaLo, $RdaHi, $Qn, $Qm",
1324                  (MVE_VRMLALDAVHu32
1325                   tGPREven:$RdaLo, tGPROdd:$RdaHi,
1326                   MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
1327def : MVEInstAlias<"vrmlalvha${vp}.u32\t$RdaLo, $RdaHi, $Qn, $Qm",
1328                  (MVE_VRMLALDAVHau32
1329                   tGPREven:$RdaLo, tGPROdd:$RdaHi,
1330                   MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
1331
1332multiclass MVE_VMLALDAV_multi<MVEVectorVTInfo VTI, list<dag> pattern=[]> {
1333  defm "" : MVE_VMLALDAVBase_AX<"vmlaldav", "s"#VTI.BitsSuffix,
1334                                VTI.Size{1}, 0b0, 0b0, 0b0, VTI.Size, pattern>;
1335  defm "" : MVE_VMLALDAVBase_A<"vmlaldav", "", "u"#VTI.BitsSuffix,
1336                               VTI.Size{1}, 0b1, 0b0, 0b0, 0b0, VTI.Size, pattern>;
1337}
1338
1339defm MVE_VMLALDAV : MVE_VMLALDAV_multi<MVE_v8i16>;
1340defm MVE_VMLALDAV : MVE_VMLALDAV_multi<MVE_v4i32>;
1341
1342let Predicates = [HasMVEInt] in {
1343  def : Pat<(ARMVMLALVs (v4i32 MQPR:$val1), (v4i32 MQPR:$val2)),
1344            (MVE_VMLALDAVs32 (v4i32 MQPR:$val1), (v4i32 MQPR:$val2))>;
1345  def : Pat<(ARMVMLALVu (v4i32 MQPR:$val1), (v4i32 MQPR:$val2)),
1346            (MVE_VMLALDAVu32 (v4i32 MQPR:$val1), (v4i32 MQPR:$val2))>;
1347  def : Pat<(ARMVMLALVs (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)),
1348            (MVE_VMLALDAVs16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))>;
1349  def : Pat<(ARMVMLALVu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)),
1350            (MVE_VMLALDAVu16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))>;
1351
1352  def : Pat<(ARMVMLALVAs tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2)),
1353            (MVE_VMLALDAVas32 tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2))>;
1354  def : Pat<(ARMVMLALVAu tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2)),
1355            (MVE_VMLALDAVau32 tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2))>;
1356  def : Pat<(ARMVMLALVAs tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)),
1357            (MVE_VMLALDAVas16 tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))>;
1358  def : Pat<(ARMVMLALVAu tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2)),
1359            (MVE_VMLALDAVau16 tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2))>;
1360
1361  // Predicated
1362  def : Pat<(ARMVMLALVps (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), (v4i1 VCCR:$pred)),
1363            (MVE_VMLALDAVs32 (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), ARMVCCThen, $pred, zero_reg)>;
1364  def : Pat<(ARMVMLALVpu (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), (v4i1 VCCR:$pred)),
1365            (MVE_VMLALDAVu32 (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), ARMVCCThen, $pred, zero_reg)>;
1366  def : Pat<(ARMVMLALVps (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)),
1367            (MVE_VMLALDAVs16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred, zero_reg)>;
1368  def : Pat<(ARMVMLALVpu (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)),
1369            (MVE_VMLALDAVu16 (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred, zero_reg)>;
1370
1371  def : Pat<(ARMVMLALVAps tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), (v4i1 VCCR:$pred)),
1372            (MVE_VMLALDAVas32 tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), ARMVCCThen, $pred, zero_reg)>;
1373  def : Pat<(ARMVMLALVApu tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), (v4i1 VCCR:$pred)),
1374            (MVE_VMLALDAVau32 tGPREven:$Rda, tGPROdd:$Rdb, (v4i32 MQPR:$val1), (v4i32 MQPR:$val2), ARMVCCThen, $pred, zero_reg)>;
1375  def : Pat<(ARMVMLALVAps tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)),
1376            (MVE_VMLALDAVas16 tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred, zero_reg)>;
1377  def : Pat<(ARMVMLALVApu tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), (v8i1 VCCR:$pred)),
1378            (MVE_VMLALDAVau16 tGPREven:$Rda, tGPROdd:$Rdb, (v8i16 MQPR:$val1), (v8i16 MQPR:$val2), ARMVCCThen, $pred, zero_reg)>;
1379}
1380
1381// vmlalv aliases vmlaldav
1382foreach acc = ["", "a"] in {
1383  foreach suffix = ["s16", "s32", "u16", "u32"] in {
1384    def : MVEInstAlias<"vmlalv" # acc # "${vp}." # suffix #
1385                          "\t$RdaLoDest, $RdaHiDest, $Qn, $Qm",
1386                       (!cast<Instruction>("MVE_VMLALDAV"#acc#suffix)
1387                       tGPREven:$RdaLoDest, tGPROdd:$RdaHiDest,
1388                       MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
1389  }
1390}
1391
1392multiclass MVE_VMLSLDAV_multi<string iname, string suffix, bit sz,
1393                              bit bit_28, bits<2> vecsize, list<dag> pattern=[]> {
1394  defm "" : MVE_VMLALDAVBase_AX<iname, suffix, sz, bit_28, 0b0, 0b1, vecsize, pattern>;
1395}
1396
1397defm MVE_VMLSLDAV   : MVE_VMLSLDAV_multi<"vmlsldav", "s16", 0b0, 0b0, 0b01>;
1398defm MVE_VMLSLDAV   : MVE_VMLSLDAV_multi<"vmlsldav", "s32", 0b1, 0b0, 0b10>;
1399defm MVE_VRMLSLDAVH : MVE_VMLSLDAV_multi<"vrmlsldavh", "s32", 0b0, 0b1, 0b10>;
1400
1401// end of mve_rDest instructions
1402
1403// start of mve_comp instructions
1404
1405class MVE_comp<InstrItinClass itin, string iname, string suffix,
1406               string cstr, bits<2> vecsize, list<dag> pattern=[]>
1407  : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm), itin, iname, suffix,
1408           "$Qd, $Qn, $Qm", vpred_r, cstr, vecsize, pattern> {
1409  bits<4> Qd;
1410  bits<4> Qn;
1411  bits<4> Qm;
1412
1413  let Inst{22} = Qd{3};
1414  let Inst{19-17} = Qn{2-0};
1415  let Inst{16} = 0b0;
1416  let Inst{15-13} = Qd{2-0};
1417  let Inst{12} = 0b0;
1418  let Inst{10-9} = 0b11;
1419  let Inst{7} = Qn{3};
1420  let Inst{5} = Qm{3};
1421  let Inst{3-1} = Qm{2-0};
1422  let Inst{0} = 0b0;
1423}
1424
1425class MVE_VMINMAXNM<string iname, string suffix, bits<2> sz, bit bit_21,
1426                    list<dag> pattern=[]>
1427  : MVE_comp<NoItinerary, iname, suffix, "", sz, pattern> {
1428
1429  let Inst{28} = 0b1;
1430  let Inst{25-24} = 0b11;
1431  let Inst{23} = 0b0;
1432  let Inst{21} = bit_21;
1433  let Inst{20} = sz{0};
1434  let Inst{11} = 0b1;
1435  let Inst{8} = 0b1;
1436  let Inst{6} = 0b1;
1437  let Inst{4} = 0b1;
1438
1439  let Predicates = [HasMVEFloat];
1440  let validForTailPredication = 1;
1441}
1442
1443multiclass MVE_VMINMAXNM_m<string iname, bit bit_4, MVEVectorVTInfo VTI, SDNode Op, Intrinsic PredInt> {
1444  def "" : MVE_VMINMAXNM<iname, VTI.Suffix, VTI.Size, bit_4>;
1445
1446  let Predicates = [HasMVEFloat] in {
1447    defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? (i32 0)), !cast<Instruction>(NAME)>;
1448  }
1449}
1450
1451defm MVE_VMAXNMf32 : MVE_VMINMAXNM_m<"vmaxnm", 0b0, MVE_v4f32, fmaxnum, int_arm_mve_max_predicated>;
1452defm MVE_VMAXNMf16 : MVE_VMINMAXNM_m<"vmaxnm", 0b0, MVE_v8f16, fmaxnum, int_arm_mve_max_predicated>;
1453defm MVE_VMINNMf32 : MVE_VMINMAXNM_m<"vminnm", 0b1, MVE_v4f32, fminnum, int_arm_mve_min_predicated>;
1454defm MVE_VMINNMf16 : MVE_VMINMAXNM_m<"vminnm", 0b1, MVE_v8f16, fminnum, int_arm_mve_min_predicated>;
1455
1456
1457class MVE_VMINMAX<string iname, string suffix, bit U, bits<2> size,
1458              bit bit_4, list<dag> pattern=[]>
1459  : MVE_comp<NoItinerary, iname, suffix, "", size, pattern> {
1460
1461  let Inst{28} = U;
1462  let Inst{25-24} = 0b11;
1463  let Inst{23} = 0b0;
1464  let Inst{21-20} = size{1-0};
1465  let Inst{11} = 0b0;
1466  let Inst{8} = 0b0;
1467  let Inst{6} = 0b1;
1468  let Inst{4} = bit_4;
1469  let validForTailPredication = 1;
1470}
1471
1472multiclass MVE_VMINMAX_m<string iname, bit bit_4, MVEVectorVTInfo VTI,
1473                      SDNode Op, Intrinsic PredInt> {
1474  def "" : MVE_VMINMAX<iname, VTI.Suffix, VTI.Unsigned, VTI.Size, bit_4>;
1475
1476  let Predicates = [HasMVEInt] in {
1477    defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? (i32 VTI.Unsigned)), !cast<Instruction>(NAME)>;
1478  }
1479}
1480
1481multiclass MVE_VMAX<MVEVectorVTInfo VTI>
1482  : MVE_VMINMAX_m<"vmax", 0b0, VTI, !if(VTI.Unsigned, umax, smax), int_arm_mve_max_predicated>;
1483multiclass MVE_VMIN<MVEVectorVTInfo VTI>
1484  : MVE_VMINMAX_m<"vmin", 0b1, VTI, !if(VTI.Unsigned, umin, smin), int_arm_mve_min_predicated>;
1485
1486defm MVE_VMINs8   : MVE_VMIN<MVE_v16s8>;
1487defm MVE_VMINs16  : MVE_VMIN<MVE_v8s16>;
1488defm MVE_VMINs32  : MVE_VMIN<MVE_v4s32>;
1489defm MVE_VMINu8   : MVE_VMIN<MVE_v16u8>;
1490defm MVE_VMINu16  : MVE_VMIN<MVE_v8u16>;
1491defm MVE_VMINu32  : MVE_VMIN<MVE_v4u32>;
1492
1493defm MVE_VMAXs8   : MVE_VMAX<MVE_v16s8>;
1494defm MVE_VMAXs16  : MVE_VMAX<MVE_v8s16>;
1495defm MVE_VMAXs32  : MVE_VMAX<MVE_v4s32>;
1496defm MVE_VMAXu8   : MVE_VMAX<MVE_v16u8>;
1497defm MVE_VMAXu16  : MVE_VMAX<MVE_v8u16>;
1498defm MVE_VMAXu32  : MVE_VMAX<MVE_v4u32>;
1499
1500// end of mve_comp instructions
1501
1502// start of mve_bit instructions
1503
1504class MVE_bit_arith<dag oops, dag iops, string iname, string suffix,
1505                    string ops, string cstr, bits<2> vecsize, list<dag> pattern=[]>
1506  : MVE_p<oops, iops, NoItinerary, iname, suffix, ops, vpred_r, cstr, vecsize, pattern> {
1507  bits<4> Qd;
1508  bits<4> Qm;
1509
1510  let Inst{22} = Qd{3};
1511  let Inst{15-13} = Qd{2-0};
1512  let Inst{5} = Qm{3};
1513  let Inst{3-1} = Qm{2-0};
1514}
1515
1516def MVE_VBIC : MVE_bit_arith<(outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm),
1517                             "vbic", "", "$Qd, $Qn, $Qm", "", 0b00> {
1518  bits<4> Qn;
1519
1520  let Inst{28} = 0b0;
1521  let Inst{25-23} = 0b110;
1522  let Inst{21-20} = 0b01;
1523  let Inst{19-17} = Qn{2-0};
1524  let Inst{16} = 0b0;
1525  let Inst{12-8} = 0b00001;
1526  let Inst{7} = Qn{3};
1527  let Inst{6} = 0b1;
1528  let Inst{4} = 0b1;
1529  let Inst{0} = 0b0;
1530  let validForTailPredication = 1;
1531}
1532
1533class MVE_VREV<string iname, string suffix, bits<2> size, bits<2> bit_8_7,
1534               bits<2> vecsize, string cstr="">
1535  : MVE_bit_arith<(outs MQPR:$Qd), (ins MQPR:$Qm), iname,
1536                  suffix, "$Qd, $Qm", cstr, vecsize> {
1537
1538  let Inst{28} = 0b1;
1539  let Inst{25-23} = 0b111;
1540  let Inst{21-20} = 0b11;
1541  let Inst{19-18} = size;
1542  let Inst{17-16} = 0b00;
1543  let Inst{12-9} = 0b0000;
1544  let Inst{8-7} = bit_8_7;
1545  let Inst{6} = 0b1;
1546  let Inst{4} = 0b0;
1547  let Inst{0} = 0b0;
1548}
1549
1550def MVE_VREV64_8  : MVE_VREV<"vrev64", "8", 0b00, 0b00, 0b11, "@earlyclobber $Qd">;
1551def MVE_VREV64_16 : MVE_VREV<"vrev64", "16", 0b01, 0b00, 0b11, "@earlyclobber $Qd">;
1552def MVE_VREV64_32 : MVE_VREV<"vrev64", "32", 0b10, 0b00, 0b11, "@earlyclobber $Qd">;
1553
1554def MVE_VREV32_8  : MVE_VREV<"vrev32", "8", 0b00, 0b01, 0b10>;
1555def MVE_VREV32_16 : MVE_VREV<"vrev32", "16", 0b01, 0b01, 0b10>;
1556
1557def MVE_VREV16_8  : MVE_VREV<"vrev16", "8", 0b00, 0b10, 0b01>;
1558
1559let Predicates = [HasMVEInt] in {
1560  def : Pat<(v8i16 (bswap (v8i16 MQPR:$src))),
1561            (v8i16 (MVE_VREV16_8 (v8i16 MQPR:$src)))>;
1562  def : Pat<(v4i32 (bswap (v4i32 MQPR:$src))),
1563            (v4i32 (MVE_VREV32_8 (v4i32 MQPR:$src)))>;
1564}
1565
1566multiclass MVE_VREV_basic_patterns<int revbits, list<MVEVectorVTInfo> VTIs,
1567                                   Instruction Inst> {
1568  defvar unpred_op = !cast<SDNode>("ARMvrev" # revbits);
1569
1570  foreach VTI = VTIs in {
1571    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$src))),
1572              (VTI.Vec (Inst (VTI.Vec MQPR:$src)))>;
1573    def : Pat<(VTI.Vec (int_arm_mve_vrev_predicated (VTI.Vec MQPR:$src),
1574                  revbits, (VTI.Pred VCCR:$pred), (VTI.Vec MQPR:$inactive))),
1575              (VTI.Vec (Inst (VTI.Vec MQPR:$src), ARMVCCThen,
1576                  (VTI.Pred VCCR:$pred), zero_reg, (VTI.Vec MQPR:$inactive)))>;
1577  }
1578}
1579
1580let Predicates = [HasMVEInt] in {
1581  defm: MVE_VREV_basic_patterns<64, [MVE_v4i32, MVE_v4f32], MVE_VREV64_32>;
1582  defm: MVE_VREV_basic_patterns<64, [MVE_v8i16, MVE_v8f16], MVE_VREV64_16>;
1583  defm: MVE_VREV_basic_patterns<64, [MVE_v16i8           ], MVE_VREV64_8>;
1584
1585  defm: MVE_VREV_basic_patterns<32, [MVE_v8i16, MVE_v8f16], MVE_VREV32_16>;
1586  defm: MVE_VREV_basic_patterns<32, [MVE_v16i8           ], MVE_VREV32_8>;
1587
1588  defm: MVE_VREV_basic_patterns<16, [MVE_v16i8           ], MVE_VREV16_8>;
1589}
1590
1591def MVE_VMVN : MVE_bit_arith<(outs MQPR:$Qd), (ins MQPR:$Qm),
1592                             "vmvn", "", "$Qd, $Qm", "", 0b00> {
1593  let Inst{28} = 0b1;
1594  let Inst{25-23} = 0b111;
1595  let Inst{21-16} = 0b110000;
1596  let Inst{12-6} = 0b0010111;
1597  let Inst{4} = 0b0;
1598  let Inst{0} = 0b0;
1599  let validForTailPredication = 1;
1600}
1601
1602let Predicates = [HasMVEInt] in {
1603  foreach VTI = [ MVE_v16i8, MVE_v8i16, MVE_v4i32, MVE_v2i64 ] in {
1604    def : Pat<(VTI.Vec (vnotq    (VTI.Vec MQPR:$val1))),
1605              (VTI.Vec (MVE_VMVN (VTI.Vec MQPR:$val1)))>;
1606    def : Pat<(VTI.Vec (int_arm_mve_mvn_predicated (VTI.Vec MQPR:$val1),
1607                       (VTI.Pred VCCR:$pred), (VTI.Vec MQPR:$inactive))),
1608              (VTI.Vec (MVE_VMVN (VTI.Vec MQPR:$val1), ARMVCCThen,
1609                       (VTI.Pred VCCR:$pred), zero_reg, (VTI.Vec MQPR:$inactive)))>;
1610  }
1611}
1612
1613class MVE_bit_ops<string iname, bits<2> bit_21_20, bit bit_28>
1614  : MVE_bit_arith<(outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm),
1615                  iname, "", "$Qd, $Qn, $Qm", "", 0b00> {
1616  bits<4> Qn;
1617
1618  let Inst{28} = bit_28;
1619  let Inst{25-23} = 0b110;
1620  let Inst{21-20} = bit_21_20;
1621  let Inst{19-17} = Qn{2-0};
1622  let Inst{16} = 0b0;
1623  let Inst{12-8} = 0b00001;
1624  let Inst{7} = Qn{3};
1625  let Inst{6} = 0b1;
1626  let Inst{4} = 0b1;
1627  let Inst{0} = 0b0;
1628  let validForTailPredication = 1;
1629}
1630
1631def MVE_VEOR : MVE_bit_ops<"veor", 0b00, 0b1>;
1632def MVE_VORN : MVE_bit_ops<"vorn", 0b11, 0b0>;
1633def MVE_VORR : MVE_bit_ops<"vorr", 0b10, 0b0>;
1634def MVE_VAND : MVE_bit_ops<"vand", 0b00, 0b0>;
1635
1636// add ignored suffixes as aliases
1637
1638foreach s=["s8", "s16", "s32", "u8", "u16", "u32", "i8", "i16", "i32", "f16", "f32"] in {
1639  def : MVEInstAlias<"vbic${vp}." # s # "\t$QdSrc, $QnSrc, $QmSrc",
1640        (MVE_VBIC MQPR:$QdSrc, MQPR:$QnSrc, MQPR:$QmSrc, vpred_r:$vp)>;
1641  def : MVEInstAlias<"veor${vp}." # s # "\t$QdSrc, $QnSrc, $QmSrc",
1642        (MVE_VEOR MQPR:$QdSrc, MQPR:$QnSrc, MQPR:$QmSrc, vpred_r:$vp)>;
1643  def : MVEInstAlias<"vorn${vp}." # s # "\t$QdSrc, $QnSrc, $QmSrc",
1644        (MVE_VORN MQPR:$QdSrc, MQPR:$QnSrc, MQPR:$QmSrc, vpred_r:$vp)>;
1645  def : MVEInstAlias<"vorr${vp}." # s # "\t$QdSrc, $QnSrc, $QmSrc",
1646        (MVE_VORR MQPR:$QdSrc, MQPR:$QnSrc, MQPR:$QmSrc, vpred_r:$vp)>;
1647  def : MVEInstAlias<"vand${vp}." # s # "\t$QdSrc, $QnSrc, $QmSrc",
1648        (MVE_VAND MQPR:$QdSrc, MQPR:$QnSrc, MQPR:$QmSrc, vpred_r:$vp)>;
1649}
1650
1651let Predicates = [HasMVEInt] in {
1652  defm : MVE_TwoOpPattern<MVE_v16i8, and, int_arm_mve_and_predicated, (? ), MVE_VAND, ARMimmAllOnesV>;
1653  defm : MVE_TwoOpPattern<MVE_v8i16, and, int_arm_mve_and_predicated, (? ), MVE_VAND, ARMimmAllOnesV>;
1654  defm : MVE_TwoOpPattern<MVE_v4i32, and, int_arm_mve_and_predicated, (? ), MVE_VAND, ARMimmAllOnesV>;
1655  defm : MVE_TwoOpPattern<MVE_v2i64, and, int_arm_mve_and_predicated, (? ), MVE_VAND, ARMimmAllOnesV>;
1656
1657  defm : MVE_TwoOpPattern<MVE_v16i8, or, int_arm_mve_orr_predicated, (? ), MVE_VORR, ARMimmAllZerosV>;
1658  defm : MVE_TwoOpPattern<MVE_v8i16, or, int_arm_mve_orr_predicated, (? ), MVE_VORR, ARMimmAllZerosV>;
1659  defm : MVE_TwoOpPattern<MVE_v4i32, or, int_arm_mve_orr_predicated, (? ), MVE_VORR, ARMimmAllZerosV>;
1660  defm : MVE_TwoOpPattern<MVE_v2i64, or, int_arm_mve_orr_predicated, (? ), MVE_VORR, ARMimmAllZerosV>;
1661
1662  defm : MVE_TwoOpPattern<MVE_v16i8, xor, int_arm_mve_eor_predicated, (? ), MVE_VEOR, ARMimmAllZerosV>;
1663  defm : MVE_TwoOpPattern<MVE_v8i16, xor, int_arm_mve_eor_predicated, (? ), MVE_VEOR, ARMimmAllZerosV>;
1664  defm : MVE_TwoOpPattern<MVE_v4i32, xor, int_arm_mve_eor_predicated, (? ), MVE_VEOR, ARMimmAllZerosV>;
1665  defm : MVE_TwoOpPattern<MVE_v2i64, xor, int_arm_mve_eor_predicated, (? ), MVE_VEOR, ARMimmAllZerosV>;
1666
1667  defm : MVE_TwoOpPattern<MVE_v16i8, BinOpFrag<(and node:$LHS, (vnotq node:$RHS))>,
1668                          int_arm_mve_bic_predicated, (? ), MVE_VBIC>;
1669  defm : MVE_TwoOpPattern<MVE_v8i16, BinOpFrag<(and node:$LHS, (vnotq node:$RHS))>,
1670                          int_arm_mve_bic_predicated, (? ), MVE_VBIC>;
1671  defm : MVE_TwoOpPattern<MVE_v4i32, BinOpFrag<(and node:$LHS, (vnotq node:$RHS))>,
1672                          int_arm_mve_bic_predicated, (? ), MVE_VBIC>;
1673  defm : MVE_TwoOpPattern<MVE_v2i64, BinOpFrag<(and node:$LHS, (vnotq node:$RHS))>,
1674                          int_arm_mve_bic_predicated, (? ), MVE_VBIC>;
1675
1676  defm : MVE_TwoOpPattern<MVE_v16i8, BinOpFrag<(or node:$LHS, (vnotq node:$RHS))>,
1677                          int_arm_mve_orn_predicated, (? ), MVE_VORN>;
1678  defm : MVE_TwoOpPattern<MVE_v8i16, BinOpFrag<(or node:$LHS, (vnotq node:$RHS))>,
1679                          int_arm_mve_orn_predicated, (? ), MVE_VORN>;
1680  defm : MVE_TwoOpPattern<MVE_v4i32, BinOpFrag<(or node:$LHS, (vnotq node:$RHS))>,
1681                          int_arm_mve_orn_predicated, (? ), MVE_VORN>;
1682  defm : MVE_TwoOpPattern<MVE_v2i64, BinOpFrag<(or node:$LHS, (vnotq node:$RHS))>,
1683                          int_arm_mve_orn_predicated, (? ), MVE_VORN>;
1684}
1685
1686class MVE_bit_cmode<string iname, string suffix, bit halfword, dag inOps, bits<2> vecsize>
1687  : MVE_p<(outs MQPR:$Qd), inOps, NoItinerary,
1688          iname, suffix, "$Qd, $imm", vpred_n, "$Qd = $Qd_src", vecsize> {
1689  bits<12> imm;
1690  bits<4> Qd;
1691
1692  let Inst{28} = imm{7};
1693  let Inst{27-23} = 0b11111;
1694  let Inst{22} = Qd{3};
1695  let Inst{21-19} = 0b000;
1696  let Inst{18-16} = imm{6-4};
1697  let Inst{15-13} = Qd{2-0};
1698  let Inst{12} = 0b0;
1699  let Inst{11} = halfword;
1700  let Inst{10} = !if(halfword, 0, imm{10});
1701  let Inst{9} = imm{9};
1702  let Inst{8} = 0b1;
1703  let Inst{7-6} = 0b01;
1704  let Inst{4} = 0b1;
1705  let Inst{3-0} = imm{3-0};
1706}
1707
1708multiclass MVE_bit_cmode_p<string iname, bit opcode,
1709                           MVEVectorVTInfo VTI, Operand imm_type, SDNode op> {
1710  def "" : MVE_bit_cmode<iname, VTI.Suffix, VTI.Size{0},
1711                         (ins MQPR:$Qd_src, imm_type:$imm), VTI.Size> {
1712    let Inst{5} = opcode;
1713    let validForTailPredication = 1;
1714  }
1715
1716  defvar Inst = !cast<Instruction>(NAME);
1717  defvar UnpredPat = (VTI.Vec (op (VTI.Vec MQPR:$src), timm:$simm));
1718
1719  let Predicates = [HasMVEInt] in {
1720    def : Pat<UnpredPat,
1721              (VTI.Vec (Inst (VTI.Vec MQPR:$src), imm_type:$simm))>;
1722    def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
1723                          UnpredPat, (VTI.Vec MQPR:$src))),
1724              (VTI.Vec (Inst (VTI.Vec MQPR:$src), imm_type:$simm,
1725                             ARMVCCThen, (VTI.Pred VCCR:$pred), zero_reg))>;
1726  }
1727}
1728
1729multiclass MVE_VORRimm<MVEVectorVTInfo VTI, Operand imm_type> {
1730  defm "": MVE_bit_cmode_p<"vorr", 0, VTI, imm_type, ARMvorrImm>;
1731}
1732multiclass MVE_VBICimm<MVEVectorVTInfo VTI, Operand imm_type> {
1733  defm "": MVE_bit_cmode_p<"vbic", 1, VTI, imm_type, ARMvbicImm>;
1734}
1735
1736defm MVE_VORRimmi16 : MVE_VORRimm<MVE_v8i16, nImmSplatI16>;
1737defm MVE_VORRimmi32 : MVE_VORRimm<MVE_v4i32, nImmSplatI32>;
1738defm MVE_VBICimmi16 : MVE_VBICimm<MVE_v8i16, nImmSplatI16>;
1739defm MVE_VBICimmi32 : MVE_VBICimm<MVE_v4i32, nImmSplatI32>;
1740
1741def MVE_VORNimmi16 : MVEInstAlias<"vorn${vp}.i16\t$Qd, $imm",
1742    (MVE_VORRimmi16 MQPR:$Qd, nImmSplatNotI16:$imm, vpred_n:$vp), 0>;
1743def MVE_VORNimmi32 : MVEInstAlias<"vorn${vp}.i32\t$Qd, $imm",
1744    (MVE_VORRimmi32 MQPR:$Qd, nImmSplatNotI32:$imm, vpred_n:$vp), 0>;
1745
1746def MVE_VANDimmi16 : MVEInstAlias<"vand${vp}.i16\t$Qd, $imm",
1747    (MVE_VBICimmi16 MQPR:$Qd, nImmSplatNotI16:$imm, vpred_n:$vp), 0>;
1748def MVE_VANDimmi32 : MVEInstAlias<"vand${vp}.i32\t$Qd, $imm",
1749    (MVE_VBICimmi32 MQPR:$Qd, nImmSplatNotI32:$imm, vpred_n:$vp), 0>;
1750
1751def MVE_VMOV : MVEInstAlias<"vmov${vp}\t$Qd, $Qm",
1752    (MVE_VORR MQPR:$Qd, MQPR:$Qm, MQPR:$Qm, vpred_r:$vp)>;
1753
1754class MVE_VMOV_lane_direction {
1755  bit bit_20;
1756  dag oops;
1757  dag iops;
1758  string ops;
1759  string cstr;
1760}
1761def MVE_VMOV_from_lane : MVE_VMOV_lane_direction {
1762  let bit_20 = 0b1;
1763  let oops = (outs rGPR:$Rt);
1764  let iops = (ins MQPR:$Qd);
1765  let ops = "$Rt, $Qd$Idx";
1766  let cstr = "";
1767}
1768def MVE_VMOV_to_lane : MVE_VMOV_lane_direction {
1769  let bit_20 = 0b0;
1770  let oops = (outs MQPR:$Qd);
1771  let iops = (ins MQPR:$Qd_src, rGPR:$Rt);
1772  let ops = "$Qd$Idx, $Rt";
1773  let cstr = "$Qd = $Qd_src";
1774}
1775
1776class MVE_VMOV_lane<string suffix, bit U, dag indexop,
1777                    MVE_VMOV_lane_direction dir>
1778  : MVE_VMOV_lane_base<dir.oops, !con(dir.iops, indexop), NoItinerary,
1779                       "vmov", suffix, dir.ops, dir.cstr, []> {
1780  bits<4> Qd;
1781  bits<4> Rt;
1782
1783  let Inst{31-24} = 0b11101110;
1784  let Inst{23} = U;
1785  let Inst{20} = dir.bit_20;
1786  let Inst{19-17} = Qd{2-0};
1787  let Inst{15-12} = Rt{3-0};
1788  let Inst{11-8} = 0b1011;
1789  let Inst{7} = Qd{3};
1790  let Inst{4-0} = 0b10000;
1791
1792  let hasSideEffects = 0;
1793}
1794
1795class MVE_VMOV_lane_32<MVE_VMOV_lane_direction dir>
1796    : MVE_VMOV_lane<"32", 0b0, (ins MVEVectorIndex<4>:$Idx), dir> {
1797  bits<2> Idx;
1798  let Inst{22} = 0b0;
1799  let Inst{6-5} = 0b00;
1800  let Inst{16} = Idx{1};
1801  let Inst{21} = Idx{0};
1802
1803  let VecSize = 0b10;
1804  let Predicates = [HasFPRegsV8_1M];
1805}
1806
1807class MVE_VMOV_lane_16<string suffix, bit U, MVE_VMOV_lane_direction dir>
1808  : MVE_VMOV_lane<suffix, U, (ins MVEVectorIndex<8>:$Idx), dir> {
1809  bits<3> Idx;
1810  let Inst{22} = 0b0;
1811  let Inst{5} = 0b1;
1812  let Inst{16} = Idx{2};
1813  let Inst{21} = Idx{1};
1814  let Inst{6} = Idx{0};
1815
1816  let VecSize = 0b01;
1817}
1818
1819class MVE_VMOV_lane_8<string suffix, bit U, MVE_VMOV_lane_direction dir>
1820  : MVE_VMOV_lane<suffix, U, (ins MVEVectorIndex<16>:$Idx), dir> {
1821  bits<4> Idx;
1822  let Inst{22} = 0b1;
1823  let Inst{16} = Idx{3};
1824  let Inst{21} = Idx{2};
1825  let Inst{6} = Idx{1};
1826  let Inst{5} = Idx{0};
1827
1828  let VecSize = 0b00;
1829}
1830
1831def MVE_VMOV_from_lane_32  : MVE_VMOV_lane_32<            MVE_VMOV_from_lane>;
1832def MVE_VMOV_from_lane_s16 : MVE_VMOV_lane_16<"s16", 0b0, MVE_VMOV_from_lane>;
1833def MVE_VMOV_from_lane_u16 : MVE_VMOV_lane_16<"u16", 0b1, MVE_VMOV_from_lane>;
1834def MVE_VMOV_from_lane_s8  : MVE_VMOV_lane_8 < "s8", 0b0, MVE_VMOV_from_lane>;
1835def MVE_VMOV_from_lane_u8  : MVE_VMOV_lane_8 < "u8", 0b1, MVE_VMOV_from_lane>;
1836let isInsertSubreg = 1 in
1837def MVE_VMOV_to_lane_32    : MVE_VMOV_lane_32<            MVE_VMOV_to_lane>;
1838def MVE_VMOV_to_lane_16    : MVE_VMOV_lane_16< "16", 0b0, MVE_VMOV_to_lane>;
1839def MVE_VMOV_to_lane_8     : MVE_VMOV_lane_8 <  "8", 0b0, MVE_VMOV_to_lane>;
1840
1841// This is the same as insertelt but allows the inserted value to be an i32 as
1842// will be used when it is the only legal type.
1843def ARMVecInsert : SDTypeProfile<1, 3, [
1844  SDTCisVT<2, i32>, SDTCisSameAs<0, 1>, SDTCisPtrTy<3>
1845]>;
1846def ARMinsertelt  : SDNode<"ISD::INSERT_VECTOR_ELT", ARMVecInsert>;
1847
1848let Predicates = [HasMVEInt] in {
1849  def : Pat<(extractelt (v2f64 MQPR:$src), imm:$lane),
1850            (f64 (EXTRACT_SUBREG MQPR:$src, (DSubReg_f64_reg imm:$lane)))>;
1851  def : Pat<(insertelt (v2f64 MQPR:$src1), DPR:$src2, imm:$lane),
1852            (INSERT_SUBREG (v2f64 (COPY_TO_REGCLASS MQPR:$src1, MQPR)), DPR:$src2, (DSubReg_f64_reg imm:$lane))>;
1853
1854  def : Pat<(extractelt (v4i32 MQPR:$src), imm:$lane),
1855            (COPY_TO_REGCLASS
1856              (i32 (EXTRACT_SUBREG MQPR:$src, (SSubReg_f32_reg imm:$lane))), rGPR)>;
1857  def : Pat<(insertelt (v4i32 MQPR:$src1), rGPR:$src2, imm:$lane),
1858            (MVE_VMOV_to_lane_32 MQPR:$src1, rGPR:$src2, imm:$lane)>;
1859  // This tries to copy from one lane to another, without going via GPR regs
1860  def : Pat<(insertelt (v4i32 MQPR:$src1), (extractelt (v4i32 MQPR:$src2), imm:$extlane), imm:$inslane),
1861            (v4i32 (COPY_TO_REGCLASS
1862                     (INSERT_SUBREG (v4f32 (COPY_TO_REGCLASS (v4i32 MQPR:$src1), MQPR)),
1863                                    (f32 (EXTRACT_SUBREG (v4f32 (COPY_TO_REGCLASS (v4i32 MQPR:$src2), MQPR)),
1864                                                         (SSubReg_f32_reg imm:$extlane))),
1865                                    (SSubReg_f32_reg imm:$inslane)),
1866                      MQPR))>;
1867
1868  def : Pat<(vector_insert (v16i8 MQPR:$src1), rGPR:$src2, imm:$lane),
1869            (MVE_VMOV_to_lane_8  MQPR:$src1, rGPR:$src2, imm:$lane)>;
1870  def : Pat<(vector_insert (v8i16 MQPR:$src1), rGPR:$src2, imm:$lane),
1871            (MVE_VMOV_to_lane_16 MQPR:$src1, rGPR:$src2, imm:$lane)>;
1872
1873  def : Pat<(ARMvgetlanes (v16i8 MQPR:$src), imm:$lane),
1874            (MVE_VMOV_from_lane_s8 MQPR:$src, imm:$lane)>;
1875  def : Pat<(ARMvgetlanes (v8i16 MQPR:$src), imm:$lane),
1876            (MVE_VMOV_from_lane_s16 MQPR:$src, imm:$lane)>;
1877  def : Pat<(ARMvgetlanes (v8f16 MQPR:$src), imm:$lane),
1878            (MVE_VMOV_from_lane_s16 MQPR:$src, imm:$lane)>;
1879  def : Pat<(ARMvgetlaneu (v16i8 MQPR:$src), imm:$lane),
1880            (MVE_VMOV_from_lane_u8 MQPR:$src, imm:$lane)>;
1881  def : Pat<(ARMvgetlaneu (v8i16 MQPR:$src), imm:$lane),
1882            (MVE_VMOV_from_lane_u16 MQPR:$src, imm:$lane)>;
1883  def : Pat<(ARMvgetlaneu (v8f16 MQPR:$src), imm:$lane),
1884            (MVE_VMOV_from_lane_u16 MQPR:$src, imm:$lane)>;
1885  // For i16's inserts being extracted from low lanes, then may use VINS.
1886  let Predicates = [HasFullFP16] in {
1887  def : Pat<(ARMinsertelt (v8i16 MQPR:$src1),
1888                          (ARMvgetlaneu (v8i16 MQPR:$src2), imm_even:$extlane),
1889                          imm_odd:$inslane),
1890            (COPY_TO_REGCLASS (INSERT_SUBREG (v4f32 (COPY_TO_REGCLASS MQPR:$src1, MQPR)),
1891                                (VINSH (EXTRACT_SUBREG MQPR:$src1, (SSubReg_f16_reg imm_odd:$inslane)),
1892                                       (EXTRACT_SUBREG MQPR:$src2, (SSubReg_f16_reg imm_even:$extlane))),
1893                                (SSubReg_f16_reg imm_odd:$inslane)), MQPR)>;
1894  }
1895
1896  def : Pat<(v16i8 (scalar_to_vector GPR:$src)),
1897            (MVE_VMOV_to_lane_8  (v16i8 (IMPLICIT_DEF)), rGPR:$src, (i32 0))>;
1898  def : Pat<(v8i16 (scalar_to_vector GPR:$src)),
1899            (MVE_VMOV_to_lane_16 (v8i16 (IMPLICIT_DEF)), rGPR:$src, (i32 0))>;
1900  def : Pat<(v4i32 (scalar_to_vector GPR:$src)),
1901            (MVE_VMOV_to_lane_32 (v4i32 (IMPLICIT_DEF)), rGPR:$src, (i32 0))>;
1902
1903  // Floating point patterns, still enabled under HasMVEInt
1904  def : Pat<(extractelt (v4f32 MQPR:$src), imm:$lane),
1905            (COPY_TO_REGCLASS (f32 (EXTRACT_SUBREG MQPR:$src, (SSubReg_f32_reg imm:$lane))), SPR)>;
1906  def : Pat<(insertelt (v4f32 MQPR:$src1), (f32 SPR:$src2), imm:$lane),
1907            (INSERT_SUBREG (v4f32 (COPY_TO_REGCLASS MQPR:$src1, MQPR)), SPR:$src2, (SSubReg_f32_reg imm:$lane))>;
1908
1909  def : Pat<(insertelt (v8f16 MQPR:$src1), (f16 HPR:$src2), imm_even:$lane),
1910            (MVE_VMOV_to_lane_16 MQPR:$src1, (COPY_TO_REGCLASS (f16 HPR:$src2), rGPR), imm:$lane)>;
1911  let Predicates = [HasFullFP16] in {
1912  def : Pat<(insertelt (v8f16 MQPR:$src1), (f16 HPR:$src2), imm_odd:$lane),
1913            (COPY_TO_REGCLASS (INSERT_SUBREG (v4f32 (COPY_TO_REGCLASS MQPR:$src1, MQPR)),
1914                                (VINSH (EXTRACT_SUBREG MQPR:$src1, (SSubReg_f16_reg imm_odd:$lane)),
1915                                       (COPY_TO_REGCLASS HPR:$src2, SPR)),
1916                                (SSubReg_f16_reg imm_odd:$lane)), MQPR)>;
1917  }
1918  def : Pat<(extractelt (v8f16 MQPR:$src), imm_even:$lane),
1919            (EXTRACT_SUBREG MQPR:$src, (SSubReg_f16_reg imm_even:$lane))>;
1920  let Predicates = [HasFullFP16] in {
1921  def : Pat<(extractelt (v8f16 MQPR:$src), imm_odd:$lane),
1922            (COPY_TO_REGCLASS
1923              (VMOVH (EXTRACT_SUBREG MQPR:$src, (SSubReg_f16_reg imm_odd:$lane))),
1924              HPR)>;
1925  }
1926
1927  def : Pat<(v2f64 (scalar_to_vector (f64 DPR:$src))),
1928            (INSERT_SUBREG (v2f64 (IMPLICIT_DEF)), DPR:$src, dsub_0)>;
1929  def : Pat<(v4f32 (scalar_to_vector SPR:$src)),
1930            (INSERT_SUBREG (v4f32 (IMPLICIT_DEF)), SPR:$src, ssub_0)>;
1931  def : Pat<(v4f32 (scalar_to_vector GPR:$src)),
1932            (MVE_VMOV_to_lane_32 (v4f32 (IMPLICIT_DEF)), rGPR:$src, (i32 0))>;
1933  def : Pat<(v8f16 (scalar_to_vector (f16 HPR:$src))),
1934            (INSERT_SUBREG (v8f16 (IMPLICIT_DEF)), (f16 HPR:$src), ssub_0)>;
1935  def : Pat<(v8f16 (scalar_to_vector GPR:$src)),
1936            (MVE_VMOV_to_lane_16 (v8f16 (IMPLICIT_DEF)), rGPR:$src, (i32 0))>;
1937}
1938
1939// end of mve_bit instructions
1940
1941// start of MVE Integer instructions
1942
1943class MVE_int<string iname, string suffix, bits<2> size, list<dag> pattern=[]>
1944  : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm), NoItinerary,
1945          iname, suffix, "$Qd, $Qn, $Qm", vpred_r, "", size, pattern> {
1946  bits<4> Qd;
1947  bits<4> Qn;
1948  bits<4> Qm;
1949
1950  let Inst{22} = Qd{3};
1951  let Inst{21-20} = size;
1952  let Inst{19-17} = Qn{2-0};
1953  let Inst{15-13} = Qd{2-0};
1954  let Inst{7} = Qn{3};
1955  let Inst{6} = 0b1;
1956  let Inst{5} = Qm{3};
1957  let Inst{3-1} = Qm{2-0};
1958}
1959
1960class MVE_VMULt1<string iname, string suffix, bits<2> size,
1961                   list<dag> pattern=[]>
1962  : MVE_int<iname, suffix, size, pattern> {
1963
1964  let Inst{28} = 0b0;
1965  let Inst{25-23} = 0b110;
1966  let Inst{16} = 0b0;
1967  let Inst{12-8} = 0b01001;
1968  let Inst{4} = 0b1;
1969  let Inst{0} = 0b0;
1970  let validForTailPredication = 1;
1971}
1972
1973multiclass MVE_VMUL_m<MVEVectorVTInfo VTI> {
1974  def "" : MVE_VMULt1<"vmul", VTI.Suffix, VTI.Size>;
1975
1976  let Predicates = [HasMVEInt] in {
1977    defm : MVE_TwoOpPattern<VTI, mul, int_arm_mve_mul_predicated, (? ),
1978                            !cast<Instruction>(NAME), ARMimmOneV>;
1979  }
1980}
1981
1982defm MVE_VMULi8  : MVE_VMUL_m<MVE_v16i8>;
1983defm MVE_VMULi16 : MVE_VMUL_m<MVE_v8i16>;
1984defm MVE_VMULi32 : MVE_VMUL_m<MVE_v4i32>;
1985
1986class MVE_VQxDMULH_Base<string iname, string suffix, bits<2> size, bit rounding,
1987                  list<dag> pattern=[]>
1988  : MVE_int<iname, suffix, size, pattern> {
1989
1990  let Inst{28} = rounding;
1991  let Inst{25-23} = 0b110;
1992  let Inst{16} = 0b0;
1993  let Inst{12-8} = 0b01011;
1994  let Inst{4} = 0b0;
1995  let Inst{0} = 0b0;
1996  let validForTailPredication = 1;
1997}
1998
1999def MVEvqdmulh : SDNode<"ARMISD::VQDMULH", SDTIntBinOp>;
2000
2001multiclass MVE_VQxDMULH_m<string iname, MVEVectorVTInfo VTI,
2002                      SDPatternOperator Op, Intrinsic unpred_int, Intrinsic pred_int,
2003                      bit rounding> {
2004  def "" : MVE_VQxDMULH_Base<iname, VTI.Suffix, VTI.Size, rounding>;
2005  defvar Inst = !cast<Instruction>(NAME);
2006
2007  let Predicates = [HasMVEInt] in {
2008    defm : MVE_TwoOpPattern<VTI, Op, pred_int, (? ), Inst>;
2009
2010    // Extra unpredicated multiply intrinsic patterns
2011    def : Pat<(VTI.Vec (unpred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn))),
2012              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
2013  }
2014}
2015
2016multiclass MVE_VQxDMULH<string iname, MVEVectorVTInfo VTI, bit rounding>
2017  : MVE_VQxDMULH_m<iname, VTI, !if(rounding, null_frag,
2018                                             MVEvqdmulh),
2019                               !if(rounding, int_arm_mve_vqrdmulh,
2020                                             int_arm_mve_vqdmulh),
2021                               !if(rounding, int_arm_mve_qrdmulh_predicated,
2022                                             int_arm_mve_qdmulh_predicated),
2023                   rounding>;
2024
2025defm MVE_VQDMULHi8  : MVE_VQxDMULH<"vqdmulh", MVE_v16s8, 0b0>;
2026defm MVE_VQDMULHi16 : MVE_VQxDMULH<"vqdmulh", MVE_v8s16, 0b0>;
2027defm MVE_VQDMULHi32 : MVE_VQxDMULH<"vqdmulh", MVE_v4s32, 0b0>;
2028
2029defm MVE_VQRDMULHi8  : MVE_VQxDMULH<"vqrdmulh", MVE_v16s8, 0b1>;
2030defm MVE_VQRDMULHi16 : MVE_VQxDMULH<"vqrdmulh", MVE_v8s16, 0b1>;
2031defm MVE_VQRDMULHi32 : MVE_VQxDMULH<"vqrdmulh", MVE_v4s32, 0b1>;
2032
2033class MVE_VADDSUB<string iname, string suffix, bits<2> size, bit subtract,
2034                    list<dag> pattern=[]>
2035  : MVE_int<iname, suffix, size, pattern> {
2036
2037  let Inst{28} = subtract;
2038  let Inst{25-23} = 0b110;
2039  let Inst{16} = 0b0;
2040  let Inst{12-8} = 0b01000;
2041  let Inst{4} = 0b0;
2042  let Inst{0} = 0b0;
2043  let validForTailPredication = 1;
2044}
2045
2046multiclass MVE_VADDSUB_m<string iname, MVEVectorVTInfo VTI, bit subtract,
2047                         SDNode Op, Intrinsic PredInt> {
2048  def "" : MVE_VADDSUB<iname, VTI.Suffix, VTI.Size, subtract>;
2049  defvar Inst = !cast<Instruction>(NAME);
2050
2051  let Predicates = [HasMVEInt] in {
2052    defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? ), !cast<Instruction>(NAME), ARMimmAllZerosV>;
2053  }
2054}
2055
2056multiclass MVE_VADD<MVEVectorVTInfo VTI>
2057  : MVE_VADDSUB_m<"vadd", VTI, 0b0, add, int_arm_mve_add_predicated>;
2058multiclass MVE_VSUB<MVEVectorVTInfo VTI>
2059  : MVE_VADDSUB_m<"vsub", VTI, 0b1, sub, int_arm_mve_sub_predicated>;
2060
2061defm MVE_VADDi8  : MVE_VADD<MVE_v16i8>;
2062defm MVE_VADDi16 : MVE_VADD<MVE_v8i16>;
2063defm MVE_VADDi32 : MVE_VADD<MVE_v4i32>;
2064
2065defm MVE_VSUBi8  : MVE_VSUB<MVE_v16i8>;
2066defm MVE_VSUBi16 : MVE_VSUB<MVE_v8i16>;
2067defm MVE_VSUBi32 : MVE_VSUB<MVE_v4i32>;
2068
2069class MVE_VQADDSUB<string iname, string suffix, bit U, bit subtract,
2070                   bits<2> size>
2071  : MVE_int<iname, suffix, size, []> {
2072
2073  let Inst{28} = U;
2074  let Inst{25-23} = 0b110;
2075  let Inst{16} = 0b0;
2076  let Inst{12-10} = 0b000;
2077  let Inst{9} = subtract;
2078  let Inst{8} = 0b0;
2079  let Inst{4} = 0b1;
2080  let Inst{0} = 0b0;
2081  let validForTailPredication = 1;
2082}
2083
2084class MVE_VQADD_<string suffix, bit U, bits<2> size>
2085  : MVE_VQADDSUB<"vqadd", suffix, U, 0b0, size>;
2086class MVE_VQSUB_<string suffix, bit U, bits<2> size>
2087  : MVE_VQADDSUB<"vqsub", suffix, U, 0b1, size>;
2088
2089multiclass MVE_VQADD_m<MVEVectorVTInfo VTI,
2090                      SDNode Op, Intrinsic PredInt> {
2091  def "" : MVE_VQADD_<VTI.Suffix, VTI.Unsigned, VTI.Size>;
2092  defvar Inst = !cast<Instruction>(NAME);
2093
2094  let Predicates = [HasMVEInt] in {
2095    defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? (i32 VTI.Unsigned)),
2096                            !cast<Instruction>(NAME)>;
2097  }
2098}
2099
2100multiclass MVE_VQADD<MVEVectorVTInfo VTI, SDNode unpred_op>
2101  : MVE_VQADD_m<VTI, unpred_op, int_arm_mve_qadd_predicated>;
2102
2103defm MVE_VQADDs8  : MVE_VQADD<MVE_v16s8, saddsat>;
2104defm MVE_VQADDs16 : MVE_VQADD<MVE_v8s16, saddsat>;
2105defm MVE_VQADDs32 : MVE_VQADD<MVE_v4s32, saddsat>;
2106defm MVE_VQADDu8  : MVE_VQADD<MVE_v16u8, uaddsat>;
2107defm MVE_VQADDu16 : MVE_VQADD<MVE_v8u16, uaddsat>;
2108defm MVE_VQADDu32 : MVE_VQADD<MVE_v4u32, uaddsat>;
2109
2110multiclass MVE_VQSUB_m<MVEVectorVTInfo VTI,
2111                      SDNode Op, Intrinsic PredInt> {
2112  def "" : MVE_VQSUB_<VTI.Suffix, VTI.Unsigned, VTI.Size>;
2113  defvar Inst = !cast<Instruction>(NAME);
2114
2115  let Predicates = [HasMVEInt] in {
2116    defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? (i32 VTI.Unsigned)),
2117                            !cast<Instruction>(NAME)>;
2118  }
2119}
2120
2121multiclass MVE_VQSUB<MVEVectorVTInfo VTI, SDNode unpred_op>
2122  : MVE_VQSUB_m<VTI, unpred_op, int_arm_mve_qsub_predicated>;
2123
2124defm MVE_VQSUBs8  : MVE_VQSUB<MVE_v16s8, ssubsat>;
2125defm MVE_VQSUBs16 : MVE_VQSUB<MVE_v8s16, ssubsat>;
2126defm MVE_VQSUBs32 : MVE_VQSUB<MVE_v4s32, ssubsat>;
2127defm MVE_VQSUBu8  : MVE_VQSUB<MVE_v16u8, usubsat>;
2128defm MVE_VQSUBu16 : MVE_VQSUB<MVE_v8u16, usubsat>;
2129defm MVE_VQSUBu32 : MVE_VQSUB<MVE_v4u32, usubsat>;
2130
2131class MVE_VABD_int<string suffix, bit U, bits<2> size,
2132                     list<dag> pattern=[]>
2133  : MVE_int<"vabd", suffix, size, pattern> {
2134
2135  let Inst{28} = U;
2136  let Inst{25-23} = 0b110;
2137  let Inst{16} = 0b0;
2138  let Inst{12-8} = 0b00111;
2139  let Inst{4} = 0b0;
2140  let Inst{0} = 0b0;
2141  let validForTailPredication = 1;
2142}
2143
2144multiclass MVE_VABD_m<MVEVectorVTInfo VTI, SDNode Op,
2145                      Intrinsic unpred_int, Intrinsic PredInt> {
2146  def "" : MVE_VABD_int<VTI.Suffix, VTI.Unsigned, VTI.Size>;
2147  defvar Inst = !cast<Instruction>(NAME);
2148
2149  let Predicates = [HasMVEInt] in {
2150    defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? (i32 VTI.Unsigned)),
2151                            !cast<Instruction>(NAME)>;
2152
2153    // Unpredicated absolute difference
2154    def : Pat<(VTI.Vec (unpred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
2155                            (i32 VTI.Unsigned))),
2156              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
2157  }
2158}
2159
2160multiclass MVE_VABD<MVEVectorVTInfo VTI, SDNode Op>
2161  : MVE_VABD_m<VTI, Op, int_arm_mve_vabd, int_arm_mve_abd_predicated>;
2162
2163defm MVE_VABDs8  : MVE_VABD<MVE_v16s8, abds>;
2164defm MVE_VABDs16 : MVE_VABD<MVE_v8s16, abds>;
2165defm MVE_VABDs32 : MVE_VABD<MVE_v4s32, abds>;
2166defm MVE_VABDu8  : MVE_VABD<MVE_v16u8, abdu>;
2167defm MVE_VABDu16 : MVE_VABD<MVE_v8u16, abdu>;
2168defm MVE_VABDu32 : MVE_VABD<MVE_v4u32, abdu>;
2169
2170class MVE_VRHADD_Base<string suffix, bit U, bits<2> size, list<dag> pattern=[]>
2171  : MVE_int<"vrhadd", suffix, size, pattern> {
2172
2173  let Inst{28} = U;
2174  let Inst{25-23} = 0b110;
2175  let Inst{16} = 0b0;
2176  let Inst{12-8} = 0b00001;
2177  let Inst{4} = 0b0;
2178  let Inst{0} = 0b0;
2179  let validForTailPredication = 1;
2180}
2181
2182def addnuw : PatFrag<(ops node:$lhs, node:$rhs),
2183                     (add node:$lhs, node:$rhs), [{
2184  return N->getFlags().hasNoUnsignedWrap();
2185}]>;
2186
2187def addnsw : PatFrag<(ops node:$lhs, node:$rhs),
2188                     (add node:$lhs, node:$rhs), [{
2189  return N->getFlags().hasNoSignedWrap();
2190}]>;
2191
2192def subnuw : PatFrag<(ops node:$lhs, node:$rhs),
2193                     (sub node:$lhs, node:$rhs), [{
2194  return N->getFlags().hasNoUnsignedWrap();
2195}]>;
2196
2197def subnsw : PatFrag<(ops node:$lhs, node:$rhs),
2198                     (sub node:$lhs, node:$rhs), [{
2199  return N->getFlags().hasNoSignedWrap();
2200}]>;
2201
2202multiclass MVE_VRHADD_m<MVEVectorVTInfo VTI, SDNode Op,
2203                      SDPatternOperator unpred_op, Intrinsic PredInt> {
2204  def "" : MVE_VRHADD_Base<VTI.Suffix, VTI.Unsigned, VTI.Size>;
2205  defvar Inst = !cast<Instruction>(NAME);
2206  defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? (i32 VTI.Unsigned)), !cast<Instruction>(NAME)>;
2207
2208  let Predicates = [HasMVEInt] in {
2209    // Unpredicated rounding add-with-divide-by-two intrinsic
2210    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
2211                            (i32 VTI.Unsigned))),
2212              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
2213  }
2214}
2215
2216multiclass MVE_VRHADD<MVEVectorVTInfo VTI, SDNode rhadd>
2217  : MVE_VRHADD_m<VTI, rhadd, int_arm_mve_vrhadd, int_arm_mve_rhadd_predicated>;
2218
2219defm MVE_VRHADDs8  : MVE_VRHADD<MVE_v16s8, avgceils>;
2220defm MVE_VRHADDs16 : MVE_VRHADD<MVE_v8s16, avgceils>;
2221defm MVE_VRHADDs32 : MVE_VRHADD<MVE_v4s32, avgceils>;
2222defm MVE_VRHADDu8  : MVE_VRHADD<MVE_v16u8, avgceilu>;
2223defm MVE_VRHADDu16 : MVE_VRHADD<MVE_v8u16, avgceilu>;
2224defm MVE_VRHADDu32 : MVE_VRHADD<MVE_v4u32, avgceilu>;
2225
2226class MVE_VHADDSUB<string iname, string suffix, bit U, bit subtract,
2227                   bits<2> size, list<dag> pattern=[]>
2228  : MVE_int<iname, suffix, size, pattern> {
2229
2230  let Inst{28} = U;
2231  let Inst{25-23} = 0b110;
2232  let Inst{16} = 0b0;
2233  let Inst{12-10} = 0b000;
2234  let Inst{9} = subtract;
2235  let Inst{8} = 0b0;
2236  let Inst{4} = 0b0;
2237  let Inst{0} = 0b0;
2238  let validForTailPredication = 1;
2239}
2240
2241class MVE_VHADD_<string suffix, bit U, bits<2> size,
2242              list<dag> pattern=[]>
2243  : MVE_VHADDSUB<"vhadd", suffix, U, 0b0, size, pattern>;
2244class MVE_VHSUB_<string suffix, bit U, bits<2> size,
2245              list<dag> pattern=[]>
2246  : MVE_VHADDSUB<"vhsub", suffix, U, 0b1, size, pattern>;
2247
2248multiclass MVE_VHADD_m<MVEVectorVTInfo VTI, SDNode Op,
2249                      SDPatternOperator unpred_op, Intrinsic PredInt> {
2250  def "" : MVE_VHADD_<VTI.Suffix, VTI.Unsigned, VTI.Size>;
2251  defvar Inst = !cast<Instruction>(NAME);
2252  defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? (i32 VTI.Unsigned)), !cast<Instruction>(NAME)>;
2253
2254  let Predicates = [HasMVEInt] in {
2255    // Unpredicated add-and-divide-by-two
2256    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn), (i32 VTI.Unsigned))),
2257              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
2258  }
2259}
2260
2261multiclass MVE_VHADD<MVEVectorVTInfo VTI, SDNode Op>
2262  : MVE_VHADD_m<VTI, Op, int_arm_mve_vhadd, int_arm_mve_hadd_predicated>;
2263
2264defm MVE_VHADDs8  : MVE_VHADD<MVE_v16s8, avgfloors>;
2265defm MVE_VHADDs16 : MVE_VHADD<MVE_v8s16, avgfloors>;
2266defm MVE_VHADDs32 : MVE_VHADD<MVE_v4s32, avgfloors>;
2267defm MVE_VHADDu8  : MVE_VHADD<MVE_v16u8, avgflooru>;
2268defm MVE_VHADDu16 : MVE_VHADD<MVE_v8u16, avgflooru>;
2269defm MVE_VHADDu32 : MVE_VHADD<MVE_v4u32, avgflooru>;
2270
2271multiclass MVE_VHSUB_m<MVEVectorVTInfo VTI,
2272                      SDPatternOperator unpred_op, Intrinsic pred_int, PatFrag sub_op,
2273                      SDNode shift_op> {
2274  def "" : MVE_VHSUB_<VTI.Suffix, VTI.Unsigned, VTI.Size>;
2275  defvar Inst = !cast<Instruction>(NAME);
2276
2277  let Predicates = [HasMVEInt] in {
2278    // Unpredicated subtract-and-divide-by-two
2279    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
2280                            (i32 VTI.Unsigned))),
2281              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
2282
2283    def : Pat<(VTI.Vec (shift_op (sub_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)), (i32 1))),
2284              (Inst MQPR:$Qm, MQPR:$Qn)>;
2285
2286
2287    // Predicated subtract-and-divide-by-two
2288    def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
2289                            (i32 VTI.Unsigned), (VTI.Pred VCCR:$mask),
2290                            (VTI.Vec MQPR:$inactive))),
2291              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
2292                             ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg,
2293                             (VTI.Vec MQPR:$inactive)))>;
2294  }
2295}
2296
2297multiclass MVE_VHSUB<MVEVectorVTInfo VTI, PatFrag sub_op, SDNode shift_op>
2298  : MVE_VHSUB_m<VTI, int_arm_mve_vhsub, int_arm_mve_hsub_predicated, sub_op,
2299                shift_op>;
2300
2301defm MVE_VHSUBs8  : MVE_VHSUB<MVE_v16s8, subnsw, ARMvshrsImm>;
2302defm MVE_VHSUBs16 : MVE_VHSUB<MVE_v8s16, subnsw, ARMvshrsImm>;
2303defm MVE_VHSUBs32 : MVE_VHSUB<MVE_v4s32, subnsw, ARMvshrsImm>;
2304defm MVE_VHSUBu8  : MVE_VHSUB<MVE_v16u8, subnuw, ARMvshruImm>;
2305defm MVE_VHSUBu16 : MVE_VHSUB<MVE_v8u16, subnuw, ARMvshruImm>;
2306defm MVE_VHSUBu32 : MVE_VHSUB<MVE_v4u32, subnuw, ARMvshruImm>;
2307
2308class MVE_VDUP<string suffix, bit B, bit E, bits<2> vecsize, list<dag> pattern=[]>
2309  : MVE_p<(outs MQPR:$Qd), (ins rGPR:$Rt), NoItinerary,
2310          "vdup", suffix, "$Qd, $Rt", vpred_r, "", vecsize, pattern> {
2311  bits<4> Qd;
2312  bits<4> Rt;
2313
2314  let Inst{28} = 0b0;
2315  let Inst{25-23} = 0b101;
2316  let Inst{22} = B;
2317  let Inst{21-20} = 0b10;
2318  let Inst{19-17} = Qd{2-0};
2319  let Inst{16} = 0b0;
2320  let Inst{15-12} = Rt;
2321  let Inst{11-8} = 0b1011;
2322  let Inst{7} = Qd{3};
2323  let Inst{6} = 0b0;
2324  let Inst{5} = E;
2325  let Inst{4-0} = 0b10000;
2326  let validForTailPredication = 1;
2327}
2328
2329def MVE_VDUP32 : MVE_VDUP<"32", 0b0, 0b0, 0b10>;
2330def MVE_VDUP16 : MVE_VDUP<"16", 0b0, 0b1, 0b01>;
2331def MVE_VDUP8  : MVE_VDUP<"8",  0b1, 0b0, 0b00>;
2332
2333let Predicates = [HasMVEInt] in {
2334  def : Pat<(v16i8 (ARMvdup (i32 rGPR:$elem))),
2335            (MVE_VDUP8  rGPR:$elem)>;
2336  def : Pat<(v8i16 (ARMvdup (i32 rGPR:$elem))),
2337            (MVE_VDUP16 rGPR:$elem)>;
2338  def : Pat<(v4i32 (ARMvdup (i32 rGPR:$elem))),
2339            (MVE_VDUP32 rGPR:$elem)>;
2340
2341  def : Pat<(v8f16 (ARMvdup (i32 rGPR:$elem))),
2342            (MVE_VDUP16 rGPR:$elem)>;
2343  def : Pat<(v4f32 (ARMvdup (i32 rGPR:$elem))),
2344            (MVE_VDUP32 rGPR:$elem)>;
2345
2346  // Match a vselect with an ARMvdup as a predicated MVE_VDUP
2347  def : Pat<(v16i8 (vselect (v16i1 VCCR:$pred),
2348                            (v16i8 (ARMvdup (i32 rGPR:$elem))),
2349                            (v16i8 MQPR:$inactive))),
2350            (MVE_VDUP8  rGPR:$elem, ARMVCCThen, (v16i1 VCCR:$pred), zero_reg,
2351                        (v16i8 MQPR:$inactive))>;
2352  def : Pat<(v8i16 (vselect (v8i1 VCCR:$pred),
2353                            (v8i16 (ARMvdup (i32 rGPR:$elem))),
2354                            (v8i16 MQPR:$inactive))),
2355            (MVE_VDUP16 rGPR:$elem, ARMVCCThen, (v8i1 VCCR:$pred), zero_reg,
2356                            (v8i16 MQPR:$inactive))>;
2357  def : Pat<(v4i32 (vselect (v4i1 VCCR:$pred),
2358                            (v4i32 (ARMvdup (i32 rGPR:$elem))),
2359                            (v4i32 MQPR:$inactive))),
2360            (MVE_VDUP32 rGPR:$elem, ARMVCCThen, (v4i1 VCCR:$pred), zero_reg,
2361                            (v4i32 MQPR:$inactive))>;
2362  def : Pat<(v4f32 (vselect (v4i1 VCCR:$pred),
2363                            (v4f32 (ARMvdup (i32 rGPR:$elem))),
2364                            (v4f32 MQPR:$inactive))),
2365            (MVE_VDUP32 rGPR:$elem, ARMVCCThen, (v4i1 VCCR:$pred), zero_reg,
2366                            (v4f32 MQPR:$inactive))>;
2367  def : Pat<(v8f16 (vselect (v8i1 VCCR:$pred),
2368                            (v8f16 (ARMvdup (i32 rGPR:$elem))),
2369                            (v8f16 MQPR:$inactive))),
2370            (MVE_VDUP16 rGPR:$elem, ARMVCCThen, (v8i1 VCCR:$pred), zero_reg,
2371                            (v8f16 MQPR:$inactive))>;
2372}
2373
2374
2375class MVEIntSingleSrc<string iname, string suffix, bits<2> size,
2376                         list<dag> pattern=[]>
2377  : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qm), NoItinerary,
2378          iname, suffix, "$Qd, $Qm", vpred_r, "", size, pattern> {
2379  bits<4> Qd;
2380  bits<4> Qm;
2381
2382  let Inst{22} = Qd{3};
2383  let Inst{19-18} = size{1-0};
2384  let Inst{15-13} = Qd{2-0};
2385  let Inst{5} = Qm{3};
2386  let Inst{3-1} = Qm{2-0};
2387}
2388
2389class MVE_VCLSCLZ<string iname, string suffix, bits<2> size,
2390                   bit count_zeroes, list<dag> pattern=[]>
2391  : MVEIntSingleSrc<iname, suffix, size, pattern> {
2392
2393  let Inst{28} = 0b1;
2394  let Inst{25-23} = 0b111;
2395  let Inst{21-20} = 0b11;
2396  let Inst{17-16} = 0b00;
2397  let Inst{12-8} = 0b00100;
2398  let Inst{7} = count_zeroes;
2399  let Inst{6} = 0b1;
2400  let Inst{4} = 0b0;
2401  let Inst{0} = 0b0;
2402  let validForTailPredication = 1;
2403}
2404
2405multiclass MVE_VCLSCLZ_p<string opname, bit opcode, MVEVectorVTInfo VTI,
2406                         SDPatternOperator unpred_op> {
2407  def "": MVE_VCLSCLZ<"v"#opname, VTI.Suffix, VTI.Size, opcode>;
2408
2409  defvar Inst     = !cast<Instruction>(NAME);
2410  defvar pred_int = !cast<Intrinsic>("int_arm_mve_"#opname#"_predicated");
2411
2412  let Predicates = [HasMVEInt] in {
2413    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$val))),
2414              (VTI.Vec (Inst (VTI.Vec MQPR:$val)))>;
2415    def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$val), (VTI.Pred VCCR:$pred),
2416                                 (VTI.Vec MQPR:$inactive))),
2417              (VTI.Vec (Inst (VTI.Vec MQPR:$val), ARMVCCThen,
2418                             (VTI.Pred VCCR:$pred), zero_reg, (VTI.Vec MQPR:$inactive)))>;
2419  }
2420}
2421
2422defm MVE_VCLSs8  : MVE_VCLSCLZ_p<"cls", 0, MVE_v16s8, int_arm_mve_vcls>;
2423defm MVE_VCLSs16 : MVE_VCLSCLZ_p<"cls", 0, MVE_v8s16, int_arm_mve_vcls>;
2424defm MVE_VCLSs32 : MVE_VCLSCLZ_p<"cls", 0, MVE_v4s32, int_arm_mve_vcls>;
2425
2426defm MVE_VCLZs8  : MVE_VCLSCLZ_p<"clz", 1, MVE_v16i8, ctlz>;
2427defm MVE_VCLZs16 : MVE_VCLSCLZ_p<"clz", 1, MVE_v8i16, ctlz>;
2428defm MVE_VCLZs32 : MVE_VCLSCLZ_p<"clz", 1, MVE_v4i32, ctlz>;
2429
2430class MVE_VABSNEG_int<string iname, string suffix, bits<2> size, bit negate,
2431                      bit saturate, list<dag> pattern=[]>
2432  : MVEIntSingleSrc<iname, suffix, size, pattern> {
2433
2434  let Inst{28} = 0b1;
2435  let Inst{25-23} = 0b111;
2436  let Inst{21-20} = 0b11;
2437  let Inst{17} = 0b0;
2438  let Inst{16} = !eq(saturate, 0);
2439  let Inst{12-11} = 0b00;
2440  let Inst{10} = saturate;
2441  let Inst{9-8} = 0b11;
2442  let Inst{7} = negate;
2443  let Inst{6} = 0b1;
2444  let Inst{4} = 0b0;
2445  let Inst{0} = 0b0;
2446  let validForTailPredication = 1;
2447}
2448
2449multiclass MVE_VABSNEG_int_m<string iname, bit negate, bit saturate,
2450                             SDPatternOperator unpred_op, Intrinsic pred_int,
2451                             MVEVectorVTInfo VTI> {
2452  def "" : MVE_VABSNEG_int<iname, VTI.Suffix, VTI.Size, negate, saturate>;
2453  defvar Inst = !cast<Instruction>(NAME);
2454
2455  let Predicates = [HasMVEInt] in {
2456    // VQABS and VQNEG have more difficult isel patterns defined elsewhere
2457    if !not(saturate) then {
2458      def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$v))),
2459                (VTI.Vec (Inst $v))>;
2460    }
2461
2462    def : Pat<(VTI.Vec (pred_int  (VTI.Vec MQPR:$v), (VTI.Pred VCCR:$mask),
2463                                  (VTI.Vec MQPR:$inactive))),
2464              (VTI.Vec (Inst $v, ARMVCCThen, $mask, zero_reg, $inactive))>;
2465  }
2466}
2467
2468foreach VTI = [ MVE_v16s8, MVE_v8s16, MVE_v4s32 ] in {
2469  defm "MVE_VABS" # VTI.Suffix : MVE_VABSNEG_int_m<
2470     "vabs",  0, 0, abs,   int_arm_mve_abs_predicated,  VTI>;
2471  defm "MVE_VQABS" # VTI.Suffix : MVE_VABSNEG_int_m<
2472     "vqabs", 0, 1, ?,     int_arm_mve_qabs_predicated, VTI>;
2473  defm "MVE_VNEG" # VTI.Suffix : MVE_VABSNEG_int_m<
2474     "vneg",  1, 0, vnegq, int_arm_mve_neg_predicated,  VTI>;
2475  defm "MVE_VQNEG" # VTI.Suffix : MVE_VABSNEG_int_m<
2476     "vqneg", 1, 1, ?,     int_arm_mve_qneg_predicated, VTI>;
2477}
2478
2479// int_min/int_max: vector containing INT_MIN/INT_MAX VTI.Size times
2480// zero_vec: v4i32-initialized zero vector, potentially wrapped in a bitconvert
2481multiclass vqabsneg_pattern<MVEVectorVTInfo VTI, dag int_min, dag int_max,
2482                         dag zero_vec,  MVE_VABSNEG_int vqabs_instruction,
2483                         MVE_VABSNEG_int vqneg_instruction> {
2484  let Predicates = [HasMVEInt] in {
2485    // The below tree can be replaced by a vqabs instruction, as it represents
2486    // the following vectorized expression (r being the value in $reg):
2487    // r > 0 ? r : (r == INT_MIN ? INT_MAX : -r)
2488    def : Pat<(VTI.Vec (vselect
2489                      (VTI.Pred (ARMvcmpz (VTI.Vec MQPR:$reg), ARMCCgt)),
2490                      (VTI.Vec MQPR:$reg),
2491                      (VTI.Vec (vselect
2492                                (VTI.Pred (ARMvcmp (VTI.Vec MQPR:$reg), int_min, ARMCCeq)),
2493                                int_max,
2494                                (sub (VTI.Vec zero_vec), (VTI.Vec MQPR:$reg)))))),
2495            (VTI.Vec (vqabs_instruction (VTI.Vec MQPR:$reg)))>;
2496    // Similarly, this tree represents vqneg, i.e. the following vectorized expression:
2497    // r == INT_MIN ? INT_MAX : -r
2498    def : Pat<(VTI.Vec (vselect
2499                        (VTI.Pred (ARMvcmp (VTI.Vec MQPR:$reg), int_min, ARMCCeq)),
2500                        int_max,
2501                        (sub (VTI.Vec zero_vec), (VTI.Vec MQPR:$reg)))),
2502               (VTI.Vec (vqneg_instruction (VTI.Vec MQPR:$reg)))>;
2503  }
2504}
2505
2506defm MVE_VQABSNEG_Ps8  : vqabsneg_pattern<MVE_v16i8,
2507                                    (v16i8 (ARMvmovImm (i32 3712))),
2508                                    (v16i8 (ARMvmovImm (i32 3711))),
2509                                    (bitconvert (v4i32 (ARMvmovImm (i32 0)))),
2510                                    MVE_VQABSs8, MVE_VQNEGs8>;
2511defm MVE_VQABSNEG_Ps16 : vqabsneg_pattern<MVE_v8i16,
2512                                    (v8i16 (ARMvmovImm (i32 2688))),
2513                                    (v8i16 (ARMvmvnImm (i32 2688))),
2514                                    (bitconvert (v4i32 (ARMvmovImm (i32 0)))),
2515                                    MVE_VQABSs16, MVE_VQNEGs16>;
2516defm MVE_VQABSNEG_Ps32 : vqabsneg_pattern<MVE_v4i32,
2517                                    (v4i32 (ARMvmovImm (i32 1664))),
2518                                    (v4i32 (ARMvmvnImm (i32 1664))),
2519                                    (ARMvmovImm (i32 0)),
2520                                    MVE_VQABSs32, MVE_VQNEGs32>;
2521
2522class MVE_mod_imm<string iname, string suffix, bits<4> cmode, bit op,
2523                  dag iops, bits<2> vecsize, list<dag> pattern=[]>
2524  : MVE_p<(outs MQPR:$Qd), iops, NoItinerary, iname, suffix, "$Qd, $imm",
2525          vpred_r, "", vecsize, pattern> {
2526  bits<13> imm;
2527  bits<4> Qd;
2528
2529  let Inst{28} = imm{7};
2530  let Inst{25-23} = 0b111;
2531  let Inst{22} = Qd{3};
2532  let Inst{21-19} = 0b000;
2533  let Inst{18-16} = imm{6-4};
2534  let Inst{15-13} = Qd{2-0};
2535  let Inst{12} = 0b0;
2536  let Inst{11-8} = cmode{3-0};
2537  let Inst{7-6} = 0b01;
2538  let Inst{5} = op;
2539  let Inst{4} = 0b1;
2540  let Inst{3-0} = imm{3-0};
2541
2542  let DecoderMethod = "DecodeMVEModImmInstruction";
2543  let validForTailPredication = 1;
2544}
2545
2546let isReMaterializable = 1 in {
2547let isAsCheapAsAMove = 1 in {
2548def MVE_VMOVimmi8  : MVE_mod_imm<"vmov", "i8",  {1,1,1,0}, 0b0, (ins nImmSplatI8:$imm), 0b00>;
2549def MVE_VMOVimmi16 : MVE_mod_imm<"vmov", "i16", {1,0,?,0}, 0b0, (ins nImmSplatI16:$imm), 0b01> {
2550  let Inst{9} = imm{9};
2551}
2552def MVE_VMOVimmi32 : MVE_mod_imm<"vmov", "i32", {?,?,?,?}, 0b0, (ins nImmVMOVI32:$imm), 0b10> {
2553  let Inst{11-8} = imm{11-8};
2554}
2555def MVE_VMOVimmi64 : MVE_mod_imm<"vmov", "i64", {1,1,1,0}, 0b1, (ins nImmSplatI64:$imm), 0b11>;
2556def MVE_VMOVimmf32 : MVE_mod_imm<"vmov", "f32", {1,1,1,1}, 0b0, (ins nImmVMOVF32:$imm), 0b10>;
2557} // let isAsCheapAsAMove = 1
2558
2559def MVE_VMVNimmi16 : MVE_mod_imm<"vmvn", "i16", {1,0,?,0}, 0b1, (ins nImmSplatI16:$imm), 0b01> {
2560  let Inst{9} = imm{9};
2561}
2562def MVE_VMVNimmi32 : MVE_mod_imm<"vmvn", "i32", {?,?,?,?}, 0b1, (ins nImmVMOVI32:$imm), 0b10> {
2563  let Inst{11-8} = imm{11-8};
2564}
2565} // let isReMaterializable = 1
2566
2567let Predicates = [HasMVEInt] in {
2568  def : Pat<(v16i8 (ARMvmovImm timm:$simm)),
2569            (v16i8 (MVE_VMOVimmi8  nImmSplatI8:$simm))>;
2570  def : Pat<(v8i16 (ARMvmovImm timm:$simm)),
2571            (v8i16 (MVE_VMOVimmi16 nImmSplatI16:$simm))>;
2572  def : Pat<(v4i32 (ARMvmovImm timm:$simm)),
2573            (v4i32 (MVE_VMOVimmi32 nImmVMOVI32:$simm))>;
2574  def : Pat<(v2i64 (ARMvmovImm timm:$simm)),
2575            (v2i64 (MVE_VMOVimmi64 nImmSplatI64:$simm))>;
2576
2577  def : Pat<(v8i16 (ARMvmvnImm timm:$simm)),
2578            (v8i16 (MVE_VMVNimmi16 nImmSplatI16:$simm))>;
2579  def : Pat<(v4i32 (ARMvmvnImm timm:$simm)),
2580            (v4i32 (MVE_VMVNimmi32 nImmVMOVI32:$simm))>;
2581
2582  def : Pat<(v4f32 (ARMvmovFPImm timm:$simm)),
2583            (v4f32 (MVE_VMOVimmf32 nImmVMOVF32:$simm))>;
2584
2585  def : Pat<(v8i16 (vselect (v8i1 VCCR:$pred), (ARMvmvnImm timm:$simm),
2586                            MQPR:$inactive)),
2587            (v8i16 (MVE_VMVNimmi16 nImmSplatI16:$simm,
2588                            ARMVCCThen, VCCR:$pred, zero_reg, MQPR:$inactive))>;
2589  def : Pat<(v4i32 (vselect (v4i1 VCCR:$pred), (ARMvmvnImm timm:$simm),
2590                            MQPR:$inactive)),
2591            (v4i32 (MVE_VMVNimmi32 nImmSplatI32:$simm,
2592                            ARMVCCThen, VCCR:$pred, zero_reg, MQPR:$inactive))>;
2593}
2594
2595class MVE_VMINMAXA<string iname, string suffix, bits<2> size,
2596                   bit bit_12, list<dag> pattern=[]>
2597  : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qd_src, MQPR:$Qm),
2598          NoItinerary, iname, suffix, "$Qd, $Qm", vpred_n, "$Qd = $Qd_src",
2599          size, pattern> {
2600  bits<4> Qd;
2601  bits<4> Qm;
2602
2603  let Inst{28} = 0b0;
2604  let Inst{25-23} = 0b100;
2605  let Inst{22} = Qd{3};
2606  let Inst{21-20} = 0b11;
2607  let Inst{19-18} = size;
2608  let Inst{17-16} = 0b11;
2609  let Inst{15-13} = Qd{2-0};
2610  let Inst{12} = bit_12;
2611  let Inst{11-6} = 0b111010;
2612  let Inst{5} = Qm{3};
2613  let Inst{4} = 0b0;
2614  let Inst{3-1} = Qm{2-0};
2615  let Inst{0} = 0b1;
2616  let validForTailPredication = 1;
2617}
2618
2619multiclass MVE_VMINMAXA_m<string iname, MVEVectorVTInfo VTI,
2620                      SDNode unpred_op, Intrinsic pred_int, bit bit_12> {
2621  def "" : MVE_VMINMAXA<iname, VTI.Suffix, VTI.Size, bit_12>;
2622  defvar Inst = !cast<Instruction>(NAME);
2623
2624  let Predicates = [HasMVEInt] in {
2625    // Unpredicated v(min|max)a
2626    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$Qd), (abs (VTI.Vec MQPR:$Qm)))),
2627              (VTI.Vec (Inst (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm)))>;
2628
2629    // Predicated v(min|max)a
2630    def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm),
2631                            (VTI.Pred VCCR:$mask))),
2632              (VTI.Vec (Inst (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm),
2633                            ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg))>;
2634  }
2635}
2636
2637multiclass MVE_VMINA<MVEVectorVTInfo VTI>
2638  : MVE_VMINMAXA_m<"vmina", VTI, umin, int_arm_mve_vmina_predicated, 0b1>;
2639
2640defm MVE_VMINAs8  : MVE_VMINA<MVE_v16s8>;
2641defm MVE_VMINAs16 : MVE_VMINA<MVE_v8s16>;
2642defm MVE_VMINAs32 : MVE_VMINA<MVE_v4s32>;
2643
2644multiclass MVE_VMAXA<MVEVectorVTInfo VTI>
2645  : MVE_VMINMAXA_m<"vmaxa", VTI, umax, int_arm_mve_vmaxa_predicated, 0b0>;
2646
2647defm MVE_VMAXAs8  : MVE_VMAXA<MVE_v16s8>;
2648defm MVE_VMAXAs16 : MVE_VMAXA<MVE_v8s16>;
2649defm MVE_VMAXAs32 : MVE_VMAXA<MVE_v4s32>;
2650
2651// end of MVE Integer instructions
2652
2653// start of mve_imm_shift instructions
2654
2655def MVE_VSHLC : MVE_p<(outs rGPR:$RdmDest, MQPR:$Qd),
2656                      (ins MQPR:$QdSrc, rGPR:$RdmSrc, long_shift:$imm),
2657                      NoItinerary, "vshlc", "", "$QdSrc, $RdmSrc, $imm",
2658                      vpred_n, "$RdmDest = $RdmSrc,$Qd = $QdSrc", 0b10> {
2659  bits<5> imm;
2660  bits<4> Qd;
2661  bits<4> RdmDest;
2662
2663  let Inst{28} = 0b0;
2664  let Inst{25-23} = 0b101;
2665  let Inst{22} = Qd{3};
2666  let Inst{21} = 0b1;
2667  let Inst{20-16} = imm{4-0};
2668  let Inst{15-13} = Qd{2-0};
2669  let Inst{12-4} = 0b011111100;
2670  let Inst{3-0} = RdmDest{3-0};
2671}
2672
2673class MVE_shift_imm<dag oops, dag iops, string iname, string suffix,
2674                    string ops, vpred_ops vpred, string cstr,
2675                    bits<2> vecsize, list<dag> pattern=[]>
2676  : MVE_p<oops, iops, NoItinerary, iname, suffix, ops, vpred, cstr, vecsize, pattern> {
2677  bits<4> Qd;
2678  bits<4> Qm;
2679
2680  let Inst{22} = Qd{3};
2681  let Inst{15-13} = Qd{2-0};
2682  let Inst{5} = Qm{3};
2683  let Inst{3-1} = Qm{2-0};
2684}
2685
2686class MVE_VMOVL<string iname, string suffix, bits<2> sz, bit U, bit top,
2687              list<dag> pattern=[]>
2688  : MVE_shift_imm<(outs MQPR:$Qd), (ins MQPR:$Qm),
2689                  iname, suffix, "$Qd, $Qm", vpred_r, "",
2690                  sz, pattern> {
2691  let Inst{28} = U;
2692  let Inst{25-23} = 0b101;
2693  let Inst{21} = 0b1;
2694  let Inst{20-19} = sz{1-0};
2695  let Inst{18-16} = 0b000;
2696  let Inst{12} = top;
2697  let Inst{11-6} = 0b111101;
2698  let Inst{4} = 0b0;
2699  let Inst{0} = 0b0;
2700  let doubleWidthResult = 1;
2701}
2702
2703multiclass MVE_VMOVL_m<bit top, string chr, MVEVectorVTInfo OutVTI,
2704                       MVEVectorVTInfo InVTI> {
2705  def "": MVE_VMOVL<"vmovl" # chr, InVTI.Suffix, OutVTI.Size,
2706                    InVTI.Unsigned, top>;
2707  defvar Inst = !cast<Instruction>(NAME);
2708
2709  def : Pat<(OutVTI.Vec (int_arm_mve_vmovl_predicated (InVTI.Vec MQPR:$src),
2710                            (i32 InVTI.Unsigned), (i32 top),
2711                            (OutVTI.Pred VCCR:$pred),
2712                            (OutVTI.Vec MQPR:$inactive))),
2713            (OutVTI.Vec (Inst (InVTI.Vec MQPR:$src), ARMVCCThen,
2714                            (OutVTI.Pred VCCR:$pred), zero_reg,
2715                            (OutVTI.Vec MQPR:$inactive)))>;
2716}
2717
2718defm MVE_VMOVLs8bh  : MVE_VMOVL_m<0, "b", MVE_v8s16, MVE_v16s8>;
2719defm MVE_VMOVLs8th  : MVE_VMOVL_m<1, "t", MVE_v8s16, MVE_v16s8>;
2720defm MVE_VMOVLu8bh  : MVE_VMOVL_m<0, "b", MVE_v8u16, MVE_v16u8>;
2721defm MVE_VMOVLu8th  : MVE_VMOVL_m<1, "t", MVE_v8u16, MVE_v16u8>;
2722defm MVE_VMOVLs16bh : MVE_VMOVL_m<0, "b", MVE_v4s32, MVE_v8s16>;
2723defm MVE_VMOVLs16th : MVE_VMOVL_m<1, "t", MVE_v4s32, MVE_v8s16>;
2724defm MVE_VMOVLu16bh : MVE_VMOVL_m<0, "b", MVE_v4s32, MVE_v8u16>;
2725defm MVE_VMOVLu16th : MVE_VMOVL_m<1, "t", MVE_v4s32, MVE_v8u16>;
2726
2727let Predicates = [HasMVEInt] in {
2728  def : Pat<(sext_inreg (v4i32 MQPR:$src), v4i16),
2729            (MVE_VMOVLs16bh MQPR:$src)>;
2730  def : Pat<(sext_inreg (v8i16 MQPR:$src), v8i8),
2731            (MVE_VMOVLs8bh MQPR:$src)>;
2732  def : Pat<(sext_inreg (v4i32 MQPR:$src), v4i8),
2733            (MVE_VMOVLs16bh (MVE_VMOVLs8bh MQPR:$src))>;
2734
2735  def : Pat<(sext_inreg (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src)))), v8i8),
2736            (MVE_VMOVLs8th MQPR:$src)>;
2737  def : Pat<(sext_inreg (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src)))), v4i16),
2738            (MVE_VMOVLs16th MQPR:$src)>;
2739
2740  // zext_inreg 8 -> 16
2741  def : Pat<(ARMvbicImm (v8i16 MQPR:$src), (i32 0xAFF)),
2742            (MVE_VMOVLu8bh MQPR:$src)>;
2743  // zext_inreg 16 -> 32
2744  def : Pat<(and (v4i32 MQPR:$src), (v4i32 (ARMvmovImm (i32 0xCFF)))),
2745            (MVE_VMOVLu16bh MQPR:$src)>;
2746  // Same zext_inreg with vrevs, picking the top half
2747  def : Pat<(ARMvbicImm (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src)))), (i32 0xAFF)),
2748            (MVE_VMOVLu8th MQPR:$src)>;
2749  def : Pat<(and (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src)))),
2750                 (v4i32 (ARMvmovImm (i32 0xCFF)))),
2751            (MVE_VMOVLu16th MQPR:$src)>;
2752}
2753
2754
2755class MVE_VSHLL_imm<string iname, string suffix, bit U, bit th,
2756                    Operand immtype, bits<2> vecsize, list<dag> pattern=[]>
2757  : MVE_shift_imm<(outs MQPR:$Qd), (ins MQPR:$Qm, immtype:$imm),
2758                  iname, suffix, "$Qd, $Qm, $imm", vpred_r, "", vecsize, pattern> {
2759  let Inst{28} = U;
2760  let Inst{25-23} = 0b101;
2761  let Inst{21} = 0b1;
2762  let Inst{12} = th;
2763  let Inst{11-6} = 0b111101;
2764  let Inst{4} = 0b0;
2765  let Inst{0} = 0b0;
2766
2767  // For the MVE_VSHLL_patterns multiclass to refer to
2768  Operand immediateType = immtype;
2769
2770  let doubleWidthResult = 1;
2771}
2772
2773// The immediate VSHLL instructions accept shift counts from 1 up to
2774// the lane width (8 or 16), but the full-width shifts have an
2775// entirely separate encoding, given below with 'lw' in the name.
2776
2777class MVE_VSHLL_imm8<string iname, string suffix,
2778                     bit U, bit th, list<dag> pattern=[]>
2779  : MVE_VSHLL_imm<iname, suffix, U, th, mve_shift_imm1_7, 0b01, pattern> {
2780  bits<3> imm;
2781  let Inst{20-19} = 0b01;
2782  let Inst{18-16} = imm;
2783}
2784
2785class MVE_VSHLL_imm16<string iname, string suffix,
2786                      bit U, bit th, list<dag> pattern=[]>
2787  : MVE_VSHLL_imm<iname, suffix, U, th, mve_shift_imm1_15, 0b10, pattern> {
2788  bits<4> imm;
2789  let Inst{20} = 0b1;
2790  let Inst{19-16} = imm;
2791}
2792
2793def MVE_VSHLL_imms8bh  : MVE_VSHLL_imm8 <"vshllb", "s8", 0b0, 0b0>;
2794def MVE_VSHLL_imms8th  : MVE_VSHLL_imm8 <"vshllt", "s8", 0b0, 0b1>;
2795def MVE_VSHLL_immu8bh  : MVE_VSHLL_imm8 <"vshllb", "u8", 0b1, 0b0>;
2796def MVE_VSHLL_immu8th  : MVE_VSHLL_imm8 <"vshllt", "u8", 0b1, 0b1>;
2797def MVE_VSHLL_imms16bh : MVE_VSHLL_imm16<"vshllb", "s16", 0b0, 0b0>;
2798def MVE_VSHLL_imms16th : MVE_VSHLL_imm16<"vshllt", "s16", 0b0, 0b1>;
2799def MVE_VSHLL_immu16bh : MVE_VSHLL_imm16<"vshllb", "u16", 0b1, 0b0>;
2800def MVE_VSHLL_immu16th : MVE_VSHLL_imm16<"vshllt", "u16", 0b1, 0b1>;
2801
2802class MVE_VSHLL_by_lane_width<string iname, string suffix, bits<2> size,
2803                              bit U, string ops, list<dag> pattern=[]>
2804  : MVE_shift_imm<(outs MQPR:$Qd), (ins MQPR:$Qm),
2805                  iname, suffix, ops, vpred_r, "", !if(size, 0b10, 0b01), pattern> {
2806  let Inst{28} = U;
2807  let Inst{25-23} = 0b100;
2808  let Inst{21-20} = 0b11;
2809  let Inst{19-18} = size{1-0};
2810  let Inst{17-16} = 0b01;
2811  let Inst{11-6} = 0b111000;
2812  let Inst{4} = 0b0;
2813  let Inst{0} = 0b1;
2814  let doubleWidthResult = 1;
2815}
2816
2817multiclass MVE_VSHLL_lw<string iname, string suffix, bits<2> sz, bit U,
2818                              string ops, list<dag> pattern=[]> {
2819  def bh : MVE_VSHLL_by_lane_width<iname#"b", suffix, sz, U, ops, pattern> {
2820    let Inst{12} = 0b0;
2821  }
2822  def th : MVE_VSHLL_by_lane_width<iname#"t", suffix, sz, U, ops, pattern> {
2823    let Inst{12} = 0b1;
2824  }
2825}
2826
2827defm MVE_VSHLL_lws8  : MVE_VSHLL_lw<"vshll", "s8",  0b00, 0b0, "$Qd, $Qm, #8">;
2828defm MVE_VSHLL_lws16 : MVE_VSHLL_lw<"vshll", "s16", 0b01, 0b0, "$Qd, $Qm, #16">;
2829defm MVE_VSHLL_lwu8  : MVE_VSHLL_lw<"vshll", "u8",  0b00, 0b1, "$Qd, $Qm, #8">;
2830defm MVE_VSHLL_lwu16 : MVE_VSHLL_lw<"vshll", "u16", 0b01, 0b1, "$Qd, $Qm, #16">;
2831
2832multiclass MVE_VSHLL_patterns<MVEVectorVTInfo VTI, int top> {
2833  defvar suffix     = !strconcat(VTI.Suffix, !if(top, "th", "bh"));
2834  defvar inst_imm   = !cast<MVE_VSHLL_imm>("MVE_VSHLL_imm" # suffix);
2835  defvar inst_lw    = !cast<MVE_VSHLL_by_lane_width>("MVE_VSHLL_lw" # suffix);
2836  defvar unpred_int = int_arm_mve_vshll_imm;
2837  defvar pred_int   = int_arm_mve_vshll_imm_predicated;
2838  defvar imm        = inst_imm.immediateType;
2839
2840  def : Pat<(VTI.DblVec (unpred_int (VTI.Vec MQPR:$src), imm:$imm,
2841                                    (i32 VTI.Unsigned), (i32 top))),
2842            (VTI.DblVec (inst_imm   (VTI.Vec MQPR:$src), imm:$imm))>;
2843  def : Pat<(VTI.DblVec (unpred_int (VTI.Vec MQPR:$src), (i32 VTI.LaneBits),
2844                                    (i32 VTI.Unsigned), (i32 top))),
2845            (VTI.DblVec (inst_lw    (VTI.Vec MQPR:$src)))>;
2846
2847  def : Pat<(VTI.DblVec (pred_int   (VTI.Vec MQPR:$src), imm:$imm,
2848                                    (i32 VTI.Unsigned), (i32 top),
2849                                    (VTI.DblPred VCCR:$mask),
2850                                    (VTI.DblVec MQPR:$inactive))),
2851            (VTI.DblVec (inst_imm   (VTI.Vec MQPR:$src), imm:$imm,
2852                                    ARMVCCThen, (VTI.DblPred VCCR:$mask), zero_reg,
2853                                    (VTI.DblVec MQPR:$inactive)))>;
2854  def : Pat<(VTI.DblVec (pred_int   (VTI.Vec MQPR:$src), (i32 VTI.LaneBits),
2855                                    (i32 VTI.Unsigned), (i32 top),
2856                                    (VTI.DblPred VCCR:$mask),
2857                                    (VTI.DblVec MQPR:$inactive))),
2858            (VTI.DblVec (inst_lw    (VTI.Vec MQPR:$src), ARMVCCThen,
2859                                    (VTI.DblPred VCCR:$mask), zero_reg,
2860                                    (VTI.DblVec MQPR:$inactive)))>;
2861}
2862
2863foreach VTI = [MVE_v16s8, MVE_v8s16, MVE_v16u8, MVE_v8u16] in
2864  foreach top = [0, 1] in
2865    defm : MVE_VSHLL_patterns<VTI, top>;
2866
2867class MVE_shift_imm_partial<Operand imm, string iname, string suffix, bits<2> vecsize>
2868  : MVE_shift_imm<(outs MQPR:$Qd), (ins MQPR:$QdSrc, MQPR:$Qm, imm:$imm),
2869                  iname, suffix, "$Qd, $Qm, $imm", vpred_n, "$Qd = $QdSrc", vecsize> {
2870  Operand immediateType = imm;
2871}
2872
2873class MVE_VxSHRN<string iname, string suffix, bit bit_12, bit bit_28,
2874                 Operand imm, bits<2> vecsize>
2875  : MVE_shift_imm_partial<imm, iname, suffix, vecsize> {
2876  bits<5> imm;
2877
2878  let Inst{28} = bit_28;
2879  let Inst{25-23} = 0b101;
2880  let Inst{21} = 0b0;
2881  let Inst{20-16} = imm{4-0};
2882  let Inst{12} = bit_12;
2883  let Inst{11-6} = 0b111111;
2884  let Inst{4} = 0b0;
2885  let Inst{0} = 0b1;
2886  let validForTailPredication = 1;
2887  let retainsPreviousHalfElement = 1;
2888}
2889
2890def MVE_VRSHRNi16bh : MVE_VxSHRN<"vrshrnb", "i16", 0b0, 0b1, shr_imm8, 0b01> {
2891  let Inst{20-19} = 0b01;
2892}
2893def MVE_VRSHRNi16th : MVE_VxSHRN<"vrshrnt", "i16", 0b1, 0b1, shr_imm8, 0b01> {
2894  let Inst{20-19} = 0b01;
2895}
2896def MVE_VRSHRNi32bh : MVE_VxSHRN<"vrshrnb", "i32", 0b0, 0b1, shr_imm16, 0b10> {
2897  let Inst{20} = 0b1;
2898}
2899def MVE_VRSHRNi32th : MVE_VxSHRN<"vrshrnt", "i32", 0b1, 0b1, shr_imm16, 0b10> {
2900  let Inst{20} = 0b1;
2901}
2902
2903def MVE_VSHRNi16bh : MVE_VxSHRN<"vshrnb", "i16", 0b0, 0b0, shr_imm8, 0b01> {
2904  let Inst{20-19} = 0b01;
2905}
2906def MVE_VSHRNi16th : MVE_VxSHRN<"vshrnt", "i16", 0b1, 0b0, shr_imm8, 0b01> {
2907  let Inst{20-19} = 0b01;
2908}
2909def MVE_VSHRNi32bh : MVE_VxSHRN<"vshrnb", "i32", 0b0, 0b0, shr_imm16, 0b10> {
2910  let Inst{20} = 0b1;
2911}
2912def MVE_VSHRNi32th : MVE_VxSHRN<"vshrnt", "i32", 0b1, 0b0, shr_imm16, 0b10> {
2913  let Inst{20} = 0b1;
2914}
2915
2916class MVE_VxQRSHRUN<string iname, string suffix, bit bit_28, bit bit_12,
2917                    Operand imm, bits<2> vecsize>
2918  : MVE_shift_imm_partial<imm, iname, suffix, vecsize> {
2919  bits<5> imm;
2920
2921  let Inst{28} = bit_28;
2922  let Inst{25-23} = 0b101;
2923  let Inst{21} = 0b0;
2924  let Inst{20-16} = imm{4-0};
2925  let Inst{12} = bit_12;
2926  let Inst{11-6} = 0b111111;
2927  let Inst{4} = 0b0;
2928  let Inst{0} = 0b0;
2929  let validForTailPredication = 1;
2930  let retainsPreviousHalfElement = 1;
2931}
2932
2933def MVE_VQRSHRUNs16bh : MVE_VxQRSHRUN<
2934    "vqrshrunb", "s16", 0b1, 0b0, shr_imm8, 0b01> {
2935  let Inst{20-19} = 0b01;
2936}
2937def MVE_VQRSHRUNs16th : MVE_VxQRSHRUN<
2938    "vqrshrunt", "s16", 0b1, 0b1, shr_imm8, 0b01> {
2939  let Inst{20-19} = 0b01;
2940}
2941def MVE_VQRSHRUNs32bh : MVE_VxQRSHRUN<
2942    "vqrshrunb", "s32", 0b1, 0b0, shr_imm16, 0b10> {
2943  let Inst{20} = 0b1;
2944}
2945def MVE_VQRSHRUNs32th : MVE_VxQRSHRUN<
2946    "vqrshrunt", "s32", 0b1, 0b1, shr_imm16, 0b10> {
2947  let Inst{20} = 0b1;
2948}
2949
2950def MVE_VQSHRUNs16bh : MVE_VxQRSHRUN<
2951    "vqshrunb", "s16", 0b0, 0b0, shr_imm8, 0b01> {
2952  let Inst{20-19} = 0b01;
2953}
2954def MVE_VQSHRUNs16th : MVE_VxQRSHRUN<
2955    "vqshrunt", "s16", 0b0, 0b1, shr_imm8, 0b01> {
2956  let Inst{20-19} = 0b01;
2957}
2958def MVE_VQSHRUNs32bh : MVE_VxQRSHRUN<
2959    "vqshrunb", "s32", 0b0, 0b0, shr_imm16, 0b10> {
2960  let Inst{20} = 0b1;
2961}
2962def MVE_VQSHRUNs32th : MVE_VxQRSHRUN<
2963    "vqshrunt", "s32", 0b0, 0b1, shr_imm16, 0b10> {
2964  let Inst{20} = 0b1;
2965}
2966
2967class MVE_VxQRSHRN<string iname, string suffix, bit bit_0, bit bit_12,
2968                   Operand imm, bits<2> vecsize>
2969  : MVE_shift_imm_partial<imm, iname, suffix, vecsize> {
2970  bits<5> imm;
2971
2972  let Inst{25-23} = 0b101;
2973  let Inst{21} = 0b0;
2974  let Inst{20-16} = imm{4-0};
2975  let Inst{12} = bit_12;
2976  let Inst{11-6} = 0b111101;
2977  let Inst{4} = 0b0;
2978  let Inst{0} = bit_0;
2979  let validForTailPredication = 1;
2980  let retainsPreviousHalfElement = 1;
2981}
2982
2983multiclass MVE_VxQRSHRN_types<string iname, bit bit_0, bit bit_12> {
2984  def s16 : MVE_VxQRSHRN<iname, "s16", bit_0, bit_12, shr_imm8, 0b01> {
2985    let Inst{28} = 0b0;
2986    let Inst{20-19} = 0b01;
2987  }
2988  def u16 : MVE_VxQRSHRN<iname, "u16", bit_0, bit_12, shr_imm8, 0b01> {
2989    let Inst{28} = 0b1;
2990    let Inst{20-19} = 0b01;
2991  }
2992  def s32 : MVE_VxQRSHRN<iname, "s32", bit_0, bit_12, shr_imm16, 0b10> {
2993    let Inst{28} = 0b0;
2994    let Inst{20} = 0b1;
2995  }
2996  def u32 : MVE_VxQRSHRN<iname, "u32", bit_0, bit_12, shr_imm16, 0b10> {
2997    let Inst{28} = 0b1;
2998    let Inst{20} = 0b1;
2999  }
3000}
3001
3002defm MVE_VQRSHRNbh : MVE_VxQRSHRN_types<"vqrshrnb", 0b1, 0b0>;
3003defm MVE_VQRSHRNth : MVE_VxQRSHRN_types<"vqrshrnt", 0b1, 0b1>;
3004defm MVE_VQSHRNbh  : MVE_VxQRSHRN_types<"vqshrnb", 0b0, 0b0>;
3005defm MVE_VQSHRNth  : MVE_VxQRSHRN_types<"vqshrnt", 0b0, 0b1>;
3006
3007multiclass MVE_VSHRN_patterns<MVE_shift_imm_partial inst,
3008                              MVEVectorVTInfo OutVTI, MVEVectorVTInfo InVTI,
3009                              bit q, bit r, bit top> {
3010  defvar inparams = (? (OutVTI.Vec MQPR:$QdSrc), (InVTI.Vec MQPR:$Qm),
3011                       (inst.immediateType:$imm), (i32 q), (i32 r),
3012                       (i32 OutVTI.Unsigned), (i32 InVTI.Unsigned), (i32 top));
3013  defvar outparams = (inst (OutVTI.Vec MQPR:$QdSrc), (InVTI.Vec MQPR:$Qm),
3014                           (imm:$imm));
3015
3016  def : Pat<(OutVTI.Vec !setdagop(inparams, int_arm_mve_vshrn)),
3017            (OutVTI.Vec outparams)>;
3018  def : Pat<(OutVTI.Vec !con(inparams, (int_arm_mve_vshrn_predicated
3019                                           (InVTI.Pred VCCR:$pred)))),
3020            (OutVTI.Vec !con(outparams, (? ARMVCCThen, VCCR:$pred, zero_reg)))>;
3021}
3022
3023defm : MVE_VSHRN_patterns<MVE_VSHRNi16bh,    MVE_v16s8, MVE_v8s16, 0,0,0>;
3024defm : MVE_VSHRN_patterns<MVE_VSHRNi16th,    MVE_v16s8, MVE_v8s16, 0,0,1>;
3025defm : MVE_VSHRN_patterns<MVE_VSHRNi32bh,    MVE_v8s16, MVE_v4s32, 0,0,0>;
3026defm : MVE_VSHRN_patterns<MVE_VSHRNi32th,    MVE_v8s16, MVE_v4s32, 0,0,1>;
3027defm : MVE_VSHRN_patterns<MVE_VSHRNi16bh,    MVE_v16u8, MVE_v8u16, 0,0,0>;
3028defm : MVE_VSHRN_patterns<MVE_VSHRNi16th,    MVE_v16u8, MVE_v8u16, 0,0,1>;
3029defm : MVE_VSHRN_patterns<MVE_VSHRNi32bh,    MVE_v8u16, MVE_v4u32, 0,0,0>;
3030defm : MVE_VSHRN_patterns<MVE_VSHRNi32th,    MVE_v8u16, MVE_v4u32, 0,0,1>;
3031defm : MVE_VSHRN_patterns<MVE_VRSHRNi16bh,   MVE_v16s8, MVE_v8s16, 0,1,0>;
3032defm : MVE_VSHRN_patterns<MVE_VRSHRNi16th,   MVE_v16s8, MVE_v8s16, 0,1,1>;
3033defm : MVE_VSHRN_patterns<MVE_VRSHRNi32bh,   MVE_v8s16, MVE_v4s32, 0,1,0>;
3034defm : MVE_VSHRN_patterns<MVE_VRSHRNi32th,   MVE_v8s16, MVE_v4s32, 0,1,1>;
3035defm : MVE_VSHRN_patterns<MVE_VRSHRNi16bh,   MVE_v16u8, MVE_v8u16, 0,1,0>;
3036defm : MVE_VSHRN_patterns<MVE_VRSHRNi16th,   MVE_v16u8, MVE_v8u16, 0,1,1>;
3037defm : MVE_VSHRN_patterns<MVE_VRSHRNi32bh,   MVE_v8u16, MVE_v4u32, 0,1,0>;
3038defm : MVE_VSHRN_patterns<MVE_VRSHRNi32th,   MVE_v8u16, MVE_v4u32, 0,1,1>;
3039defm : MVE_VSHRN_patterns<MVE_VQSHRNbhs16,   MVE_v16s8, MVE_v8s16, 1,0,0>;
3040defm : MVE_VSHRN_patterns<MVE_VQSHRNths16,   MVE_v16s8, MVE_v8s16, 1,0,1>;
3041defm : MVE_VSHRN_patterns<MVE_VQSHRNbhs32,   MVE_v8s16, MVE_v4s32, 1,0,0>;
3042defm : MVE_VSHRN_patterns<MVE_VQSHRNths32,   MVE_v8s16, MVE_v4s32, 1,0,1>;
3043defm : MVE_VSHRN_patterns<MVE_VQSHRNbhu16,   MVE_v16u8, MVE_v8u16, 1,0,0>;
3044defm : MVE_VSHRN_patterns<MVE_VQSHRNthu16,   MVE_v16u8, MVE_v8u16, 1,0,1>;
3045defm : MVE_VSHRN_patterns<MVE_VQSHRNbhu32,   MVE_v8u16, MVE_v4u32, 1,0,0>;
3046defm : MVE_VSHRN_patterns<MVE_VQSHRNthu32,   MVE_v8u16, MVE_v4u32, 1,0,1>;
3047defm : MVE_VSHRN_patterns<MVE_VQRSHRNbhs16,  MVE_v16s8, MVE_v8s16, 1,1,0>;
3048defm : MVE_VSHRN_patterns<MVE_VQRSHRNths16,  MVE_v16s8, MVE_v8s16, 1,1,1>;
3049defm : MVE_VSHRN_patterns<MVE_VQRSHRNbhs32,  MVE_v8s16, MVE_v4s32, 1,1,0>;
3050defm : MVE_VSHRN_patterns<MVE_VQRSHRNths32,  MVE_v8s16, MVE_v4s32, 1,1,1>;
3051defm : MVE_VSHRN_patterns<MVE_VQRSHRNbhu16,  MVE_v16u8, MVE_v8u16, 1,1,0>;
3052defm : MVE_VSHRN_patterns<MVE_VQRSHRNthu16,  MVE_v16u8, MVE_v8u16, 1,1,1>;
3053defm : MVE_VSHRN_patterns<MVE_VQRSHRNbhu32,  MVE_v8u16, MVE_v4u32, 1,1,0>;
3054defm : MVE_VSHRN_patterns<MVE_VQRSHRNthu32,  MVE_v8u16, MVE_v4u32, 1,1,1>;
3055defm : MVE_VSHRN_patterns<MVE_VQSHRUNs16bh,  MVE_v16u8, MVE_v8s16, 1,0,0>;
3056defm : MVE_VSHRN_patterns<MVE_VQSHRUNs16th,  MVE_v16u8, MVE_v8s16, 1,0,1>;
3057defm : MVE_VSHRN_patterns<MVE_VQSHRUNs32bh,  MVE_v8u16, MVE_v4s32, 1,0,0>;
3058defm : MVE_VSHRN_patterns<MVE_VQSHRUNs32th,  MVE_v8u16, MVE_v4s32, 1,0,1>;
3059defm : MVE_VSHRN_patterns<MVE_VQRSHRUNs16bh, MVE_v16u8, MVE_v8s16, 1,1,0>;
3060defm : MVE_VSHRN_patterns<MVE_VQRSHRUNs16th, MVE_v16u8, MVE_v8s16, 1,1,1>;
3061defm : MVE_VSHRN_patterns<MVE_VQRSHRUNs32bh, MVE_v8u16, MVE_v4s32, 1,1,0>;
3062defm : MVE_VSHRN_patterns<MVE_VQRSHRUNs32th, MVE_v8u16, MVE_v4s32, 1,1,1>;
3063
3064// end of mve_imm_shift instructions
3065
3066// start of mve_shift instructions
3067
3068class MVE_shift_by_vec<string iname, string suffix, bit U,
3069                       bits<2> size, bit bit_4, bit bit_8>
3070  : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qm, MQPR:$Qn), NoItinerary,
3071           iname, suffix, "$Qd, $Qm, $Qn", vpred_r, "", size, []> {
3072  // Shift instructions which take a vector of shift counts
3073  bits<4> Qd;
3074  bits<4> Qm;
3075  bits<4> Qn;
3076
3077  let Inst{28} = U;
3078  let Inst{25-24} = 0b11;
3079  let Inst{23} = 0b0;
3080  let Inst{22} = Qd{3};
3081  let Inst{21-20} = size;
3082  let Inst{19-17} = Qn{2-0};
3083  let Inst{16} = 0b0;
3084  let Inst{15-13} = Qd{2-0};
3085  let Inst{12-9} = 0b0010;
3086  let Inst{8} = bit_8;
3087  let Inst{7} = Qn{3};
3088  let Inst{6} = 0b1;
3089  let Inst{5} = Qm{3};
3090  let Inst{4} = bit_4;
3091  let Inst{3-1} = Qm{2-0};
3092  let Inst{0} = 0b0;
3093  let validForTailPredication = 1;
3094}
3095
3096multiclass MVE_shift_by_vec_p<string iname, MVEVectorVTInfo VTI, bit q, bit r> {
3097  def "" : MVE_shift_by_vec<iname, VTI.Suffix, VTI.Unsigned, VTI.Size, q, r>;
3098  defvar Inst = !cast<Instruction>(NAME);
3099
3100  def : Pat<(VTI.Vec (int_arm_mve_vshl_vector
3101                         (VTI.Vec MQPR:$in), (VTI.Vec MQPR:$sh),
3102                         (i32 q), (i32 r), (i32 VTI.Unsigned))),
3103            (VTI.Vec (Inst (VTI.Vec MQPR:$in), (VTI.Vec MQPR:$sh)))>;
3104
3105  def : Pat<(VTI.Vec (int_arm_mve_vshl_vector_predicated
3106                         (VTI.Vec MQPR:$in), (VTI.Vec MQPR:$sh),
3107                         (i32 q), (i32 r), (i32 VTI.Unsigned),
3108                         (VTI.Pred VCCR:$mask), (VTI.Vec MQPR:$inactive))),
3109            (VTI.Vec (Inst (VTI.Vec MQPR:$in), (VTI.Vec MQPR:$sh),
3110                           ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg,
3111                           (VTI.Vec MQPR:$inactive)))>;
3112}
3113
3114multiclass mve_shift_by_vec_multi<string iname, bit bit_4, bit bit_8> {
3115  defm s8  : MVE_shift_by_vec_p<iname, MVE_v16s8, bit_4, bit_8>;
3116  defm s16 : MVE_shift_by_vec_p<iname, MVE_v8s16, bit_4, bit_8>;
3117  defm s32 : MVE_shift_by_vec_p<iname, MVE_v4s32, bit_4, bit_8>;
3118  defm u8  : MVE_shift_by_vec_p<iname, MVE_v16u8, bit_4, bit_8>;
3119  defm u16 : MVE_shift_by_vec_p<iname, MVE_v8u16, bit_4, bit_8>;
3120  defm u32 : MVE_shift_by_vec_p<iname, MVE_v4u32, bit_4, bit_8>;
3121}
3122
3123defm MVE_VSHL_by_vec   : mve_shift_by_vec_multi<"vshl",   0b0, 0b0>;
3124defm MVE_VQSHL_by_vec  : mve_shift_by_vec_multi<"vqshl",  0b1, 0b0>;
3125defm MVE_VQRSHL_by_vec : mve_shift_by_vec_multi<"vqrshl", 0b1, 0b1>;
3126defm MVE_VRSHL_by_vec  : mve_shift_by_vec_multi<"vrshl",  0b0, 0b1>;
3127
3128let Predicates = [HasMVEInt] in {
3129  defm : MVE_TwoOpPattern<MVE_v16i8, ARMvshlu, int_arm_mve_vshl_vector_predicated,
3130                          (? (i32 0), (i32 0), (i32 1)), MVE_VSHL_by_vecu8, null_frag>;
3131  defm : MVE_TwoOpPattern<MVE_v8i16, ARMvshlu, int_arm_mve_vshl_vector_predicated,
3132                          (? (i32 0), (i32 0), (i32 1)), MVE_VSHL_by_vecu16, null_frag>;
3133  defm : MVE_TwoOpPattern<MVE_v4i32, ARMvshlu, int_arm_mve_vshl_vector_predicated,
3134                          (? (i32 0), (i32 0), (i32 1)), MVE_VSHL_by_vecu32, null_frag>;
3135  defm : MVE_TwoOpPattern<MVE_v16i8, ARMvshls, int_arm_mve_vshl_vector_predicated,
3136                          (? (i32 0), (i32 0), (i32 0)), MVE_VSHL_by_vecs8, null_frag>;
3137  defm : MVE_TwoOpPattern<MVE_v8i16, ARMvshls, int_arm_mve_vshl_vector_predicated,
3138                          (? (i32 0), (i32 0), (i32 0)), MVE_VSHL_by_vecs16, null_frag>;
3139  defm : MVE_TwoOpPattern<MVE_v4i32, ARMvshls, int_arm_mve_vshl_vector_predicated,
3140                          (? (i32 0), (i32 0), (i32 0)), MVE_VSHL_by_vecs32, null_frag>;
3141}
3142
3143class MVE_shift_with_imm<string iname, string suffix, dag oops, dag iops,
3144                         string ops, vpred_ops vpred, string cstr,
3145                         bits<2> vecsize, list<dag> pattern=[]>
3146  : MVE_p<oops, iops, NoItinerary, iname, suffix, ops, vpred, cstr, vecsize, pattern> {
3147  bits<4> Qd;
3148  bits<4> Qm;
3149
3150  let Inst{23} = 0b1;
3151  let Inst{22} = Qd{3};
3152  let Inst{15-13} = Qd{2-0};
3153  let Inst{12-11} = 0b00;
3154  let Inst{7-6} = 0b01;
3155  let Inst{5} = Qm{3};
3156  let Inst{4} = 0b1;
3157  let Inst{3-1} = Qm{2-0};
3158  let Inst{0} = 0b0;
3159  let validForTailPredication = 1;
3160
3161  // For the MVE_shift_imm_patterns multiclass to refer to
3162  MVEVectorVTInfo VTI;
3163  Operand immediateType;
3164  Intrinsic unpred_int;
3165  Intrinsic pred_int;
3166  dag unsignedFlag = (?);
3167}
3168
3169class MVE_VSxI_imm<string iname, string suffix, bit bit_8, Operand immType, bits<2> vecsize>
3170  : MVE_shift_with_imm<iname, suffix, (outs MQPR:$Qd),
3171                       (ins MQPR:$Qd_src, MQPR:$Qm, immType:$imm),
3172                       "$Qd, $Qm, $imm", vpred_n, "$Qd = $Qd_src", vecsize> {
3173  bits<6> imm;
3174  let Inst{28} = 0b1;
3175  let Inst{25-24} = 0b11;
3176  let Inst{21-16} = imm;
3177  let Inst{10-9} = 0b10;
3178  let Inst{8} = bit_8;
3179  let validForTailPredication = 1;
3180
3181  Operand immediateType = immType;
3182}
3183
3184def MVE_VSRIimm8 : MVE_VSxI_imm<"vsri", "8", 0b0, shr_imm8, 0b00> {
3185  let Inst{21-19} = 0b001;
3186}
3187
3188def MVE_VSRIimm16 : MVE_VSxI_imm<"vsri", "16", 0b0, shr_imm16, 0b01> {
3189  let Inst{21-20} = 0b01;
3190}
3191
3192def MVE_VSRIimm32 : MVE_VSxI_imm<"vsri", "32", 0b0, shr_imm32, 0b10> {
3193  let Inst{21} = 0b1;
3194}
3195
3196def MVE_VSLIimm8 : MVE_VSxI_imm<"vsli", "8", 0b1, imm0_7, 0b00> {
3197  let Inst{21-19} = 0b001;
3198}
3199
3200def MVE_VSLIimm16 : MVE_VSxI_imm<"vsli", "16", 0b1, imm0_15, 0b01> {
3201  let Inst{21-20} = 0b01;
3202}
3203
3204def MVE_VSLIimm32 : MVE_VSxI_imm<"vsli", "32", 0b1,imm0_31, 0b10> {
3205  let Inst{21} = 0b1;
3206}
3207
3208multiclass MVE_VSxI_patterns<MVE_VSxI_imm inst, string name,
3209                              MVEVectorVTInfo VTI> {
3210  defvar inparams = (? (VTI.Vec MQPR:$QdSrc), (VTI.Vec MQPR:$Qm),
3211                       (inst.immediateType:$imm));
3212  defvar outparams = (inst (VTI.Vec MQPR:$QdSrc), (VTI.Vec MQPR:$Qm),
3213                           (inst.immediateType:$imm));
3214  defvar unpred_int = !cast<Intrinsic>("int_arm_mve_" # name);
3215  defvar pred_int = !cast<Intrinsic>("int_arm_mve_" # name # "_predicated");
3216
3217  def : Pat<(VTI.Vec !setdagop(inparams, unpred_int)),
3218            (VTI.Vec outparams)>;
3219  def : Pat<(VTI.Vec !con(inparams, (pred_int (VTI.Pred VCCR:$pred)))),
3220            (VTI.Vec !con(outparams, (? ARMVCCThen, VCCR:$pred, zero_reg)))>;
3221}
3222
3223defm : MVE_VSxI_patterns<MVE_VSLIimm8,  "vsli", MVE_v16i8>;
3224defm : MVE_VSxI_patterns<MVE_VSLIimm16, "vsli", MVE_v8i16>;
3225defm : MVE_VSxI_patterns<MVE_VSLIimm32, "vsli", MVE_v4i32>;
3226defm : MVE_VSxI_patterns<MVE_VSRIimm8,  "vsri", MVE_v16i8>;
3227defm : MVE_VSxI_patterns<MVE_VSRIimm16, "vsri", MVE_v8i16>;
3228defm : MVE_VSxI_patterns<MVE_VSRIimm32, "vsri", MVE_v4i32>;
3229
3230class MVE_VQSHL_imm<MVEVectorVTInfo VTI_, Operand immType>
3231  : MVE_shift_with_imm<"vqshl", VTI_.Suffix, (outs MQPR:$Qd),
3232                       (ins MQPR:$Qm, immType:$imm), "$Qd, $Qm, $imm",
3233                       vpred_r, "", VTI_.Size> {
3234  bits<6> imm;
3235
3236  let Inst{28} = VTI_.Unsigned;
3237  let Inst{25-24} = 0b11;
3238  let Inst{21-16} = imm;
3239  let Inst{10-8} = 0b111;
3240
3241  let VTI = VTI_;
3242  let immediateType = immType;
3243  let unsignedFlag = (? (i32 VTI.Unsigned));
3244}
3245
3246let unpred_int = int_arm_mve_vqshl_imm,
3247    pred_int = int_arm_mve_vqshl_imm_predicated in {
3248  def MVE_VQSHLimms8 : MVE_VQSHL_imm<MVE_v16s8, imm0_7> {
3249    let Inst{21-19} = 0b001;
3250  }
3251  def MVE_VQSHLimmu8 : MVE_VQSHL_imm<MVE_v16u8, imm0_7> {
3252    let Inst{21-19} = 0b001;
3253  }
3254
3255  def MVE_VQSHLimms16 : MVE_VQSHL_imm<MVE_v8s16, imm0_15> {
3256    let Inst{21-20} = 0b01;
3257  }
3258  def MVE_VQSHLimmu16 : MVE_VQSHL_imm<MVE_v8u16, imm0_15> {
3259    let Inst{21-20} = 0b01;
3260  }
3261
3262  def MVE_VQSHLimms32 : MVE_VQSHL_imm<MVE_v4s32, imm0_31> {
3263    let Inst{21} = 0b1;
3264  }
3265  def MVE_VQSHLimmu32 : MVE_VQSHL_imm<MVE_v4u32, imm0_31> {
3266    let Inst{21} = 0b1;
3267  }
3268}
3269
3270class MVE_VQSHLU_imm<MVEVectorVTInfo VTI_, Operand immType>
3271  : MVE_shift_with_imm<"vqshlu", VTI_.Suffix, (outs MQPR:$Qd),
3272                       (ins MQPR:$Qm, immType:$imm), "$Qd, $Qm, $imm",
3273                       vpred_r, "", VTI_.Size> {
3274  bits<6> imm;
3275
3276  let Inst{28} = 0b1;
3277  let Inst{25-24} = 0b11;
3278  let Inst{21-16} = imm;
3279  let Inst{10-8} = 0b110;
3280
3281  let VTI = VTI_;
3282  let immediateType = immType;
3283}
3284
3285let unpred_int = int_arm_mve_vqshlu_imm,
3286    pred_int = int_arm_mve_vqshlu_imm_predicated in {
3287  def MVE_VQSHLU_imms8 : MVE_VQSHLU_imm<MVE_v16s8, imm0_7> {
3288    let Inst{21-19} = 0b001;
3289  }
3290
3291  def MVE_VQSHLU_imms16 : MVE_VQSHLU_imm<MVE_v8s16, imm0_15> {
3292    let Inst{21-20} = 0b01;
3293  }
3294
3295  def MVE_VQSHLU_imms32 : MVE_VQSHLU_imm<MVE_v4s32, imm0_31> {
3296    let Inst{21} = 0b1;
3297  }
3298}
3299
3300class MVE_VRSHR_imm<MVEVectorVTInfo VTI_, Operand immType>
3301  : MVE_shift_with_imm<"vrshr", VTI_.Suffix, (outs MQPR:$Qd),
3302                       (ins MQPR:$Qm, immType:$imm), "$Qd, $Qm, $imm",
3303                       vpred_r, "", VTI_.Size> {
3304  bits<6> imm;
3305
3306  let Inst{28} = VTI_.Unsigned;
3307  let Inst{25-24} = 0b11;
3308  let Inst{21-16} = imm;
3309  let Inst{10-8} = 0b010;
3310
3311  let VTI = VTI_;
3312  let immediateType = immType;
3313  let unsignedFlag = (? (i32 VTI.Unsigned));
3314}
3315
3316let unpred_int = int_arm_mve_vrshr_imm,
3317    pred_int = int_arm_mve_vrshr_imm_predicated in {
3318  def MVE_VRSHR_imms8 : MVE_VRSHR_imm<MVE_v16s8, shr_imm8> {
3319    let Inst{21-19} = 0b001;
3320  }
3321
3322  def MVE_VRSHR_immu8 : MVE_VRSHR_imm<MVE_v16u8, shr_imm8> {
3323    let Inst{21-19} = 0b001;
3324  }
3325
3326  def MVE_VRSHR_imms16 : MVE_VRSHR_imm<MVE_v8s16, shr_imm16> {
3327    let Inst{21-20} = 0b01;
3328  }
3329
3330  def MVE_VRSHR_immu16 : MVE_VRSHR_imm<MVE_v8u16, shr_imm16> {
3331    let Inst{21-20} = 0b01;
3332  }
3333
3334  def MVE_VRSHR_imms32 : MVE_VRSHR_imm<MVE_v4s32, shr_imm32> {
3335    let Inst{21} = 0b1;
3336  }
3337
3338  def MVE_VRSHR_immu32 : MVE_VRSHR_imm<MVE_v4u32, shr_imm32> {
3339    let Inst{21} = 0b1;
3340  }
3341}
3342
3343multiclass MVE_shift_imm_patterns<MVE_shift_with_imm inst> {
3344  def : Pat<(inst.VTI.Vec !con((inst.unpred_int (inst.VTI.Vec MQPR:$src),
3345                                                inst.immediateType:$imm),
3346                               inst.unsignedFlag)),
3347            (inst.VTI.Vec (inst (inst.VTI.Vec MQPR:$src),
3348                                inst.immediateType:$imm))>;
3349
3350  def : Pat<(inst.VTI.Vec !con((inst.pred_int (inst.VTI.Vec MQPR:$src),
3351                                              inst.immediateType:$imm),
3352                               inst.unsignedFlag,
3353                               (? (inst.VTI.Pred VCCR:$mask),
3354                                  (inst.VTI.Vec MQPR:$inactive)))),
3355            (inst.VTI.Vec (inst (inst.VTI.Vec MQPR:$src),
3356                                inst.immediateType:$imm,
3357                                ARMVCCThen, (inst.VTI.Pred VCCR:$mask), zero_reg,
3358                                (inst.VTI.Vec MQPR:$inactive)))>;
3359}
3360
3361defm : MVE_shift_imm_patterns<MVE_VQSHLimms8>;
3362defm : MVE_shift_imm_patterns<MVE_VQSHLimmu8>;
3363defm : MVE_shift_imm_patterns<MVE_VQSHLimms16>;
3364defm : MVE_shift_imm_patterns<MVE_VQSHLimmu16>;
3365defm : MVE_shift_imm_patterns<MVE_VQSHLimms32>;
3366defm : MVE_shift_imm_patterns<MVE_VQSHLimmu32>;
3367defm : MVE_shift_imm_patterns<MVE_VQSHLU_imms8>;
3368defm : MVE_shift_imm_patterns<MVE_VQSHLU_imms16>;
3369defm : MVE_shift_imm_patterns<MVE_VQSHLU_imms32>;
3370defm : MVE_shift_imm_patterns<MVE_VRSHR_imms8>;
3371defm : MVE_shift_imm_patterns<MVE_VRSHR_immu8>;
3372defm : MVE_shift_imm_patterns<MVE_VRSHR_imms16>;
3373defm : MVE_shift_imm_patterns<MVE_VRSHR_immu16>;
3374defm : MVE_shift_imm_patterns<MVE_VRSHR_imms32>;
3375defm : MVE_shift_imm_patterns<MVE_VRSHR_immu32>;
3376
3377class MVE_VSHR_imm<string suffix, dag imm, bits<2> vecsize>
3378  : MVE_shift_with_imm<"vshr", suffix, (outs MQPR:$Qd),
3379                       !con((ins MQPR:$Qm), imm), "$Qd, $Qm, $imm",
3380                       vpred_r, "", vecsize> {
3381  bits<6> imm;
3382
3383  let Inst{25-24} = 0b11;
3384  let Inst{21-16} = imm;
3385  let Inst{10-8} = 0b000;
3386}
3387
3388def MVE_VSHR_imms8 : MVE_VSHR_imm<"s8", (ins shr_imm8:$imm), 0b00> {
3389  let Inst{28} = 0b0;
3390  let Inst{21-19} = 0b001;
3391}
3392
3393def MVE_VSHR_immu8 : MVE_VSHR_imm<"u8", (ins shr_imm8:$imm), 0b00> {
3394  let Inst{28} = 0b1;
3395  let Inst{21-19} = 0b001;
3396}
3397
3398def MVE_VSHR_imms16 : MVE_VSHR_imm<"s16", (ins shr_imm16:$imm), 0b01> {
3399  let Inst{28} = 0b0;
3400  let Inst{21-20} = 0b01;
3401}
3402
3403def MVE_VSHR_immu16 : MVE_VSHR_imm<"u16", (ins shr_imm16:$imm), 0b01> {
3404  let Inst{28} = 0b1;
3405  let Inst{21-20} = 0b01;
3406}
3407
3408def MVE_VSHR_imms32 : MVE_VSHR_imm<"s32", (ins shr_imm32:$imm), 0b10> {
3409  let Inst{28} = 0b0;
3410  let Inst{21} = 0b1;
3411}
3412
3413def MVE_VSHR_immu32 : MVE_VSHR_imm<"u32", (ins shr_imm32:$imm), 0b10> {
3414  let Inst{28} = 0b1;
3415  let Inst{21} = 0b1;
3416}
3417
3418class MVE_VSHL_imm<string suffix, dag imm, bits<2> vecsize>
3419  : MVE_shift_with_imm<"vshl", suffix, (outs MQPR:$Qd),
3420                       !con((ins MQPR:$Qm), imm), "$Qd, $Qm, $imm",
3421                       vpred_r, "", vecsize> {
3422  bits<6> imm;
3423
3424  let Inst{28} = 0b0;
3425  let Inst{25-24} = 0b11;
3426  let Inst{21-16} = imm;
3427  let Inst{10-8} = 0b101;
3428}
3429
3430def MVE_VSHL_immi8 : MVE_VSHL_imm<"i8", (ins imm0_7:$imm), 0b00> {
3431  let Inst{21-19} = 0b001;
3432}
3433
3434def MVE_VSHL_immi16 : MVE_VSHL_imm<"i16", (ins imm0_15:$imm), 0b01> {
3435  let Inst{21-20} = 0b01;
3436}
3437
3438def MVE_VSHL_immi32 : MVE_VSHL_imm<"i32", (ins imm0_31:$imm), 0b10> {
3439  let Inst{21} = 0b1;
3440}
3441
3442multiclass MVE_immediate_shift_patterns_inner<
3443    MVEVectorVTInfo VTI, Operand imm_operand_type, SDNode unpred_op,
3444    Intrinsic pred_int, Instruction inst, list<int> unsignedFlag = []> {
3445
3446  def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$src), imm_operand_type:$imm)),
3447            (VTI.Vec (inst (VTI.Vec MQPR:$src), imm_operand_type:$imm))>;
3448
3449  def : Pat<(VTI.Vec !con((pred_int (VTI.Vec MQPR:$src), imm_operand_type:$imm),
3450                          !dag(pred_int, unsignedFlag, ?),
3451                          (pred_int (VTI.Pred VCCR:$mask),
3452                                   (VTI.Vec MQPR:$inactive)))),
3453            (VTI.Vec (inst (VTI.Vec MQPR:$src), imm_operand_type:$imm,
3454                           ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg,
3455                           (VTI.Vec MQPR:$inactive)))>;
3456}
3457
3458multiclass MVE_immediate_shift_patterns<MVEVectorVTInfo VTI,
3459                                        Operand imm_operand_type> {
3460  defm : MVE_immediate_shift_patterns_inner<VTI, imm_operand_type,
3461      ARMvshlImm, int_arm_mve_shl_imm_predicated,
3462      !cast<Instruction>("MVE_VSHL_immi" # VTI.BitsSuffix)>;
3463  defm : MVE_immediate_shift_patterns_inner<VTI, imm_operand_type,
3464      ARMvshruImm, int_arm_mve_shr_imm_predicated,
3465      !cast<Instruction>("MVE_VSHR_immu" # VTI.BitsSuffix), [1]>;
3466  defm : MVE_immediate_shift_patterns_inner<VTI, imm_operand_type,
3467      ARMvshrsImm, int_arm_mve_shr_imm_predicated,
3468      !cast<Instruction>("MVE_VSHR_imms" # VTI.BitsSuffix), [0]>;
3469}
3470
3471let Predicates = [HasMVEInt] in {
3472  defm : MVE_immediate_shift_patterns<MVE_v16i8, imm0_7>;
3473  defm : MVE_immediate_shift_patterns<MVE_v8i16, imm0_15>;
3474  defm : MVE_immediate_shift_patterns<MVE_v4i32, imm0_31>;
3475}
3476
3477// end of mve_shift instructions
3478
3479// start of MVE Floating Point instructions
3480
3481class MVE_float<string iname, string suffix, dag oops, dag iops, string ops,
3482                vpred_ops vpred, string cstr, bits<2> vecsize, list<dag> pattern=[]>
3483  : MVE_f<oops, iops, NoItinerary, iname, suffix, ops, vpred, cstr, vecsize, pattern> {
3484  bits<4> Qm;
3485
3486  let Inst{12} = 0b0;
3487  let Inst{6} = 0b1;
3488  let Inst{5} = Qm{3};
3489  let Inst{3-1} = Qm{2-0};
3490  let Inst{0} = 0b0;
3491}
3492
3493class MVE_VRINT<string rmode, bits<3> op, string suffix, bits<2> size,
3494                list<dag> pattern=[]>
3495  : MVE_float<!strconcat("vrint", rmode), suffix, (outs MQPR:$Qd),
3496              (ins MQPR:$Qm), "$Qd, $Qm", vpred_r, "", size, pattern> {
3497  bits<4> Qd;
3498
3499  let Inst{28} = 0b1;
3500  let Inst{25-23} = 0b111;
3501  let Inst{22} = Qd{3};
3502  let Inst{21-20} = 0b11;
3503  let Inst{19-18} = size;
3504  let Inst{17-16} = 0b10;
3505  let Inst{15-13} = Qd{2-0};
3506  let Inst{11-10} = 0b01;
3507  let Inst{9-7} = op{2-0};
3508  let Inst{4} = 0b0;
3509  let validForTailPredication = 1;
3510
3511}
3512
3513multiclass MVE_VRINT_m<MVEVectorVTInfo VTI, string suffix, bits<3> opcode,
3514                       SDPatternOperator unpred_op> {
3515  def "": MVE_VRINT<suffix, opcode, VTI.Suffix, VTI.Size>;
3516  defvar Inst = !cast<Instruction>(NAME);
3517  defvar pred_int = !cast<Intrinsic>("int_arm_mve_vrint"#suffix#"_predicated");
3518
3519  let Predicates = [HasMVEFloat] in {
3520    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$val))),
3521              (VTI.Vec (Inst (VTI.Vec MQPR:$val)))>;
3522    def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$val), (VTI.Pred VCCR:$pred),
3523                                 (VTI.Vec MQPR:$inactive))),
3524              (VTI.Vec (Inst (VTI.Vec MQPR:$val), ARMVCCThen,
3525                             (VTI.Pred VCCR:$pred), zero_reg, (VTI.Vec MQPR:$inactive)))>;
3526  }
3527}
3528
3529multiclass MVE_VRINT_ops<MVEVectorVTInfo VTI> {
3530  defm N : MVE_VRINT_m<VTI, "n", 0b000, int_arm_mve_vrintn>;
3531  defm X : MVE_VRINT_m<VTI, "x", 0b001, frint>;
3532  defm A : MVE_VRINT_m<VTI, "a", 0b010, fround>;
3533  defm Z : MVE_VRINT_m<VTI, "z", 0b011, ftrunc>;
3534  defm M : MVE_VRINT_m<VTI, "m", 0b101, ffloor>;
3535  defm P : MVE_VRINT_m<VTI, "p", 0b111, fceil>;
3536}
3537
3538defm MVE_VRINTf16 : MVE_VRINT_ops<MVE_v8f16>;
3539defm MVE_VRINTf32 : MVE_VRINT_ops<MVE_v4f32>;
3540
3541class MVEFloatArithNeon<string iname, string suffix, bit size,
3542                           dag oops, dag iops, string ops,
3543                           vpred_ops vpred, string cstr, bits<2> vecsize, list<dag> pattern=[]>
3544  : MVE_float<iname, suffix, oops, iops, ops, vpred, cstr, vecsize, pattern> {
3545  let Inst{20} = size;
3546  let Inst{16} = 0b0;
3547}
3548
3549class MVE_VMUL_fp<string iname, string suffix, bits<2> size, list<dag> pattern=[]>
3550  : MVEFloatArithNeon<iname, suffix, size{0}, (outs MQPR:$Qd),
3551                      (ins MQPR:$Qn, MQPR:$Qm), "$Qd, $Qn, $Qm", vpred_r, "",
3552                      size, pattern> {
3553  bits<4> Qd;
3554  bits<4> Qn;
3555
3556  let Inst{28} = 0b1;
3557  let Inst{25-23} = 0b110;
3558  let Inst{22} = Qd{3};
3559  let Inst{21} = 0b0;
3560  let Inst{19-17} = Qn{2-0};
3561  let Inst{15-13} = Qd{2-0};
3562  let Inst{12-8} = 0b01101;
3563  let Inst{7} = Qn{3};
3564  let Inst{4} = 0b1;
3565  let validForTailPredication = 1;
3566}
3567
3568multiclass MVE_VMULT_fp_m<string iname, MVEVectorVTInfo VTI, SDNode Op,
3569                          Intrinsic PredInt, SDPatternOperator IdentityVec> {
3570  def "" : MVE_VMUL_fp<iname, VTI.Suffix, VTI.Size>;
3571  defvar Inst = !cast<Instruction>(NAME);
3572
3573  let Predicates = [HasMVEFloat] in {
3574    defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? ), !cast<Instruction>(NAME), IdentityVec>;
3575  }
3576}
3577
3578multiclass MVE_VMUL_fp_m<MVEVectorVTInfo VTI, SDPatternOperator IdentityVec>
3579  : MVE_VMULT_fp_m<"vmul", VTI, fmul, int_arm_mve_mul_predicated, IdentityVec>;
3580
3581def ARMimmOneF: PatLeaf<(bitconvert (v4f32 (ARMvmovFPImm (i32 112))))>; // 1.0 float
3582def ARMimmOneH: PatLeaf<(bitconvert (v8i16 (ARMvmovImm (i32 2620))))>; // 1.0 half
3583
3584defm MVE_VMULf32 : MVE_VMUL_fp_m<MVE_v4f32, ARMimmOneF>;
3585defm MVE_VMULf16 : MVE_VMUL_fp_m<MVE_v8f16, ARMimmOneH>;
3586
3587class MVE_VCMLA<string suffix, bits<2> size, string cstr>
3588  : MVEFloatArithNeon<"vcmla", suffix, size{1}, (outs MQPR:$Qd),
3589                         (ins MQPR:$Qd_src, MQPR:$Qn, MQPR:$Qm, complexrotateop:$rot),
3590                         "$Qd, $Qn, $Qm, $rot", vpred_n, "$Qd = $Qd_src"#cstr, size, []> {
3591  bits<4> Qd;
3592  bits<4> Qn;
3593  bits<2> rot;
3594
3595  let Inst{28} = 0b1;
3596  let Inst{25} = 0b0;
3597  let Inst{24-23} = rot;
3598  let Inst{22} = Qd{3};
3599  let Inst{21} = 0b1;
3600  let Inst{19-17} = Qn{2-0};
3601  let Inst{15-13} = Qd{2-0};
3602  let Inst{12-8} = 0b01000;
3603  let Inst{7} = Qn{3};
3604  let Inst{4} = 0b0;
3605}
3606
3607multiclass MVE_VCMLA_m<MVEVectorVTInfo VTI, string cstr=""> {
3608  def "" : MVE_VCMLA<VTI.Suffix, VTI.Size, cstr>;
3609  defvar Inst = !cast<Instruction>(NAME);
3610
3611  let Predicates = [HasMVEFloat] in {
3612    def : Pat<(VTI.Vec (int_arm_mve_vcmlaq
3613                            imm:$rot, (VTI.Vec MQPR:$Qd_src),
3614                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
3615              (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src),
3616                             (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
3617                             imm:$rot))>;
3618
3619    def: Pat<(VTI.Vec (fadd_contract MQPR:$Qd_src,
3620                            (int_arm_mve_vcmulq imm:$rot,
3621                              (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm)))),
3622              (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src),
3623                             (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
3624                             imm:$rot))>;
3625
3626    def : Pat<(VTI.Vec (int_arm_mve_vcmlaq_predicated
3627                            imm:$rot, (VTI.Vec MQPR:$Qd_src),
3628                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
3629                            (VTI.Pred VCCR:$mask))),
3630              (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src), (VTI.Vec MQPR:$Qn),
3631                             (VTI.Vec MQPR:$Qm), imm:$rot,
3632                             ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg))>;
3633  }
3634}
3635
3636defm MVE_VCMLAf16 : MVE_VCMLA_m<MVE_v8f16>;
3637defm MVE_VCMLAf32 : MVE_VCMLA_m<MVE_v4f32, ",@earlyclobber $Qd">;
3638
3639class MVE_VADDSUBFMA_fp<string iname, string suffix, bits<2> size, bit bit_4,
3640                        bit bit_8, bit bit_21, dag iops=(ins),
3641                        vpred_ops vpred=vpred_r, string cstr="",
3642                        list<dag> pattern=[]>
3643  : MVEFloatArithNeon<iname, suffix, size{0}, (outs MQPR:$Qd),
3644                      !con(iops, (ins MQPR:$Qn, MQPR:$Qm)), "$Qd, $Qn, $Qm",
3645                      vpred, cstr, size, pattern> {
3646  bits<4> Qd;
3647  bits<4> Qn;
3648
3649  let Inst{28} = 0b0;
3650  let Inst{25-23} = 0b110;
3651  let Inst{22} = Qd{3};
3652  let Inst{21} = bit_21;
3653  let Inst{19-17} = Qn{2-0};
3654  let Inst{15-13} = Qd{2-0};
3655  let Inst{11-9} = 0b110;
3656  let Inst{8} = bit_8;
3657  let Inst{7} = Qn{3};
3658  let Inst{4} = bit_4;
3659  let validForTailPredication = 1;
3660}
3661
3662multiclass MVE_VFMA_fp_multi<string iname, bit fms, MVEVectorVTInfo VTI> {
3663  def "" : MVE_VADDSUBFMA_fp<iname, VTI.Suffix, VTI.Size, 0b1, 0b0, fms,
3664                             (ins MQPR:$Qd_src), vpred_n, "$Qd = $Qd_src">;
3665  defvar Inst = !cast<Instruction>(NAME);
3666  defvar pred_int = int_arm_mve_fma_predicated;
3667  defvar m1   = (VTI.Vec MQPR:$m1);
3668  defvar m2   = (VTI.Vec MQPR:$m2);
3669  defvar add  = (VTI.Vec MQPR:$add);
3670  defvar pred = (VTI.Pred VCCR:$pred);
3671
3672  let Predicates = [HasMVEFloat] in {
3673    if fms then {
3674      def : Pat<(VTI.Vec (fma (fneg m1), m2, add)),
3675                (Inst $add, $m1, $m2)>;
3676      def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
3677                                  (VTI.Vec (fma (fneg m1), m2, add)),
3678                                  add)),
3679                (Inst $add, $m1, $m2, ARMVCCThen, $pred, zero_reg)>;
3680      def : Pat<(VTI.Vec (pred_int (fneg m1), m2, add, pred)),
3681                (Inst $add, $m1, $m2, ARMVCCThen, $pred, zero_reg)>;
3682      def : Pat<(VTI.Vec (pred_int m1, (fneg m2), add, pred)),
3683                (Inst $add, $m1, $m2, ARMVCCThen, $pred, zero_reg)>;
3684    } else {
3685      def : Pat<(VTI.Vec (fma m1, m2, add)),
3686                (Inst $add, $m1, $m2)>;
3687      def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
3688                                  (VTI.Vec (fma m1, m2, add)),
3689                                  add)),
3690                (Inst $add, $m1, $m2, ARMVCCThen, $pred, zero_reg)>;
3691      def : Pat<(VTI.Vec (pred_int m1, m2, add, pred)),
3692                (Inst $add, $m1, $m2, ARMVCCThen, $pred, zero_reg)>;
3693    }
3694  }
3695}
3696
3697defm MVE_VFMAf32 : MVE_VFMA_fp_multi<"vfma", 0, MVE_v4f32>;
3698defm MVE_VFMAf16 : MVE_VFMA_fp_multi<"vfma", 0, MVE_v8f16>;
3699defm MVE_VFMSf32 : MVE_VFMA_fp_multi<"vfms", 1, MVE_v4f32>;
3700defm MVE_VFMSf16 : MVE_VFMA_fp_multi<"vfms", 1, MVE_v8f16>;
3701
3702multiclass MVE_VADDSUB_fp_m<string iname, bit bit_21, MVEVectorVTInfo VTI,
3703                            SDNode Op, Intrinsic PredInt, SDPatternOperator IdentityVec> {
3704  def "" : MVE_VADDSUBFMA_fp<iname, VTI.Suffix, VTI.Size, 0, 1, bit_21> {
3705    let validForTailPredication = 1;
3706  }
3707  defvar Inst = !cast<Instruction>(NAME);
3708
3709  let Predicates = [HasMVEFloat] in {
3710    defm : MVE_TwoOpPattern<VTI, Op, PredInt, (? ), !cast<Instruction>(NAME), IdentityVec>;
3711  }
3712}
3713
3714multiclass MVE_VADD_fp_m<MVEVectorVTInfo VTI, SDPatternOperator IdentityVec>
3715  : MVE_VADDSUB_fp_m<"vadd", 0, VTI, fadd, int_arm_mve_add_predicated, IdentityVec>;
3716multiclass MVE_VSUB_fp_m<MVEVectorVTInfo VTI, SDPatternOperator IdentityVec>
3717  : MVE_VADDSUB_fp_m<"vsub", 1, VTI, fsub, int_arm_mve_sub_predicated, IdentityVec>;
3718
3719def ARMimmMinusZeroF: PatLeaf<(bitconvert (v4i32 (ARMvmovImm (i32 1664))))>; // -0.0 float
3720def ARMimmMinusZeroH: PatLeaf<(bitconvert (v8i16 (ARMvmovImm (i32 2688))))>; // -0.0 half
3721
3722defm MVE_VADDf32 : MVE_VADD_fp_m<MVE_v4f32, ARMimmMinusZeroF>;
3723defm MVE_VADDf16 : MVE_VADD_fp_m<MVE_v8f16, ARMimmMinusZeroH>;
3724
3725defm MVE_VSUBf32 : MVE_VSUB_fp_m<MVE_v4f32, ARMimmAllZerosV>;
3726defm MVE_VSUBf16 : MVE_VSUB_fp_m<MVE_v8f16, ARMimmAllZerosV>;
3727
3728class MVE_VCADD<string suffix, bits<2> size, string cstr="">
3729  : MVEFloatArithNeon<"vcadd", suffix, size{1}, (outs MQPR:$Qd),
3730                         (ins MQPR:$Qn, MQPR:$Qm, complexrotateopodd:$rot),
3731                         "$Qd, $Qn, $Qm, $rot", vpred_r, cstr, size, []> {
3732  bits<4> Qd;
3733  bits<4> Qn;
3734  bit rot;
3735
3736  let Inst{28} = 0b1;
3737  let Inst{25} = 0b0;
3738  let Inst{24} = rot;
3739  let Inst{23} = 0b1;
3740  let Inst{22} = Qd{3};
3741  let Inst{21} = 0b0;
3742  let Inst{19-17} = Qn{2-0};
3743  let Inst{15-13} = Qd{2-0};
3744  let Inst{12-8} = 0b01000;
3745  let Inst{7} = Qn{3};
3746  let Inst{4} = 0b0;
3747}
3748
3749multiclass MVE_VCADD_m<MVEVectorVTInfo VTI, string cstr=""> {
3750  def "" : MVE_VCADD<VTI.Suffix, VTI.Size, cstr>;
3751  defvar Inst = !cast<Instruction>(NAME);
3752
3753  let Predicates = [HasMVEFloat] in {
3754    def : Pat<(VTI.Vec (int_arm_mve_vcaddq (i32 1),
3755                            imm:$rot, (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
3756              (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
3757                             imm:$rot))>;
3758
3759    def : Pat<(VTI.Vec (int_arm_mve_vcaddq_predicated (i32 1),
3760                            imm:$rot, (VTI.Vec MQPR:$inactive),
3761                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
3762                            (VTI.Pred VCCR:$mask))),
3763              (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
3764                             imm:$rot, ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg,
3765                             (VTI.Vec MQPR:$inactive)))>;
3766
3767  }
3768}
3769
3770defm MVE_VCADDf16 : MVE_VCADD_m<MVE_v8f16>;
3771defm MVE_VCADDf32 : MVE_VCADD_m<MVE_v4f32, "@earlyclobber $Qd">;
3772
3773class MVE_VABD_fp<string suffix, bits<2> size>
3774  : MVE_float<"vabd", suffix, (outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm),
3775              "$Qd, $Qn, $Qm", vpred_r, "", size> {
3776  bits<4> Qd;
3777  bits<4> Qn;
3778
3779  let Inst{28} = 0b1;
3780  let Inst{25-23} = 0b110;
3781  let Inst{22} = Qd{3};
3782  let Inst{21} = 0b1;
3783  let Inst{20} = size{0};
3784  let Inst{19-17} = Qn{2-0};
3785  let Inst{16} = 0b0;
3786  let Inst{15-13} = Qd{2-0};
3787  let Inst{11-8} = 0b1101;
3788  let Inst{7} = Qn{3};
3789  let Inst{4} = 0b0;
3790  let validForTailPredication = 1;
3791}
3792
3793multiclass MVE_VABDT_fp_m<MVEVectorVTInfo VTI,
3794                            Intrinsic unpred_int, Intrinsic pred_int> {
3795  def "" : MVE_VABD_fp<VTI.Suffix, VTI.Size>;
3796  defvar Inst = !cast<Instruction>(NAME);
3797
3798  let Predicates = [HasMVEFloat] in {
3799    def : Pat<(VTI.Vec (unpred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
3800                            (i32 0))),
3801              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
3802    def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
3803                            (i32 0), (VTI.Pred VCCR:$mask),
3804                            (VTI.Vec MQPR:$inactive))),
3805              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
3806                             ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg,
3807                             (VTI.Vec MQPR:$inactive)))>;
3808  }
3809}
3810
3811multiclass MVE_VABD_fp_m<MVEVectorVTInfo VTI>
3812  : MVE_VABDT_fp_m<VTI, int_arm_mve_vabd, int_arm_mve_abd_predicated>;
3813
3814defm MVE_VABDf32 : MVE_VABD_fp_m<MVE_v4f32>;
3815defm MVE_VABDf16 : MVE_VABD_fp_m<MVE_v8f16>;
3816
3817let Predicates = [HasMVEFloat] in {
3818  def : Pat<(v8f16 (fabs (fsub (v8f16 MQPR:$Qm), (v8f16 MQPR:$Qn)))),
3819            (MVE_VABDf16 MQPR:$Qm, MQPR:$Qn)>;
3820  def : Pat<(v4f32 (fabs (fsub (v4f32 MQPR:$Qm), (v4f32 MQPR:$Qn)))),
3821            (MVE_VABDf32 MQPR:$Qm, MQPR:$Qn)>;
3822}
3823
3824class MVE_VCVT_fix<string suffix, bit fsi, bit U, bit op,
3825                   Operand imm_operand_type>
3826  : MVE_float<"vcvt", suffix,
3827              (outs MQPR:$Qd), (ins MQPR:$Qm, imm_operand_type:$imm6),
3828              "$Qd, $Qm, $imm6", vpred_r, "", !if(fsi, 0b10, 0b01), []> {
3829  bits<4> Qd;
3830  bits<6> imm6;
3831
3832  let Inst{28} = U;
3833  let Inst{25-23} = 0b111;
3834  let Inst{22} = Qd{3};
3835  let Inst{21} = 0b1;
3836  let Inst{19-16} = imm6{3-0};
3837  let Inst{15-13} = Qd{2-0};
3838  let Inst{11-10} = 0b11;
3839  let Inst{9} = fsi;
3840  let Inst{8} = op;
3841  let Inst{7} = 0b0;
3842  let Inst{4} = 0b1;
3843
3844  let DecoderMethod = "DecodeMVEVCVTt1fp";
3845  let validForTailPredication = 1;
3846}
3847
3848class MVE_VCVT_imm_asmop<int Bits> : AsmOperandClass {
3849  let PredicateMethod = "isImmediate<1," # Bits # ">";
3850  let DiagnosticString =
3851      "MVE fixed-point immediate operand must be between 1 and " # Bits;
3852  let Name = "MVEVcvtImm" # Bits;
3853  let RenderMethod = "addImmOperands";
3854}
3855class MVE_VCVT_imm<int Bits>: Operand<i32> {
3856  let ParserMatchClass = MVE_VCVT_imm_asmop<Bits>;
3857  let EncoderMethod = "getNEONVcvtImm32OpValue";
3858  let DecoderMethod = "DecodeVCVTImmOperand";
3859}
3860
3861class MVE_VCVT_fix_f32<string suffix, bit U, bit op>
3862    : MVE_VCVT_fix<suffix, 0b1, U, op, MVE_VCVT_imm<32>> {
3863  let Inst{20} = imm6{4};
3864}
3865class MVE_VCVT_fix_f16<string suffix, bit U, bit op>
3866    : MVE_VCVT_fix<suffix, 0b0, U, op, MVE_VCVT_imm<16>> {
3867  let Inst{20} = 0b1;
3868}
3869
3870multiclass MVE_VCVT_fix_patterns<Instruction Inst, bit U, MVEVectorVTInfo DestVTI,
3871                                 MVEVectorVTInfo SrcVTI> {
3872  let Predicates = [HasMVEFloat] in {
3873    def : Pat<(DestVTI.Vec (int_arm_mve_vcvt_fix
3874                              (i32 U), (SrcVTI.Vec MQPR:$Qm), imm:$scale)),
3875              (DestVTI.Vec (Inst (SrcVTI.Vec MQPR:$Qm), imm:$scale))>;
3876    def : Pat<(DestVTI.Vec (int_arm_mve_vcvt_fix_predicated (i32 U),
3877                              (DestVTI.Vec MQPR:$inactive),
3878                              (SrcVTI.Vec MQPR:$Qm),
3879                              imm:$scale,
3880                              (DestVTI.Pred VCCR:$mask))),
3881              (DestVTI.Vec (Inst (SrcVTI.Vec MQPR:$Qm), imm:$scale,
3882                             ARMVCCThen, (DestVTI.Pred VCCR:$mask), zero_reg,
3883                             (DestVTI.Vec MQPR:$inactive)))>;
3884  }
3885}
3886
3887multiclass MVE_VCVT_fix_f32_m<bit U, bit op,
3888                              MVEVectorVTInfo DestVTI, MVEVectorVTInfo SrcVTI> {
3889  def "" : MVE_VCVT_fix_f32<DestVTI.Suffix#"."#SrcVTI.Suffix, U, op>;
3890  defm : MVE_VCVT_fix_patterns<!cast<Instruction>(NAME), U, DestVTI, SrcVTI>;
3891}
3892
3893multiclass MVE_VCVT_fix_f16_m<bit U, bit op,
3894                              MVEVectorVTInfo DestVTI, MVEVectorVTInfo SrcVTI> {
3895  def "" : MVE_VCVT_fix_f16<DestVTI.Suffix#"."#SrcVTI.Suffix, U, op>;
3896  defm : MVE_VCVT_fix_patterns<!cast<Instruction>(NAME), U, DestVTI, SrcVTI>;
3897}
3898
3899defm MVE_VCVTf16s16_fix : MVE_VCVT_fix_f16_m<0b0, 0b0, MVE_v8f16, MVE_v8s16>;
3900defm MVE_VCVTs16f16_fix : MVE_VCVT_fix_f16_m<0b0, 0b1, MVE_v8s16, MVE_v8f16>;
3901defm MVE_VCVTf16u16_fix : MVE_VCVT_fix_f16_m<0b1, 0b0, MVE_v8f16, MVE_v8u16>;
3902defm MVE_VCVTu16f16_fix : MVE_VCVT_fix_f16_m<0b1, 0b1, MVE_v8u16, MVE_v8f16>;
3903defm MVE_VCVTf32s32_fix : MVE_VCVT_fix_f32_m<0b0, 0b0, MVE_v4f32, MVE_v4s32>;
3904defm MVE_VCVTs32f32_fix : MVE_VCVT_fix_f32_m<0b0, 0b1, MVE_v4s32, MVE_v4f32>;
3905defm MVE_VCVTf32u32_fix : MVE_VCVT_fix_f32_m<0b1, 0b0, MVE_v4f32, MVE_v4u32>;
3906defm MVE_VCVTu32f32_fix : MVE_VCVT_fix_f32_m<0b1, 0b1, MVE_v4u32, MVE_v4f32>;
3907
3908class MVE_VCVT_fp_int_anpm<string suffix, bits<2> size, bit op, string anpm,
3909                bits<2> rm, list<dag> pattern=[]>
3910  : MVE_float<!strconcat("vcvt", anpm), suffix, (outs MQPR:$Qd),
3911              (ins MQPR:$Qm), "$Qd, $Qm", vpred_r, "", size, pattern> {
3912  bits<4> Qd;
3913
3914  let Inst{28} = 0b1;
3915  let Inst{25-23} = 0b111;
3916  let Inst{22} = Qd{3};
3917  let Inst{21-20} = 0b11;
3918  let Inst{19-18} = size;
3919  let Inst{17-16} = 0b11;
3920  let Inst{15-13} = Qd{2-0};
3921  let Inst{12-10} = 0b000;
3922  let Inst{9-8} = rm;
3923  let Inst{7} = op;
3924  let Inst{4} = 0b0;
3925  let validForTailPredication = 1;
3926}
3927
3928multiclass MVE_VCVT_fp_int_anpm_inner<MVEVectorVTInfo Int, MVEVectorVTInfo Flt,
3929                                      string anpm, bits<2> rm> {
3930  def "": MVE_VCVT_fp_int_anpm<Int.Suffix # "." # Flt.Suffix, Int.Size,
3931                               Int.Unsigned, anpm, rm>;
3932
3933  defvar Inst         = !cast<Instruction>(NAME);
3934  defvar IntrBaseName = "int_arm_mve_vcvt" # anpm;
3935  defvar UnpredIntr   = !cast<Intrinsic>(IntrBaseName);
3936  defvar PredIntr     = !cast<Intrinsic>(IntrBaseName # "_predicated");
3937
3938  let Predicates = [HasMVEFloat] in {
3939    def : Pat<(Int.Vec (UnpredIntr (i32 Int.Unsigned), (Flt.Vec MQPR:$in))),
3940              (Int.Vec (Inst (Flt.Vec MQPR:$in)))>;
3941
3942    def : Pat<(Int.Vec (PredIntr (i32 Int.Unsigned), (Int.Vec MQPR:$inactive),
3943                                 (Flt.Vec MQPR:$in), (Flt.Pred VCCR:$pred))),
3944              (Int.Vec (Inst (Flt.Vec MQPR:$in), ARMVCCThen,
3945                             (Flt.Pred VCCR:$pred), zero_reg, (Int.Vec MQPR:$inactive)))>;
3946  }
3947}
3948
3949multiclass MVE_VCVT_fp_int_anpm_outer<MVEVectorVTInfo Int,
3950                                      MVEVectorVTInfo Flt> {
3951  defm a : MVE_VCVT_fp_int_anpm_inner<Int, Flt, "a", 0b00>;
3952  defm n : MVE_VCVT_fp_int_anpm_inner<Int, Flt, "n", 0b01>;
3953  defm p : MVE_VCVT_fp_int_anpm_inner<Int, Flt, "p", 0b10>;
3954  defm m : MVE_VCVT_fp_int_anpm_inner<Int, Flt, "m", 0b11>;
3955}
3956
3957// This defines instructions such as MVE_VCVTu16f16a, with an explicit
3958// rounding-mode suffix on the mnemonic. The class below will define
3959// the bare MVE_VCVTu16f16 (with implied rounding toward zero).
3960defm MVE_VCVTs16f16 : MVE_VCVT_fp_int_anpm_outer<MVE_v8s16, MVE_v8f16>;
3961defm MVE_VCVTu16f16 : MVE_VCVT_fp_int_anpm_outer<MVE_v8u16, MVE_v8f16>;
3962defm MVE_VCVTs32f32 : MVE_VCVT_fp_int_anpm_outer<MVE_v4s32, MVE_v4f32>;
3963defm MVE_VCVTu32f32 : MVE_VCVT_fp_int_anpm_outer<MVE_v4u32, MVE_v4f32>;
3964
3965class MVE_VCVT_fp_int<string suffix, bits<2> size, bit toint, bit unsigned,
3966                      list<dag> pattern=[]>
3967  : MVE_float<"vcvt", suffix, (outs MQPR:$Qd),
3968              (ins MQPR:$Qm), "$Qd, $Qm", vpred_r, "", size, pattern> {
3969  bits<4> Qd;
3970
3971  let Inst{28} = 0b1;
3972  let Inst{25-23} = 0b111;
3973  let Inst{22} = Qd{3};
3974  let Inst{21-20} = 0b11;
3975  let Inst{19-18} = size;
3976  let Inst{17-16} = 0b11;
3977  let Inst{15-13} = Qd{2-0};
3978  let Inst{12-9} = 0b0011;
3979  let Inst{8} = toint;
3980  let Inst{7} = unsigned;
3981  let Inst{4} = 0b0;
3982  let validForTailPredication = 1;
3983}
3984
3985multiclass MVE_VCVT_fp_int_m<MVEVectorVTInfo Dest, MVEVectorVTInfo Src,
3986                             SDNode unpred_op> {
3987  defvar Unsigned = !or(!eq(Dest.SuffixLetter,"u"), !eq(Src.SuffixLetter,"u"));
3988  defvar ToInt = !eq(Src.SuffixLetter,"f");
3989
3990  def "" : MVE_VCVT_fp_int<Dest.Suffix # "." # Src.Suffix, Dest.Size,
3991                           ToInt, Unsigned>;
3992  defvar Inst = !cast<Instruction>(NAME);
3993
3994  let Predicates = [HasMVEFloat] in {
3995    def : Pat<(Dest.Vec (unpred_op (Src.Vec MQPR:$src))),
3996              (Dest.Vec (Inst (Src.Vec MQPR:$src)))>;
3997    def : Pat<(Dest.Vec (int_arm_mve_vcvt_fp_int_predicated
3998                             (Src.Vec MQPR:$src), (i32 Unsigned),
3999                             (Src.Pred VCCR:$mask), (Dest.Vec MQPR:$inactive))),
4000              (Dest.Vec (Inst (Src.Vec MQPR:$src), ARMVCCThen,
4001                              (Src.Pred VCCR:$mask), zero_reg,
4002                              (Dest.Vec MQPR:$inactive)))>;
4003  }
4004}
4005// The unsuffixed VCVT for float->int implicitly rounds toward zero,
4006// which I reflect here in the llvm instruction names
4007defm MVE_VCVTs16f16z : MVE_VCVT_fp_int_m<MVE_v8s16, MVE_v8f16, fp_to_sint>;
4008defm MVE_VCVTu16f16z : MVE_VCVT_fp_int_m<MVE_v8u16, MVE_v8f16, fp_to_uint>;
4009defm MVE_VCVTs32f32z : MVE_VCVT_fp_int_m<MVE_v4s32, MVE_v4f32, fp_to_sint>;
4010defm MVE_VCVTu32f32z : MVE_VCVT_fp_int_m<MVE_v4u32, MVE_v4f32, fp_to_uint>;
4011// Whereas VCVT for int->float rounds to nearest
4012defm MVE_VCVTf16s16n : MVE_VCVT_fp_int_m<MVE_v8f16, MVE_v8s16, sint_to_fp>;
4013defm MVE_VCVTf16u16n : MVE_VCVT_fp_int_m<MVE_v8f16, MVE_v8u16, uint_to_fp>;
4014defm MVE_VCVTf32s32n : MVE_VCVT_fp_int_m<MVE_v4f32, MVE_v4s32, sint_to_fp>;
4015defm MVE_VCVTf32u32n : MVE_VCVT_fp_int_m<MVE_v4f32, MVE_v4u32, uint_to_fp>;
4016
4017let Predicates = [HasMVEFloat] in {
4018  def : Pat<(v4i32 (fp_to_sint_sat v4f32:$src, i32)),
4019            (MVE_VCVTs32f32z v4f32:$src)>;
4020  def : Pat<(v4i32 (fp_to_uint_sat v4f32:$src, i32)),
4021            (MVE_VCVTu32f32z v4f32:$src)>;
4022  def : Pat<(v8i16 (fp_to_sint_sat v8f16:$src, i16)),
4023            (MVE_VCVTs16f16z v8f16:$src)>;
4024  def : Pat<(v8i16 (fp_to_uint_sat v8f16:$src, i16)),
4025            (MVE_VCVTu16f16z v8f16:$src)>;
4026}
4027
4028class MVE_VABSNEG_fp<string iname, string suffix, bits<2> size, bit negate,
4029                   list<dag> pattern=[]>
4030  : MVE_float<iname, suffix, (outs MQPR:$Qd),
4031              (ins MQPR:$Qm), "$Qd, $Qm", vpred_r, "", size, pattern> {
4032  bits<4> Qd;
4033
4034  let Inst{28} = 0b1;
4035  let Inst{25-23} = 0b111;
4036  let Inst{22} = Qd{3};
4037  let Inst{21-20} = 0b11;
4038  let Inst{19-18} = size;
4039  let Inst{17-16} = 0b01;
4040  let Inst{15-13} = Qd{2-0};
4041  let Inst{11-8} = 0b0111;
4042  let Inst{7} = negate;
4043  let Inst{4} = 0b0;
4044  let validForTailPredication = 1;
4045}
4046
4047multiclass MVE_VABSNEG_fp_m<string iname, SDNode unpred_op, Intrinsic pred_int,
4048                            MVEVectorVTInfo VTI, bit opcode> {
4049  def "" : MVE_VABSNEG_fp<iname, VTI.Suffix, VTI.Size, opcode>;
4050  defvar Inst = !cast<Instruction>(NAME);
4051
4052  let Predicates = [HasMVEInt] in {
4053    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$v))),
4054              (VTI.Vec (Inst $v))>;
4055    def : Pat<(VTI.Vec (pred_int  (VTI.Vec MQPR:$v), (VTI.Pred VCCR:$mask),
4056                                  (VTI.Vec MQPR:$inactive))),
4057              (VTI.Vec (Inst $v, ARMVCCThen, $mask, zero_reg, $inactive))>;
4058  }
4059}
4060
4061defm MVE_VABSf16 : MVE_VABSNEG_fp_m<"vabs", fabs, int_arm_mve_abs_predicated,
4062                                    MVE_v8f16, 0>;
4063defm MVE_VABSf32 : MVE_VABSNEG_fp_m<"vabs", fabs, int_arm_mve_abs_predicated,
4064                                    MVE_v4f32, 0>;
4065defm MVE_VNEGf16 : MVE_VABSNEG_fp_m<"vneg", fneg, int_arm_mve_neg_predicated,
4066                                    MVE_v8f16, 1>;
4067defm MVE_VNEGf32 : MVE_VABSNEG_fp_m<"vneg", fneg, int_arm_mve_neg_predicated,
4068                                    MVE_v4f32, 1>;
4069
4070class MVE_VMAXMINNMA<string iname, string suffix, bits<2> size, bit bit_12,
4071                     list<dag> pattern=[]>
4072  : MVE_f<(outs MQPR:$Qd), (ins MQPR:$Qd_src, MQPR:$Qm),
4073          NoItinerary, iname, suffix, "$Qd, $Qm", vpred_n, "$Qd = $Qd_src",
4074          size, pattern> {
4075  bits<4> Qd;
4076  bits<4> Qm;
4077
4078  let Inst{28} = size{0};
4079  let Inst{25-23} = 0b100;
4080  let Inst{22} = Qd{3};
4081  let Inst{21-16} = 0b111111;
4082  let Inst{15-13} = Qd{2-0};
4083  let Inst{12} = bit_12;
4084  let Inst{11-6} = 0b111010;
4085  let Inst{5} = Qm{3};
4086  let Inst{4} = 0b0;
4087  let Inst{3-1} = Qm{2-0};
4088  let Inst{0} = 0b1;
4089
4090  let isCommutable = 1;
4091  let validForTailPredication = 1;
4092}
4093
4094multiclass MVE_VMAXMINNMA_m<string iname, MVEVectorVTInfo VTI,
4095                      SDNode unpred_op, Intrinsic pred_int,
4096                      bit bit_12> {
4097  def "" : MVE_VMAXMINNMA<iname, VTI.Suffix, VTI.Size, bit_12>;
4098  defvar Inst = !cast<Instruction>(NAME);
4099
4100  let Predicates = [HasMVEInt] in {
4101    // Unpredicated v(max|min)nma
4102    def : Pat<(VTI.Vec (unpred_op (fabs (VTI.Vec MQPR:$Qd)),
4103                                  (fabs (VTI.Vec MQPR:$Qm)))),
4104              (VTI.Vec (Inst (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm)))>;
4105
4106    // Predicated v(max|min)nma
4107    def : Pat<(VTI.Vec (pred_int (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm),
4108                            (VTI.Pred VCCR:$mask))),
4109              (VTI.Vec (Inst (VTI.Vec MQPR:$Qd), (VTI.Vec MQPR:$Qm),
4110                            ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg))>;
4111  }
4112}
4113
4114multiclass MVE_VMAXNMA<MVEVectorVTInfo VTI, bit bit_12>
4115  : MVE_VMAXMINNMA_m<"vmaxnma", VTI, fmaxnum, int_arm_mve_vmaxnma_predicated, bit_12>;
4116
4117defm MVE_VMAXNMAf32 : MVE_VMAXNMA<MVE_v4f32, 0b0>;
4118defm MVE_VMAXNMAf16 : MVE_VMAXNMA<MVE_v8f16, 0b0>;
4119
4120multiclass MVE_VMINNMA<MVEVectorVTInfo VTI, bit bit_12>
4121  : MVE_VMAXMINNMA_m<"vminnma", VTI, fminnum, int_arm_mve_vminnma_predicated, bit_12>;
4122
4123defm MVE_VMINNMAf32 : MVE_VMINNMA<MVE_v4f32, 0b1>;
4124defm MVE_VMINNMAf16 : MVE_VMINNMA<MVE_v8f16, 0b1>;
4125
4126// end of MVE Floating Point instructions
4127
4128// start of MVE compares
4129
4130class MVE_VCMPqq<string suffix, bit bit_28, bits<2> bits_21_20,
4131                 VCMPPredicateOperand predtype, bits<2> vecsize, list<dag> pattern=[]>
4132  : MVE_p<(outs VCCR:$P0), (ins MQPR:$Qn, MQPR:$Qm, predtype:$fc),
4133           NoItinerary, "vcmp", suffix, "$fc, $Qn, $Qm", vpred_n, "", vecsize, pattern> {
4134  // Base class for comparing two vector registers
4135  bits<3> fc;
4136  bits<4> Qn;
4137  bits<4> Qm;
4138
4139  let Inst{28} = bit_28;
4140  let Inst{25-22} = 0b1000;
4141  let Inst{21-20} = bits_21_20;
4142  let Inst{19-17} = Qn{2-0};
4143  let Inst{16-13} = 0b1000;
4144  let Inst{12} = fc{2};
4145  let Inst{11-8} = 0b1111;
4146  let Inst{7} = fc{0};
4147  let Inst{6} = 0b0;
4148  let Inst{5} = Qm{3};
4149  let Inst{4} = 0b0;
4150  let Inst{3-1} = Qm{2-0};
4151  let Inst{0} = fc{1};
4152
4153  let Constraints = "";
4154
4155  // We need a custom decoder method for these instructions because of
4156  // the output VCCR operand, which isn't encoded in the instruction
4157  // bits anywhere (there is only one choice for it) but has to be
4158  // included in the MC operands so that codegen will be able to track
4159  // its data flow between instructions, spill/reload it when
4160  // necessary, etc. There seems to be no way to get the Tablegen
4161  // decoder to emit an operand that isn't affected by any instruction
4162  // bit.
4163  let DecoderMethod = "DecodeMVEVCMP<false," # predtype.DecoderMethod # ">";
4164  let validForTailPredication = 1;
4165}
4166
4167class MVE_VCMPqqf<string suffix, bit size>
4168    : MVE_VCMPqq<suffix, size, 0b11, pred_basic_fp, !if(size, 0b01, 0b10)> {
4169  let Predicates = [HasMVEFloat];
4170}
4171
4172class MVE_VCMPqqi<string suffix, bits<2> size>
4173    : MVE_VCMPqq<suffix, 0b1, size, pred_basic_i, size> {
4174  let Inst{12} = 0b0;
4175  let Inst{0} = 0b0;
4176}
4177
4178class MVE_VCMPqqu<string suffix, bits<2> size>
4179    : MVE_VCMPqq<suffix, 0b1, size, pred_basic_u, size> {
4180  let Inst{12} = 0b0;
4181  let Inst{0} = 0b1;
4182}
4183
4184class MVE_VCMPqqs<string suffix, bits<2> size>
4185    : MVE_VCMPqq<suffix, 0b1, size, pred_basic_s, size> {
4186  let Inst{12} = 0b1;
4187}
4188
4189def MVE_VCMPf32 : MVE_VCMPqqf<"f32", 0b0>;
4190def MVE_VCMPf16 : MVE_VCMPqqf<"f16", 0b1>;
4191
4192def MVE_VCMPi8  : MVE_VCMPqqi<"i8",  0b00>;
4193def MVE_VCMPi16 : MVE_VCMPqqi<"i16", 0b01>;
4194def MVE_VCMPi32 : MVE_VCMPqqi<"i32", 0b10>;
4195
4196def MVE_VCMPu8  : MVE_VCMPqqu<"u8",  0b00>;
4197def MVE_VCMPu16 : MVE_VCMPqqu<"u16", 0b01>;
4198def MVE_VCMPu32 : MVE_VCMPqqu<"u32", 0b10>;
4199
4200def MVE_VCMPs8  : MVE_VCMPqqs<"s8",  0b00>;
4201def MVE_VCMPs16 : MVE_VCMPqqs<"s16", 0b01>;
4202def MVE_VCMPs32 : MVE_VCMPqqs<"s32", 0b10>;
4203
4204class MVE_VCMPqr<string suffix, bit bit_28, bits<2> bits_21_20,
4205                 VCMPPredicateOperand predtype, bits<2> vecsize, list<dag> pattern=[]>
4206  : MVE_p<(outs VCCR:$P0), (ins MQPR:$Qn, GPRwithZR:$Rm, predtype:$fc),
4207           NoItinerary, "vcmp", suffix, "$fc, $Qn, $Rm", vpred_n, "", vecsize, pattern> {
4208  // Base class for comparing a vector register with a scalar
4209  bits<3> fc;
4210  bits<4> Qn;
4211  bits<4> Rm;
4212
4213  let Inst{28} = bit_28;
4214  let Inst{25-22} = 0b1000;
4215  let Inst{21-20} = bits_21_20;
4216  let Inst{19-17} = Qn{2-0};
4217  let Inst{16-13} = 0b1000;
4218  let Inst{12} = fc{2};
4219  let Inst{11-8} = 0b1111;
4220  let Inst{7} = fc{0};
4221  let Inst{6} = 0b1;
4222  let Inst{5} = fc{1};
4223  let Inst{4} = 0b0;
4224  let Inst{3-0} = Rm{3-0};
4225
4226  let Constraints = "";
4227  // Custom decoder method, for the same reason as MVE_VCMPqq
4228  let DecoderMethod = "DecodeMVEVCMP<true," # predtype.DecoderMethod # ">";
4229  let validForTailPredication = 1;
4230}
4231
4232class MVE_VCMPqrf<string suffix, bit size>
4233    : MVE_VCMPqr<suffix, size, 0b11, pred_basic_fp, !if(size, 0b01, 0b10)> {
4234  let Predicates = [HasMVEFloat];
4235}
4236
4237class MVE_VCMPqri<string suffix, bits<2> size>
4238    : MVE_VCMPqr<suffix, 0b1, size, pred_basic_i, size> {
4239  let Inst{12} = 0b0;
4240  let Inst{5} = 0b0;
4241}
4242
4243class MVE_VCMPqru<string suffix, bits<2> size>
4244    : MVE_VCMPqr<suffix, 0b1, size, pred_basic_u, size> {
4245  let Inst{12} = 0b0;
4246  let Inst{5} = 0b1;
4247}
4248
4249class MVE_VCMPqrs<string suffix, bits<2> size>
4250    : MVE_VCMPqr<suffix, 0b1, size, pred_basic_s, size> {
4251  let Inst{12} = 0b1;
4252}
4253
4254def MVE_VCMPf32r : MVE_VCMPqrf<"f32", 0b0>;
4255def MVE_VCMPf16r : MVE_VCMPqrf<"f16", 0b1>;
4256
4257def MVE_VCMPi8r  : MVE_VCMPqri<"i8",  0b00>;
4258def MVE_VCMPi16r : MVE_VCMPqri<"i16", 0b01>;
4259def MVE_VCMPi32r : MVE_VCMPqri<"i32", 0b10>;
4260
4261def MVE_VCMPu8r  : MVE_VCMPqru<"u8",  0b00>;
4262def MVE_VCMPu16r : MVE_VCMPqru<"u16", 0b01>;
4263def MVE_VCMPu32r : MVE_VCMPqru<"u32", 0b10>;
4264
4265def MVE_VCMPs8r  : MVE_VCMPqrs<"s8",  0b00>;
4266def MVE_VCMPs16r : MVE_VCMPqrs<"s16", 0b01>;
4267def MVE_VCMPs32r : MVE_VCMPqrs<"s32", 0b10>;
4268
4269multiclass unpred_vcmp_z<string suffix, PatLeaf fc> {
4270  def i8  : Pat<(v16i1 (ARMvcmpz (v16i8 MQPR:$v1), fc)),
4271                (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8r") (v16i8 MQPR:$v1), ZR, fc))>;
4272  def i16 : Pat<(v8i1 (ARMvcmpz (v8i16 MQPR:$v1), fc)),
4273                (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16r") (v8i16 MQPR:$v1), ZR, fc))>;
4274  def i32 : Pat<(v4i1 (ARMvcmpz (v4i32 MQPR:$v1), fc)),
4275                (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32r") (v4i32 MQPR:$v1), ZR, fc))>;
4276
4277  def : Pat<(v16i1 (and (v16i1 VCCR:$p1), (v16i1 (ARMvcmpz (v16i8 MQPR:$v1), fc)))),
4278            (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8r") (v16i8 MQPR:$v1), ZR, fc, ARMVCCThen, VCCR:$p1, zero_reg))>;
4279  def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmpz (v8i16 MQPR:$v1), fc)))),
4280            (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16r") (v8i16 MQPR:$v1), ZR, fc, ARMVCCThen, VCCR:$p1, zero_reg))>;
4281  def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmpz (v4i32 MQPR:$v1), fc)))),
4282            (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32r") (v4i32 MQPR:$v1), ZR, fc, ARMVCCThen, VCCR:$p1, zero_reg))>;
4283}
4284
4285multiclass unpred_vcmp_r<string suffix, PatLeaf fc> {
4286  def i8  : Pat<(v16i1 (ARMvcmp (v16i8 MQPR:$v1), (v16i8 MQPR:$v2), fc)),
4287                (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8") (v16i8 MQPR:$v1), (v16i8 MQPR:$v2), fc))>;
4288  def i16 : Pat<(v8i1 (ARMvcmp (v8i16 MQPR:$v1), (v8i16 MQPR:$v2), fc)),
4289                (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16") (v8i16 MQPR:$v1), (v8i16 MQPR:$v2), fc))>;
4290  def i32 : Pat<(v4i1 (ARMvcmp (v4i32 MQPR:$v1), (v4i32 MQPR:$v2), fc)),
4291                (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32") (v4i32 MQPR:$v1), (v4i32 MQPR:$v2), fc))>;
4292
4293  def i8r  : Pat<(v16i1 (ARMvcmp (v16i8 MQPR:$v1), (v16i8 (ARMvdup rGPR:$v2)), fc)),
4294                 (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8r") (v16i8 MQPR:$v1), (i32 rGPR:$v2), fc))>;
4295  def i16r : Pat<(v8i1 (ARMvcmp (v8i16 MQPR:$v1), (v8i16 (ARMvdup rGPR:$v2)), fc)),
4296                 (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16r") (v8i16 MQPR:$v1), (i32 rGPR:$v2), fc))>;
4297  def i32r : Pat<(v4i1 (ARMvcmp (v4i32 MQPR:$v1), (v4i32 (ARMvdup rGPR:$v2)), fc)),
4298                 (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32r") (v4i32 MQPR:$v1), (i32 rGPR:$v2), fc))>;
4299
4300  def : Pat<(v16i1 (and (v16i1 VCCR:$p1), (v16i1 (ARMvcmp (v16i8 MQPR:$v1), (v16i8 MQPR:$v2), fc)))),
4301            (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8") (v16i8 MQPR:$v1), (v16i8 MQPR:$v2), fc, ARMVCCThen, VCCR:$p1, zero_reg))>;
4302  def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmp (v8i16 MQPR:$v1), (v8i16 MQPR:$v2), fc)))),
4303            (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16") (v8i16 MQPR:$v1), (v8i16 MQPR:$v2), fc, ARMVCCThen, VCCR:$p1, zero_reg))>;
4304  def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmp (v4i32 MQPR:$v1), (v4i32 MQPR:$v2), fc)))),
4305            (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32") (v4i32 MQPR:$v1), (v4i32 MQPR:$v2), fc, ARMVCCThen, VCCR:$p1, zero_reg))>;
4306
4307  def : Pat<(v16i1 (and (v16i1 VCCR:$p1), (v16i1 (ARMvcmp (v16i8 MQPR:$v1), (v16i8 (ARMvdup rGPR:$v2)), fc)))),
4308            (v16i1 (!cast<Instruction>("MVE_VCMP"#suffix#"8r") (v16i8 MQPR:$v1), (i32 rGPR:$v2), fc, ARMVCCThen, VCCR:$p1, zero_reg))>;
4309  def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmp (v8i16 MQPR:$v1), (v8i16 (ARMvdup rGPR:$v2)), fc)))),
4310            (v8i1 (!cast<Instruction>("MVE_VCMP"#suffix#"16r") (v8i16 MQPR:$v1), (i32 rGPR:$v2), fc, ARMVCCThen, VCCR:$p1, zero_reg))>;
4311  def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmp (v4i32 MQPR:$v1), (v4i32 (ARMvdup rGPR:$v2)), fc)))),
4312            (v4i1 (!cast<Instruction>("MVE_VCMP"#suffix#"32r") (v4i32 MQPR:$v1), (i32 rGPR:$v2), fc, ARMVCCThen, VCCR:$p1, zero_reg))>;
4313}
4314
4315multiclass unpred_vcmpf_z<PatLeaf fc> {
4316  def f16 : Pat<(v8i1 (ARMvcmpz (v8f16 MQPR:$v1), fc)),
4317                (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), ZR, fc))>;
4318  def f32 : Pat<(v4i1 (ARMvcmpz (v4f32 MQPR:$v1), fc)),
4319                (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), ZR, fc))>;
4320
4321  def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmpz (v8f16 MQPR:$v1), fc)))),
4322            (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), ZR, fc, ARMVCCThen, VCCR:$p1, zero_reg))>;
4323  def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmpz (v4f32 MQPR:$v1), fc)))),
4324            (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), ZR, fc, ARMVCCThen, VCCR:$p1, zero_reg))>;
4325}
4326
4327multiclass unpred_vcmpf_r<PatLeaf fc> {
4328  def : Pat<(v8i1 (ARMvcmp (v8f16 MQPR:$v1), (v8f16 MQPR:$v2), fc)),
4329            (v8i1 (MVE_VCMPf16 (v8f16 MQPR:$v1), (v8f16 MQPR:$v2), fc))>;
4330  def : Pat<(v4i1 (ARMvcmp (v4f32 MQPR:$v1), (v4f32 MQPR:$v2), fc)),
4331            (v4i1 (MVE_VCMPf32 (v4f32 MQPR:$v1), (v4f32 MQPR:$v2), fc))>;
4332
4333  def : Pat<(v8i1 (ARMvcmp (v8f16 MQPR:$v1), (v8f16 (ARMvdup rGPR:$v2)), fc)),
4334            (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), (i32 rGPR:$v2), fc))>;
4335  def : Pat<(v4i1 (ARMvcmp (v4f32 MQPR:$v1), (v4f32 (ARMvdup rGPR:$v2)), fc)),
4336            (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), (i32 rGPR:$v2), fc))>;
4337
4338  def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmp (v8f16 MQPR:$v1), (v8f16 MQPR:$v2), fc)))),
4339            (v8i1 (MVE_VCMPf16 (v8f16 MQPR:$v1), (v8f16 MQPR:$v2), fc, ARMVCCThen, VCCR:$p1, zero_reg))>;
4340  def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmp (v4f32 MQPR:$v1), (v4f32 MQPR:$v2), fc)))),
4341            (v4i1 (MVE_VCMPf32 (v4f32 MQPR:$v1), (v4f32 MQPR:$v2), fc, ARMVCCThen, VCCR:$p1, zero_reg))>;
4342
4343  def : Pat<(v8i1 (and (v8i1 VCCR:$p1), (v8i1 (ARMvcmp (v8f16 MQPR:$v1), (v8f16 (ARMvdup rGPR:$v2)), fc)))),
4344            (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), (i32 rGPR:$v2), fc, ARMVCCThen, VCCR:$p1, zero_reg))>;
4345  def : Pat<(v4i1 (and (v4i1 VCCR:$p1), (v4i1 (ARMvcmp (v4f32 MQPR:$v1), (v4f32 (ARMvdup rGPR:$v2)), fc)))),
4346            (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), (i32 rGPR:$v2), fc, ARMVCCThen, VCCR:$p1, zero_reg))>;
4347}
4348
4349let Predicates = [HasMVEInt] in {
4350  defm MVE_VCEQZ  : unpred_vcmp_z<"i", ARMCCeq>;
4351  defm MVE_VCNEZ  : unpred_vcmp_z<"i", ARMCCne>;
4352  defm MVE_VCGEZ  : unpred_vcmp_z<"s", ARMCCge>;
4353  defm MVE_VCLTZ  : unpred_vcmp_z<"s", ARMCClt>;
4354  defm MVE_VCGTZ  : unpred_vcmp_z<"s", ARMCCgt>;
4355  defm MVE_VCLEZ  : unpred_vcmp_z<"s", ARMCCle>;
4356  defm MVE_VCGTUZ : unpred_vcmp_z<"u", ARMCChi>;
4357  defm MVE_VCGEUZ : unpred_vcmp_z<"u", ARMCChs>;
4358
4359  defm MVE_VCEQ   : unpred_vcmp_r<"i", ARMCCeq>;
4360  defm MVE_VCNE   : unpred_vcmp_r<"i", ARMCCne>;
4361  defm MVE_VCGE   : unpred_vcmp_r<"s", ARMCCge>;
4362  defm MVE_VCLT   : unpred_vcmp_r<"s", ARMCClt>;
4363  defm MVE_VCGT   : unpred_vcmp_r<"s", ARMCCgt>;
4364  defm MVE_VCLE   : unpred_vcmp_r<"s", ARMCCle>;
4365  defm MVE_VCGTU  : unpred_vcmp_r<"u", ARMCChi>;
4366  defm MVE_VCGEU  : unpred_vcmp_r<"u", ARMCChs>;
4367}
4368
4369let Predicates = [HasMVEFloat] in {
4370  defm MVE_VFCEQZ  : unpred_vcmpf_z<ARMCCeq>;
4371  defm MVE_VFCNEZ  : unpred_vcmpf_z<ARMCCne>;
4372  defm MVE_VFCGEZ  : unpred_vcmpf_z<ARMCCge>;
4373  defm MVE_VFCLTZ  : unpred_vcmpf_z<ARMCClt>;
4374  defm MVE_VFCGTZ  : unpred_vcmpf_z<ARMCCgt>;
4375  defm MVE_VFCLEZ  : unpred_vcmpf_z<ARMCCle>;
4376
4377  defm MVE_VFCEQ   : unpred_vcmpf_r<ARMCCeq>;
4378  defm MVE_VFCNE   : unpred_vcmpf_r<ARMCCne>;
4379  defm MVE_VFCGE   : unpred_vcmpf_r<ARMCCge>;
4380  defm MVE_VFCLT   : unpred_vcmpf_r<ARMCClt>;
4381  defm MVE_VFCGT   : unpred_vcmpf_r<ARMCCgt>;
4382  defm MVE_VFCLE   : unpred_vcmpf_r<ARMCCle>;
4383}
4384
4385
4386// Extra "worst case" and/or/xor patterns, going into and out of GRP
4387multiclass two_predops<SDPatternOperator opnode, Instruction insn> {
4388  def v16i1 : Pat<(v16i1 (opnode (v16i1 VCCR:$p1), (v16i1 VCCR:$p2))),
4389                  (v16i1 (COPY_TO_REGCLASS
4390                           (insn (i32 (COPY_TO_REGCLASS (v16i1 VCCR:$p1), rGPR)),
4391                                 (i32 (COPY_TO_REGCLASS (v16i1 VCCR:$p2), rGPR))),
4392                           VCCR))>;
4393  def v8i1  : Pat<(v8i1 (opnode (v8i1 VCCR:$p1), (v8i1 VCCR:$p2))),
4394                  (v8i1 (COPY_TO_REGCLASS
4395                          (insn (i32 (COPY_TO_REGCLASS (v8i1 VCCR:$p1), rGPR)),
4396                                (i32 (COPY_TO_REGCLASS (v8i1 VCCR:$p2), rGPR))),
4397                          VCCR))>;
4398  def v4i1  : Pat<(v4i1 (opnode (v4i1 VCCR:$p1), (v4i1 VCCR:$p2))),
4399                  (v4i1 (COPY_TO_REGCLASS
4400                          (insn (i32 (COPY_TO_REGCLASS (v4i1 VCCR:$p1), rGPR)),
4401                                (i32 (COPY_TO_REGCLASS (v4i1 VCCR:$p2), rGPR))),
4402                          VCCR))>;
4403  def v2i1  : Pat<(v2i1 (opnode (v2i1 VCCR:$p1), (v2i1 VCCR:$p2))),
4404                  (v2i1 (COPY_TO_REGCLASS
4405                          (insn (i32 (COPY_TO_REGCLASS (v2i1 VCCR:$p1), rGPR)),
4406                                (i32 (COPY_TO_REGCLASS (v2i1 VCCR:$p2), rGPR))),
4407                          VCCR))>;
4408}
4409
4410let Predicates = [HasMVEInt] in {
4411  defm POR    : two_predops<or,  t2ORRrr>;
4412  defm PAND   : two_predops<and, t2ANDrr>;
4413  defm PEOR   : two_predops<xor, t2EORrr>;
4414}
4415
4416// Occasionally we need to cast between a i32 and a boolean vector, for
4417// example when moving between rGPR and VPR.P0 as part of predicate vector
4418// shuffles. We also sometimes need to cast between different predicate
4419// vector types (v4i1<>v8i1, etc.) also as part of lowering vector shuffles.
4420def predicate_cast : SDNode<"ARMISD::PREDICATE_CAST", SDTUnaryOp>;
4421
4422def load_align4 : PatFrag<(ops node:$ptr), (load node:$ptr), [{
4423  return cast<LoadSDNode>(N)->getAlign() >= 4;
4424}]>;
4425
4426let Predicates = [HasMVEInt] in {
4427  foreach VT = [ v2i1, v4i1, v8i1, v16i1 ] in {
4428    def : Pat<(i32 (predicate_cast (VT VCCR:$src))),
4429              (i32 (COPY_TO_REGCLASS (VT VCCR:$src), VCCR))>;
4430    def : Pat<(VT  (predicate_cast (i32 VCCR:$src))),
4431              (VT  (COPY_TO_REGCLASS (i32 VCCR:$src), VCCR))>;
4432
4433    foreach VT2 = [ v2i1, v4i1, v8i1, v16i1 ] in
4434      def : Pat<(VT  (predicate_cast (VT2 VCCR:$src))),
4435                (VT  (COPY_TO_REGCLASS (VT2 VCCR:$src), VCCR))>;
4436  }
4437
4438  // If we happen to be casting from a load we can convert that straight
4439  // into a predicate load, so long as the load is of the correct type.
4440  foreach VT = [ v2i1, v4i1, v8i1, v16i1 ] in {
4441    def : Pat<(VT (predicate_cast (i32 (load_align4 taddrmode_imm7<2>:$addr)))),
4442              (VT (VLDR_P0_off taddrmode_imm7<2>:$addr))>;
4443  }
4444
4445  // Here we match the specific SDNode type 'ARMVectorRegCastImpl'
4446  // rather than the more general 'ARMVectorRegCast' which would also
4447  // match some bitconverts. If we use the latter in cases where the
4448  // input and output types are the same, the bitconvert gets elided
4449  // and we end up generating a nonsense match of nothing.
4450
4451  foreach VT = [ v16i8, v8i16, v8f16, v4i32, v4f32, v2i64, v2f64 ] in
4452    foreach VT2 = [ v16i8, v8i16, v8f16, v4i32, v4f32, v2i64, v2f64 ] in
4453      def : Pat<(VT (ARMVectorRegCastImpl (VT2 MQPR:$src))),
4454                (VT MQPR:$src)>;
4455}
4456
4457// end of MVE compares
4458
4459// start of MVE_qDest_qSrc
4460
4461class MVE_qDest_qSrc<string iname, string suffix, dag oops, dag iops,
4462                     string ops, vpred_ops vpred, string cstr,
4463                     bits<2> vecsize, list<dag> pattern=[]>
4464  : MVE_p<oops, iops, NoItinerary, iname, suffix,
4465          ops, vpred, cstr, vecsize, pattern> {
4466  bits<4> Qd;
4467  bits<4> Qm;
4468
4469  let Inst{25-23} = 0b100;
4470  let Inst{22} = Qd{3};
4471  let Inst{15-13} = Qd{2-0};
4472  let Inst{11-9} = 0b111;
4473  let Inst{6} = 0b0;
4474  let Inst{5} = Qm{3};
4475  let Inst{4} = 0b0;
4476  let Inst{3-1} = Qm{2-0};
4477}
4478
4479class MVE_VQxDMLxDH<string iname, bit exch, bit round, bit subtract,
4480                    string suffix, bits<2> size, string cstr="",
4481                    list<dag> pattern=[]>
4482  : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
4483                   (ins MQPR:$Qd_src, MQPR:$Qn, MQPR:$Qm), "$Qd, $Qn, $Qm",
4484                   vpred_n, "$Qd = $Qd_src"#cstr, size, pattern> {
4485  bits<4> Qn;
4486
4487  let Inst{28} = subtract;
4488  let Inst{21-20} = size;
4489  let Inst{19-17} = Qn{2-0};
4490  let Inst{16} = 0b0;
4491  let Inst{12} = exch;
4492  let Inst{8} = 0b0;
4493  let Inst{7} = Qn{3};
4494  let Inst{0} = round;
4495}
4496
4497multiclass MVE_VQxDMLxDH_p<string iname, bit exch, bit round, bit subtract,
4498                           MVEVectorVTInfo VTI> {
4499  def "": MVE_VQxDMLxDH<iname, exch, round, subtract, VTI.Suffix, VTI.Size,
4500                        !if(!eq(VTI.LaneBits, 32), ",@earlyclobber $Qd", "")>;
4501  defvar Inst = !cast<Instruction>(NAME);
4502  defvar ConstParams = (? (i32 exch), (i32 round), (i32 subtract));
4503  defvar unpred_intr = int_arm_mve_vqdmlad;
4504  defvar pred_intr = int_arm_mve_vqdmlad_predicated;
4505
4506  def : Pat<(VTI.Vec !con((unpred_intr (VTI.Vec MQPR:$a), (VTI.Vec MQPR:$b),
4507                                       (VTI.Vec MQPR:$c)), ConstParams)),
4508            (VTI.Vec (Inst (VTI.Vec MQPR:$a), (VTI.Vec MQPR:$b),
4509                           (VTI.Vec MQPR:$c)))>;
4510  def : Pat<(VTI.Vec !con((pred_intr (VTI.Vec MQPR:$a), (VTI.Vec MQPR:$b),
4511                                     (VTI.Vec MQPR:$c)), ConstParams,
4512                          (? (VTI.Pred VCCR:$pred)))),
4513            (VTI.Vec (Inst (VTI.Vec MQPR:$a), (VTI.Vec MQPR:$b),
4514                           (VTI.Vec MQPR:$c),
4515                           ARMVCCThen, (VTI.Pred VCCR:$pred), zero_reg))>;
4516}
4517
4518multiclass MVE_VQxDMLxDH_multi<string iname, bit exch,
4519                               bit round, bit subtract> {
4520  defm s8  : MVE_VQxDMLxDH_p<iname, exch, round, subtract, MVE_v16s8>;
4521  defm s16 : MVE_VQxDMLxDH_p<iname, exch, round, subtract, MVE_v8s16>;
4522  defm s32 : MVE_VQxDMLxDH_p<iname, exch, round, subtract, MVE_v4s32>;
4523}
4524
4525defm MVE_VQDMLADH   : MVE_VQxDMLxDH_multi<"vqdmladh",   0b0, 0b0, 0b0>;
4526defm MVE_VQDMLADHX  : MVE_VQxDMLxDH_multi<"vqdmladhx",  0b1, 0b0, 0b0>;
4527defm MVE_VQRDMLADH  : MVE_VQxDMLxDH_multi<"vqrdmladh",  0b0, 0b1, 0b0>;
4528defm MVE_VQRDMLADHX : MVE_VQxDMLxDH_multi<"vqrdmladhx", 0b1, 0b1, 0b0>;
4529defm MVE_VQDMLSDH   : MVE_VQxDMLxDH_multi<"vqdmlsdh",   0b0, 0b0, 0b1>;
4530defm MVE_VQDMLSDHX  : MVE_VQxDMLxDH_multi<"vqdmlsdhx",  0b1, 0b0, 0b1>;
4531defm MVE_VQRDMLSDH  : MVE_VQxDMLxDH_multi<"vqrdmlsdh",  0b0, 0b1, 0b1>;
4532defm MVE_VQRDMLSDHX : MVE_VQxDMLxDH_multi<"vqrdmlsdhx", 0b1, 0b1, 0b1>;
4533
4534class MVE_VCMUL<string iname, string suffix, bits<2> size, string cstr="">
4535  : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
4536                   (ins MQPR:$Qn, MQPR:$Qm, complexrotateop:$rot),
4537                   "$Qd, $Qn, $Qm, $rot", vpred_r, cstr, size,
4538                   []> {
4539  bits<4> Qn;
4540  bits<2> rot;
4541
4542  let Inst{28} = size{1};
4543  let Inst{21-20} = 0b11;
4544  let Inst{19-17} = Qn{2-0};
4545  let Inst{16} = 0b0;
4546  let Inst{12} = rot{1};
4547  let Inst{8} = 0b0;
4548  let Inst{7} = Qn{3};
4549  let Inst{0} = rot{0};
4550
4551  let Predicates = [HasMVEFloat];
4552}
4553
4554multiclass MVE_VCMUL_m<string iname, MVEVectorVTInfo VTI,
4555                       string cstr=""> {
4556  def "" : MVE_VCMUL<iname, VTI.Suffix, VTI.Size, cstr>;
4557  defvar Inst = !cast<Instruction>(NAME);
4558
4559  let Predicates = [HasMVEFloat] in {
4560    def : Pat<(VTI.Vec (int_arm_mve_vcmulq
4561                            imm:$rot, (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
4562              (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
4563                             imm:$rot))>;
4564
4565    def : Pat<(VTI.Vec (int_arm_mve_vcmulq_predicated
4566                            imm:$rot, (VTI.Vec MQPR:$inactive),
4567                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
4568                            (VTI.Pred VCCR:$mask))),
4569              (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
4570                             imm:$rot, ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg,
4571                             (VTI.Vec MQPR:$inactive)))>;
4572
4573  }
4574}
4575
4576defm MVE_VCMULf16 : MVE_VCMUL_m<"vcmul", MVE_v8f16>;
4577defm MVE_VCMULf32 : MVE_VCMUL_m<"vcmul", MVE_v4f32, "@earlyclobber $Qd">;
4578
4579class MVE_VMULL<string iname, string suffix, bit bit_28, bits<2> bits_21_20,
4580                bit T, string cstr, bits<2> vecsize, list<dag> pattern=[]>
4581  : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
4582                   (ins MQPR:$Qn, MQPR:$Qm), "$Qd, $Qn, $Qm",
4583                   vpred_r, cstr, vecsize, pattern> {
4584  bits<4> Qd;
4585  bits<4> Qn;
4586  bits<4> Qm;
4587
4588  let Inst{28} = bit_28;
4589  let Inst{21-20} = bits_21_20;
4590  let Inst{19-17} = Qn{2-0};
4591  let Inst{16} = 0b1;
4592  let Inst{12} = T;
4593  let Inst{8} = 0b0;
4594  let Inst{7} = Qn{3};
4595  let Inst{0} = 0b0;
4596  let validForTailPredication = 1;
4597  let doubleWidthResult = 1;
4598}
4599
4600multiclass MVE_VMULL_m<MVEVectorVTInfo VTI,
4601                       SDPatternOperator unpred_op, Intrinsic pred_int,
4602                       bit Top, bits<2> vecsize, string cstr=""> {
4603  def "" : MVE_VMULL<"vmull" # !if(Top, "t", "b"), VTI.Suffix, VTI.Unsigned,
4604                     VTI.Size, Top, cstr, vecsize>;
4605  defvar Inst = !cast<Instruction>(NAME);
4606
4607  let Predicates = [HasMVEInt] in {
4608    defvar uflag = !if(!eq(VTI.SuffixLetter, "p"), (?), (? (i32 VTI.Unsigned)));
4609
4610    // Unpredicated multiply
4611    def : Pat<(VTI.DblVec !con((unpred_op (VTI.Vec MQPR:$Qm),
4612                                          (VTI.Vec MQPR:$Qn)),
4613                               uflag, (? (i32 Top)))),
4614              (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
4615
4616    // Predicated multiply
4617    def : Pat<(VTI.DblVec !con((pred_int (VTI.Vec MQPR:$Qm),
4618                                         (VTI.Vec MQPR:$Qn)),
4619                               uflag, (? (i32 Top), (VTI.DblPred VCCR:$mask),
4620                                         (VTI.DblVec MQPR:$inactive)))),
4621              (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
4622                                ARMVCCThen, (VTI.DblPred VCCR:$mask), zero_reg,
4623                                (VTI.DblVec MQPR:$inactive)))>;
4624  }
4625}
4626
4627// For polynomial multiplies, the size bits take the unused value 0b11, and
4628// the unsigned bit switches to encoding the size.
4629
4630defm MVE_VMULLBs8  : MVE_VMULL_m<MVE_v16s8, int_arm_mve_vmull,
4631                                 int_arm_mve_mull_int_predicated, 0b0, 0b01>;
4632defm MVE_VMULLTs8  : MVE_VMULL_m<MVE_v16s8, int_arm_mve_vmull,
4633                                 int_arm_mve_mull_int_predicated, 0b1, 0b01>;
4634defm MVE_VMULLBs16 : MVE_VMULL_m<MVE_v8s16, int_arm_mve_vmull,
4635                                 int_arm_mve_mull_int_predicated, 0b0, 0b10>;
4636defm MVE_VMULLTs16 : MVE_VMULL_m<MVE_v8s16, int_arm_mve_vmull,
4637                                 int_arm_mve_mull_int_predicated, 0b1, 0b10>;
4638defm MVE_VMULLBs32 : MVE_VMULL_m<MVE_v4s32, int_arm_mve_vmull,
4639                                 int_arm_mve_mull_int_predicated, 0b0, 0b11,
4640                                 "@earlyclobber $Qd">;
4641defm MVE_VMULLTs32 : MVE_VMULL_m<MVE_v4s32, int_arm_mve_vmull,
4642                                 int_arm_mve_mull_int_predicated, 0b1, 0b11,
4643                                 "@earlyclobber $Qd">;
4644
4645defm MVE_VMULLBu8  : MVE_VMULL_m<MVE_v16u8, int_arm_mve_vmull,
4646                                 int_arm_mve_mull_int_predicated, 0b0, 0b01>;
4647defm MVE_VMULLTu8  : MVE_VMULL_m<MVE_v16u8, int_arm_mve_vmull,
4648                                 int_arm_mve_mull_int_predicated, 0b1, 0b01>;
4649defm MVE_VMULLBu16 : MVE_VMULL_m<MVE_v8u16, int_arm_mve_vmull,
4650                                 int_arm_mve_mull_int_predicated, 0b0, 0b10>;
4651defm MVE_VMULLTu16 : MVE_VMULL_m<MVE_v8u16, int_arm_mve_vmull,
4652                                 int_arm_mve_mull_int_predicated, 0b1, 0b10>;
4653defm MVE_VMULLBu32 : MVE_VMULL_m<MVE_v4u32, int_arm_mve_vmull,
4654                                 int_arm_mve_mull_int_predicated, 0b0, 0b11,
4655                                 "@earlyclobber $Qd">;
4656defm MVE_VMULLTu32 : MVE_VMULL_m<MVE_v4u32, int_arm_mve_vmull,
4657                                 int_arm_mve_mull_int_predicated, 0b1, 0b11,
4658                                 "@earlyclobber $Qd">;
4659
4660defm MVE_VMULLBp8  : MVE_VMULL_m<MVE_v16p8, int_arm_mve_vmull_poly,
4661                                 int_arm_mve_mull_poly_predicated, 0b0, 0b01>;
4662defm MVE_VMULLTp8  : MVE_VMULL_m<MVE_v16p8, int_arm_mve_vmull_poly,
4663                                 int_arm_mve_mull_poly_predicated, 0b1, 0b01>;
4664defm MVE_VMULLBp16 : MVE_VMULL_m<MVE_v8p16, int_arm_mve_vmull_poly,
4665                                 int_arm_mve_mull_poly_predicated, 0b0, 0b10>;
4666defm MVE_VMULLTp16 : MVE_VMULL_m<MVE_v8p16, int_arm_mve_vmull_poly,
4667                                 int_arm_mve_mull_poly_predicated, 0b1, 0b10>;
4668
4669let Predicates = [HasMVEInt] in {
4670  def : Pat<(v2i64 (ARMvmulls (v4i32 MQPR:$src1), (v4i32 MQPR:$src2))),
4671            (MVE_VMULLBs32 MQPR:$src1, MQPR:$src2)>;
4672  def : Pat<(v2i64 (ARMvmulls (v4i32 (ARMvrev64 (v4i32 MQPR:$src1))),
4673                              (v4i32 (ARMvrev64 (v4i32 MQPR:$src2))))),
4674            (MVE_VMULLTs32 MQPR:$src1, MQPR:$src2)>;
4675
4676  def : Pat<(mul (sext_inreg (v4i32 MQPR:$src1), v4i16),
4677                 (sext_inreg (v4i32 MQPR:$src2), v4i16)),
4678            (MVE_VMULLBs16 MQPR:$src1, MQPR:$src2)>;
4679  def : Pat<(mul (sext_inreg (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src1)))), v4i16),
4680                 (sext_inreg (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src2)))), v4i16)),
4681            (MVE_VMULLTs16 MQPR:$src1, MQPR:$src2)>;
4682
4683  def : Pat<(mul (sext_inreg (v8i16 MQPR:$src1), v8i8),
4684                 (sext_inreg (v8i16 MQPR:$src2), v8i8)),
4685            (MVE_VMULLBs8 MQPR:$src1, MQPR:$src2)>;
4686  def : Pat<(mul (sext_inreg (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src1)))), v8i8),
4687                 (sext_inreg (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src2)))), v8i8)),
4688            (MVE_VMULLTs8 MQPR:$src1, MQPR:$src2)>;
4689
4690  def : Pat<(v2i64 (ARMvmullu (v4i32 MQPR:$src1), (v4i32 MQPR:$src2))),
4691            (MVE_VMULLBu32 MQPR:$src1, MQPR:$src2)>;
4692  def : Pat<(v2i64 (ARMvmullu (v4i32 (ARMvrev64 (v4i32 MQPR:$src1))),
4693                              (v4i32 (ARMvrev64 (v4i32 MQPR:$src2))))),
4694            (MVE_VMULLTu32 MQPR:$src1, MQPR:$src2)>;
4695
4696  def : Pat<(mul (and (v4i32 MQPR:$src1), (v4i32 (ARMvmovImm (i32 0xCFF)))),
4697                 (and (v4i32 MQPR:$src2), (v4i32 (ARMvmovImm (i32 0xCFF))))),
4698            (MVE_VMULLBu16 MQPR:$src1, MQPR:$src2)>;
4699  def : Pat<(mul (and (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src1)))),
4700                      (v4i32 (ARMvmovImm (i32 0xCFF)))),
4701                 (and (v4i32 (ARMVectorRegCast (ARMvrev32 (v8i16 MQPR:$src2)))),
4702                      (v4i32 (ARMvmovImm (i32 0xCFF))))),
4703            (MVE_VMULLTu16 MQPR:$src1, MQPR:$src2)>;
4704
4705  def : Pat<(mul (ARMvbicImm (v8i16 MQPR:$src1), (i32 0xAFF)),
4706                 (ARMvbicImm (v8i16 MQPR:$src2), (i32 0xAFF))),
4707            (MVE_VMULLBu8 MQPR:$src1, MQPR:$src2)>;
4708  def : Pat<(mul (ARMvbicImm (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src1)))), (i32 0xAFF)),
4709                 (ARMvbicImm (v8i16 (ARMVectorRegCast (ARMvrev16 (v16i8 MQPR:$src2)))), (i32 0xAFF))),
4710            (MVE_VMULLTu8 MQPR:$src1, MQPR:$src2)>;
4711}
4712
4713class MVE_VxMULH<string iname, string suffix, bit U, bits<2> size, bit round,
4714                 list<dag> pattern=[]>
4715  : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
4716                   (ins MQPR:$Qn, MQPR:$Qm), "$Qd, $Qn, $Qm",
4717                   vpred_r, "", size, pattern> {
4718  bits<4> Qn;
4719
4720  let Inst{28} = U;
4721  let Inst{21-20} = size;
4722  let Inst{19-17} = Qn{2-0};
4723  let Inst{16} = 0b1;
4724  let Inst{12} = round;
4725  let Inst{8} = 0b0;
4726  let Inst{7} = Qn{3};
4727  let Inst{0} = 0b1;
4728  let validForTailPredication = 1;
4729}
4730
4731multiclass MVE_VxMULH_m<string iname, MVEVectorVTInfo VTI, SDPatternOperator unpred_op,
4732                        Intrinsic PredInt, bit round> {
4733  def "" : MVE_VxMULH<iname, VTI.Suffix, VTI.Unsigned, VTI.Size, round>;
4734  defvar Inst = !cast<Instruction>(NAME);
4735
4736  let Predicates = [HasMVEInt] in {
4737    if !eq(round, 0b0) then {
4738      defvar mulh = !if(VTI.Unsigned, mulhu, mulhs);
4739      defm : MVE_TwoOpPattern<VTI, mulh, PredInt, (? (i32 VTI.Unsigned)),
4740                              !cast<Instruction>(NAME)>;
4741    } else {
4742      // Predicated multiply returning high bits
4743      def : Pat<(VTI.Vec (PredInt (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
4744                              (i32 VTI.Unsigned), (VTI.Pred VCCR:$mask),
4745                              (VTI.Vec MQPR:$inactive))),
4746                (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
4747                              ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg,
4748                              (VTI.Vec MQPR:$inactive)))>;
4749    }
4750
4751    // Unpredicated intrinsic
4752    def : Pat<(VTI.Vec (unpred_op (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
4753                            (i32 VTI.Unsigned))),
4754              (VTI.Vec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
4755  }
4756}
4757
4758multiclass MVE_VMULT<string iname, MVEVectorVTInfo VTI, bit round>
4759  : MVE_VxMULH_m<iname, VTI, !if(round, int_arm_mve_vrmulh, int_arm_mve_vmulh),
4760                 !if(round, int_arm_mve_rmulh_predicated,
4761                            int_arm_mve_mulh_predicated),
4762                 round>;
4763
4764defm MVE_VMULHs8   : MVE_VMULT<"vmulh",  MVE_v16s8, 0b0>;
4765defm MVE_VMULHs16  : MVE_VMULT<"vmulh",  MVE_v8s16, 0b0>;
4766defm MVE_VMULHs32  : MVE_VMULT<"vmulh",  MVE_v4s32, 0b0>;
4767defm MVE_VMULHu8   : MVE_VMULT<"vmulh",  MVE_v16u8, 0b0>;
4768defm MVE_VMULHu16  : MVE_VMULT<"vmulh",  MVE_v8u16, 0b0>;
4769defm MVE_VMULHu32  : MVE_VMULT<"vmulh",  MVE_v4u32, 0b0>;
4770
4771defm MVE_VRMULHs8  : MVE_VMULT<"vrmulh", MVE_v16s8, 0b1>;
4772defm MVE_VRMULHs16 : MVE_VMULT<"vrmulh", MVE_v8s16, 0b1>;
4773defm MVE_VRMULHs32 : MVE_VMULT<"vrmulh", MVE_v4s32, 0b1>;
4774defm MVE_VRMULHu8  : MVE_VMULT<"vrmulh", MVE_v16u8, 0b1>;
4775defm MVE_VRMULHu16 : MVE_VMULT<"vrmulh", MVE_v8u16, 0b1>;
4776defm MVE_VRMULHu32 : MVE_VMULT<"vrmulh", MVE_v4u32, 0b1>;
4777
4778class MVE_VxMOVxN<string iname, string suffix, bit bit_28, bit bit_17,
4779                  bits<2> size, bit T, list<dag> pattern=[]>
4780  : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
4781                   (ins MQPR:$Qd_src, MQPR:$Qm), "$Qd, $Qm",
4782                   vpred_n, "$Qd = $Qd_src", !if(size, 0b10, 0b01), pattern> {
4783
4784  let Inst{28} = bit_28;
4785  let Inst{21-20} = 0b11;
4786  let Inst{19-18} = size;
4787  let Inst{17} = bit_17;
4788  let Inst{16} = 0b1;
4789  let Inst{12} = T;
4790  let Inst{8} = 0b0;
4791  let Inst{7} = !not(bit_17);
4792  let Inst{0} = 0b1;
4793  let validForTailPredication = 1;
4794  let retainsPreviousHalfElement = 1;
4795}
4796
4797multiclass MVE_VxMOVxN_halves<string iname, string suffix,
4798                              bit bit_28, bit bit_17, bits<2> size> {
4799  def bh : MVE_VxMOVxN<iname # "b", suffix, bit_28, bit_17, size, 0b0>;
4800  def th : MVE_VxMOVxN<iname # "t", suffix, bit_28, bit_17, size, 0b1>;
4801}
4802
4803defm MVE_VMOVNi16   : MVE_VxMOVxN_halves<"vmovn",   "i16", 0b1, 0b0, 0b00>;
4804defm MVE_VMOVNi32   : MVE_VxMOVxN_halves<"vmovn",   "i32", 0b1, 0b0, 0b01>;
4805defm MVE_VQMOVNs16  : MVE_VxMOVxN_halves<"vqmovn",  "s16", 0b0, 0b1, 0b00>;
4806defm MVE_VQMOVNs32  : MVE_VxMOVxN_halves<"vqmovn",  "s32", 0b0, 0b1, 0b01>;
4807defm MVE_VQMOVNu16  : MVE_VxMOVxN_halves<"vqmovn",  "u16", 0b1, 0b1, 0b00>;
4808defm MVE_VQMOVNu32  : MVE_VxMOVxN_halves<"vqmovn",  "u32", 0b1, 0b1, 0b01>;
4809defm MVE_VQMOVUNs16 : MVE_VxMOVxN_halves<"vqmovun", "s16", 0b0, 0b0, 0b00>;
4810defm MVE_VQMOVUNs32 : MVE_VxMOVxN_halves<"vqmovun", "s32", 0b0, 0b0, 0b01>;
4811
4812def MVEvmovn       : SDNode<"ARMISD::VMOVN", SDTARMVEXT>;
4813
4814multiclass MVE_VMOVN_p<Instruction Inst, bit top,
4815                       MVEVectorVTInfo VTI, MVEVectorVTInfo InVTI> {
4816  // Match the most obvious MVEvmovn(a,b,t), which overwrites the odd or even
4817  // lanes of a (depending on t) with the even lanes of b.
4818  def : Pat<(VTI.Vec (MVEvmovn (VTI.Vec MQPR:$Qd_src),
4819                               (VTI.Vec MQPR:$Qm), (i32 top))),
4820            (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src), (VTI.Vec MQPR:$Qm)))>;
4821
4822  if !not(top) then {
4823    // If we see MVEvmovn(a,ARMvrev(b),1), that wants to overwrite the odd
4824    // lanes of a with the odd lanes of b. In other words, the lanes we're
4825    // _keeping_ from a are the even ones. So we can flip it round and say that
4826    // this is the same as overwriting the even lanes of b with the even lanes
4827    // of a, i.e. it's a VMOVNB with the operands reversed.
4828    defvar vrev = !cast<SDNode>("ARMvrev" # InVTI.LaneBits);
4829    def : Pat<(VTI.Vec (MVEvmovn (VTI.Vec MQPR:$Qm),
4830                                 (VTI.Vec (vrev MQPR:$Qd_src)), (i32 1))),
4831              (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src), (VTI.Vec MQPR:$Qm)))>;
4832  }
4833
4834  // Match the IR intrinsic for a predicated VMOVN. This regards the Qm input
4835  // as having wider lanes that we're narrowing, instead of already-narrow
4836  // lanes that we're taking every other one of.
4837  def : Pat<(VTI.Vec (int_arm_mve_vmovn_predicated (VTI.Vec MQPR:$Qd_src),
4838                                  (InVTI.Vec MQPR:$Qm), (i32 top),
4839                                  (InVTI.Pred VCCR:$pred))),
4840            (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src),
4841                              (InVTI.Vec MQPR:$Qm),
4842                              ARMVCCThen, (InVTI.Pred VCCR:$pred), zero_reg))>;
4843}
4844
4845defm : MVE_VMOVN_p<MVE_VMOVNi32bh, 0, MVE_v8i16, MVE_v4i32>;
4846defm : MVE_VMOVN_p<MVE_VMOVNi32th, 1, MVE_v8i16, MVE_v4i32>;
4847defm : MVE_VMOVN_p<MVE_VMOVNi16bh, 0, MVE_v16i8, MVE_v8i16>;
4848defm : MVE_VMOVN_p<MVE_VMOVNi16th, 1, MVE_v16i8, MVE_v8i16>;
4849
4850multiclass MVE_VQMOVN_p<Instruction Inst, bit outU, bit inU, bit top,
4851                        MVEVectorVTInfo VTI, MVEVectorVTInfo InVTI> {
4852  def : Pat<(VTI.Vec (int_arm_mve_vqmovn (VTI.Vec MQPR:$Qd_src),
4853                                  (InVTI.Vec MQPR:$Qm),
4854                                  (i32 outU), (i32 inU), (i32 top))),
4855            (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src),
4856                              (InVTI.Vec MQPR:$Qm)))>;
4857
4858  def : Pat<(VTI.Vec (int_arm_mve_vqmovn_predicated (VTI.Vec MQPR:$Qd_src),
4859                                  (InVTI.Vec MQPR:$Qm),
4860                                  (i32 outU), (i32 inU), (i32 top),
4861                                  (InVTI.Pred VCCR:$pred))),
4862            (VTI.Vec (Inst (VTI.Vec MQPR:$Qd_src),
4863                              (InVTI.Vec MQPR:$Qm),
4864                              ARMVCCThen, (InVTI.Pred VCCR:$pred), zero_reg))>;
4865}
4866
4867defm : MVE_VQMOVN_p<MVE_VQMOVNs32bh,  0, 0, 0, MVE_v8i16, MVE_v4i32>;
4868defm : MVE_VQMOVN_p<MVE_VQMOVNs32th,  0, 0, 1, MVE_v8i16, MVE_v4i32>;
4869defm : MVE_VQMOVN_p<MVE_VQMOVNs16bh,  0, 0, 0, MVE_v16i8, MVE_v8i16>;
4870defm : MVE_VQMOVN_p<MVE_VQMOVNs16th,  0, 0, 1, MVE_v16i8, MVE_v8i16>;
4871defm : MVE_VQMOVN_p<MVE_VQMOVNu32bh,  1, 1, 0, MVE_v8i16, MVE_v4i32>;
4872defm : MVE_VQMOVN_p<MVE_VQMOVNu32th,  1, 1, 1, MVE_v8i16, MVE_v4i32>;
4873defm : MVE_VQMOVN_p<MVE_VQMOVNu16bh,  1, 1, 0, MVE_v16i8, MVE_v8i16>;
4874defm : MVE_VQMOVN_p<MVE_VQMOVNu16th,  1, 1, 1, MVE_v16i8, MVE_v8i16>;
4875defm : MVE_VQMOVN_p<MVE_VQMOVUNs32bh, 1, 0, 0, MVE_v8i16, MVE_v4i32>;
4876defm : MVE_VQMOVN_p<MVE_VQMOVUNs32th, 1, 0, 1, MVE_v8i16, MVE_v4i32>;
4877defm : MVE_VQMOVN_p<MVE_VQMOVUNs16bh, 1, 0, 0, MVE_v16i8, MVE_v8i16>;
4878defm : MVE_VQMOVN_p<MVE_VQMOVUNs16th, 1, 0, 1, MVE_v16i8, MVE_v8i16>;
4879
4880def SDTARMVMOVNQ : SDTypeProfile<1, 3, [SDTCisVec<0>, SDTCisSameAs<0, 1>,
4881                                        SDTCisVec<2>, SDTCisVT<3, i32>]>;
4882def MVEvqmovns   : SDNode<"ARMISD::VQMOVNs", SDTARMVMOVNQ>;
4883def MVEvqmovnu   : SDNode<"ARMISD::VQMOVNu", SDTARMVMOVNQ>;
4884
4885let Predicates = [HasMVEInt] in {
4886  def : Pat<(v8i16 (MVEvqmovns (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), (i32 0))),
4887            (v8i16 (MVE_VQMOVNs32bh (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm)))>;
4888  def : Pat<(v8i16 (MVEvqmovns (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), (i32 1))),
4889            (v8i16 (MVE_VQMOVNs32th (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm)))>;
4890  def : Pat<(v16i8 (MVEvqmovns (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), (i32 0))),
4891            (v16i8 (MVE_VQMOVNs16bh (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm)))>;
4892  def : Pat<(v16i8 (MVEvqmovns (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), (i32 1))),
4893            (v16i8 (MVE_VQMOVNs16th (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm)))>;
4894
4895  def : Pat<(v8i16 (MVEvqmovnu (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), (i32 0))),
4896            (v8i16 (MVE_VQMOVNu32bh (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm)))>;
4897  def : Pat<(v8i16 (MVEvqmovnu (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), (i32 1))),
4898            (v8i16 (MVE_VQMOVNu32th (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm)))>;
4899  def : Pat<(v16i8 (MVEvqmovnu (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), (i32 0))),
4900            (v16i8 (MVE_VQMOVNu16bh (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm)))>;
4901  def : Pat<(v16i8 (MVEvqmovnu (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), (i32 1))),
4902            (v16i8 (MVE_VQMOVNu16th (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm)))>;
4903
4904  def : Pat<(v8i16 (MVEvqmovns (v8i16 MQPR:$Qd_src), (v4i32 (ARMvshrsImm (v4i32 MQPR:$Qm), imm0_31:$imm)), (i32 0))),
4905            (v8i16 (MVE_VQSHRNbhs32 (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), imm0_31:$imm))>;
4906  def : Pat<(v16i8 (MVEvqmovns (v16i8 MQPR:$Qd_src), (v8i16 (ARMvshrsImm (v8i16 MQPR:$Qm), imm0_15:$imm)), (i32 0))),
4907            (v16i8 (MVE_VQSHRNbhs16 (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), imm0_15:$imm))>;
4908  def : Pat<(v8i16 (MVEvqmovns (v8i16 MQPR:$Qd_src), (v4i32 (ARMvshrsImm (v4i32 MQPR:$Qm), imm0_31:$imm)), (i32 1))),
4909            (v8i16 (MVE_VQSHRNths32 (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), imm0_31:$imm))>;
4910  def : Pat<(v16i8 (MVEvqmovns (v16i8 MQPR:$Qd_src), (v8i16 (ARMvshrsImm (v8i16 MQPR:$Qm), imm0_15:$imm)), (i32 1))),
4911            (v16i8 (MVE_VQSHRNths16 (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), imm0_15:$imm))>;
4912
4913  def : Pat<(v8i16 (MVEvqmovnu (v8i16 MQPR:$Qd_src), (v4i32 (ARMvshruImm (v4i32 MQPR:$Qm), imm0_31:$imm)), (i32 0))),
4914            (v8i16 (MVE_VQSHRNbhu32 (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), imm0_31:$imm))>;
4915  def : Pat<(v16i8 (MVEvqmovnu (v16i8 MQPR:$Qd_src), (v8i16 (ARMvshruImm (v8i16 MQPR:$Qm), imm0_15:$imm)), (i32 0))),
4916            (v16i8 (MVE_VQSHRNbhu16 (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), imm0_15:$imm))>;
4917  def : Pat<(v8i16 (MVEvqmovnu (v8i16 MQPR:$Qd_src), (v4i32 (ARMvshruImm (v4i32 MQPR:$Qm), imm0_31:$imm)), (i32 1))),
4918            (v8i16 (MVE_VQSHRNthu32 (v8i16 MQPR:$Qd_src), (v4i32 MQPR:$Qm), imm0_31:$imm))>;
4919  def : Pat<(v16i8 (MVEvqmovnu (v16i8 MQPR:$Qd_src), (v8i16 (ARMvshruImm (v8i16 MQPR:$Qm), imm0_15:$imm)), (i32 1))),
4920            (v16i8 (MVE_VQSHRNthu16 (v16i8 MQPR:$Qd_src), (v8i16 MQPR:$Qm), imm0_15:$imm))>;
4921}
4922
4923class MVE_VCVT_ff<string iname, string suffix, bit op, bit T,
4924                  dag iops_extra, vpred_ops vpred, string cstr>
4925  : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
4926                   !con(iops_extra, (ins MQPR:$Qm)), "$Qd, $Qm",
4927                   vpred, cstr, 0b10, []> {
4928  let Inst{28} = op;
4929  let Inst{21-16} = 0b111111;
4930  let Inst{12} = T;
4931  let Inst{8-7} = 0b00;
4932  let Inst{0} = 0b1;
4933
4934  let Predicates = [HasMVEFloat];
4935  let retainsPreviousHalfElement = 1;
4936}
4937
4938def SDTARMVCVTL    : SDTypeProfile<1, 2, [SDTCisVec<0>, SDTCisVec<1>,
4939                                         SDTCisVT<2, i32>]>;
4940def MVEvcvtn       : SDNode<"ARMISD::VCVTN", SDTARMVMOVNQ>;
4941def MVEvcvtl       : SDNode<"ARMISD::VCVTL", SDTARMVCVTL>;
4942
4943multiclass MVE_VCVT_f2h_m<string iname, int half> {
4944  def "": MVE_VCVT_ff<iname, "f16.f32", 0b0, half,
4945                      (ins MQPR:$Qd_src), vpred_n, "$Qd = $Qd_src">;
4946  defvar Inst = !cast<Instruction>(NAME);
4947
4948  let Predicates = [HasMVEFloat] in {
4949    def : Pat<(v8f16 (int_arm_mve_vcvt_narrow
4950                         (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm), (i32 half))),
4951              (v8f16 (Inst (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm)))>;
4952    def : Pat<(v8f16 (int_arm_mve_vcvt_narrow_predicated
4953                         (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm), (i32 half),
4954                         (v4i1 VCCR:$mask))),
4955              (v8f16 (Inst (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm),
4956                           ARMVCCThen, (v4i1 VCCR:$mask), zero_reg))>;
4957
4958    def : Pat<(v8f16 (MVEvcvtn (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm), (i32 half))),
4959              (v8f16 (Inst (v8f16 MQPR:$Qd_src), (v4f32 MQPR:$Qm)))>;
4960  }
4961}
4962
4963multiclass MVE_VCVT_h2f_m<string iname, int half> {
4964  def "": MVE_VCVT_ff<iname, "f32.f16", 0b1, half, (ins), vpred_r, "">;
4965  defvar Inst = !cast<Instruction>(NAME);
4966
4967  let Predicates = [HasMVEFloat] in {
4968    def : Pat<(v4f32 (int_arm_mve_vcvt_widen (v8f16 MQPR:$Qm), (i32 half))),
4969              (v4f32 (Inst (v8f16 MQPR:$Qm)))>;
4970    def : Pat<(v4f32 (int_arm_mve_vcvt_widen_predicated
4971                         (v4f32 MQPR:$inactive), (v8f16 MQPR:$Qm), (i32 half),
4972                         (v4i1 VCCR:$mask))),
4973              (v4f32 (Inst (v8f16 MQPR:$Qm), ARMVCCThen,
4974                           (v4i1 VCCR:$mask), zero_reg, (v4f32 MQPR:$inactive)))>;
4975
4976    def : Pat<(v4f32 (MVEvcvtl (v8f16 MQPR:$Qm), (i32 half))),
4977              (v4f32 (Inst (v8f16 MQPR:$Qm)))>;
4978  }
4979}
4980
4981defm MVE_VCVTf16f32bh : MVE_VCVT_f2h_m<"vcvtb", 0b0>;
4982defm MVE_VCVTf16f32th : MVE_VCVT_f2h_m<"vcvtt", 0b1>;
4983defm MVE_VCVTf32f16bh : MVE_VCVT_h2f_m<"vcvtb", 0b0>;
4984defm MVE_VCVTf32f16th : MVE_VCVT_h2f_m<"vcvtt", 0b1>;
4985
4986class MVE_VxCADD<string iname, string suffix, bits<2> size, bit halve,
4987                 string cstr="">
4988  : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
4989                   (ins MQPR:$Qn, MQPR:$Qm, complexrotateopodd:$rot),
4990                   "$Qd, $Qn, $Qm, $rot", vpred_r, cstr, size, []> {
4991  bits<4> Qn;
4992  bit rot;
4993
4994  let Inst{28} = halve;
4995  let Inst{21-20} = size;
4996  let Inst{19-17} = Qn{2-0};
4997  let Inst{16} = 0b0;
4998  let Inst{12} = rot;
4999  let Inst{8} = 0b1;
5000  let Inst{7} = Qn{3};
5001  let Inst{0} = 0b0;
5002}
5003
5004multiclass MVE_VxCADD_m<string iname, MVEVectorVTInfo VTI,
5005                        bit halve, string cstr=""> {
5006  def "" : MVE_VxCADD<iname, VTI.Suffix, VTI.Size, halve, cstr>;
5007  defvar Inst = !cast<Instruction>(NAME);
5008
5009  let Predicates = [HasMVEInt] in {
5010    def : Pat<(VTI.Vec (int_arm_mve_vcaddq halve,
5011                            imm:$rot, (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm))),
5012              (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
5013                             imm:$rot))>;
5014
5015    def : Pat<(VTI.Vec (int_arm_mve_vcaddq_predicated halve,
5016                            imm:$rot, (VTI.Vec MQPR:$inactive),
5017                            (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
5018                            (VTI.Pred VCCR:$mask))),
5019              (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (VTI.Vec MQPR:$Qm),
5020                             imm:$rot, ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg,
5021                             (VTI.Vec MQPR:$inactive)))>;
5022
5023  }
5024}
5025
5026defm MVE_VCADDi8   : MVE_VxCADD_m<"vcadd", MVE_v16i8, 0b1>;
5027defm MVE_VCADDi16  : MVE_VxCADD_m<"vcadd", MVE_v8i16, 0b1>;
5028defm MVE_VCADDi32  : MVE_VxCADD_m<"vcadd", MVE_v4i32, 0b1, "@earlyclobber $Qd">;
5029
5030defm MVE_VHCADDs8  : MVE_VxCADD_m<"vhcadd", MVE_v16s8, 0b0>;
5031defm MVE_VHCADDs16 : MVE_VxCADD_m<"vhcadd", MVE_v8s16, 0b0>;
5032defm MVE_VHCADDs32 : MVE_VxCADD_m<"vhcadd", MVE_v4s32, 0b0, "@earlyclobber $Qd">;
5033
5034class MVE_VADCSBC<string iname, bit I, bit subtract,
5035                  dag carryin, list<dag> pattern=[]>
5036  : MVE_qDest_qSrc<iname, "i32", (outs MQPR:$Qd, cl_FPSCR_NZCV:$carryout),
5037                   !con((ins MQPR:$Qn, MQPR:$Qm), carryin),
5038                   "$Qd, $Qn, $Qm", vpred_r, "", 0b10, pattern> {
5039  bits<4> Qn;
5040
5041  let Inst{28} = subtract;
5042  let Inst{21-20} = 0b11;
5043  let Inst{19-17} = Qn{2-0};
5044  let Inst{16} = 0b0;
5045  let Inst{12} = I;
5046  let Inst{8} = 0b1;
5047  let Inst{7} = Qn{3};
5048  let Inst{0} = 0b0;
5049
5050  // Custom decoder method in order to add the FPSCR operand(s), which
5051  // Tablegen won't do right
5052  let DecoderMethod = "DecodeMVEVADCInstruction";
5053}
5054
5055def MVE_VADC  : MVE_VADCSBC<"vadc",  0b0, 0b0, (ins cl_FPSCR_NZCV:$carryin)>;
5056def MVE_VADCI : MVE_VADCSBC<"vadci", 0b1, 0b0, (ins)>;
5057
5058def MVE_VSBC  : MVE_VADCSBC<"vsbc",  0b0, 0b1, (ins cl_FPSCR_NZCV:$carryin)>;
5059def MVE_VSBCI : MVE_VADCSBC<"vsbci", 0b1, 0b1, (ins)>;
5060
5061class MVE_VQDMULL<string iname, string suffix, bit size, bit T,
5062                  string cstr="", list<dag> pattern=[]>
5063  : MVE_qDest_qSrc<iname, suffix, (outs MQPR:$Qd),
5064                   (ins MQPR:$Qn, MQPR:$Qm), "$Qd, $Qn, $Qm",
5065                   vpred_r, cstr, !if(size, 0b10, 0b01), pattern> {
5066  bits<4> Qn;
5067
5068  let Inst{28} = size;
5069  let Inst{21-20} = 0b11;
5070  let Inst{19-17} = Qn{2-0};
5071  let Inst{16} = 0b0;
5072  let Inst{12} = T;
5073  let Inst{8} = 0b1;
5074  let Inst{7} = Qn{3};
5075  let Inst{0} = 0b1;
5076  let validForTailPredication = 1;
5077  let doubleWidthResult = 1;
5078}
5079
5080multiclass MVE_VQDMULL_m<string iname, MVEVectorVTInfo VTI, bit size, bit T,
5081                         string cstr> {
5082  def "" : MVE_VQDMULL<iname, VTI.Suffix, size, T, cstr>;
5083  defvar Inst = !cast<Instruction>(NAME);
5084
5085  let Predicates = [HasMVEInt] in {
5086    // Unpredicated saturating multiply
5087    def : Pat<(VTI.DblVec (int_arm_mve_vqdmull (VTI.Vec MQPR:$Qm),
5088                                               (VTI.Vec MQPR:$Qn), (i32 T))),
5089              (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn)))>;
5090    // Predicated saturating multiply
5091    def : Pat<(VTI.DblVec (int_arm_mve_vqdmull_predicated
5092                                    (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
5093                                    (i32 T), (VTI.DblPred VCCR:$mask),
5094                                    (VTI.DblVec MQPR:$inactive))),
5095              (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (VTI.Vec MQPR:$Qn),
5096                                ARMVCCThen, (VTI.DblPred VCCR:$mask), zero_reg,
5097                                (VTI.DblVec MQPR:$inactive)))>;
5098  }
5099}
5100
5101multiclass MVE_VQDMULL_halves<MVEVectorVTInfo VTI, bit size, string cstr=""> {
5102  defm bh : MVE_VQDMULL_m<"vqdmullb", VTI, size, 0b0, cstr>;
5103  defm th : MVE_VQDMULL_m<"vqdmullt", VTI, size, 0b1, cstr>;
5104}
5105
5106defm MVE_VQDMULLs16 : MVE_VQDMULL_halves<MVE_v8s16, 0b0>;
5107defm MVE_VQDMULLs32 : MVE_VQDMULL_halves<MVE_v4s32, 0b1, "@earlyclobber $Qd">;
5108
5109// end of mve_qDest_qSrc
5110
5111// start of mve_qDest_rSrc
5112
5113class MVE_qr_base<dag oops, dag iops, string iname, string suffix, string ops,
5114                  vpred_ops vpred, string cstr, bits<2> vecsize, list<dag> pattern=[]>
5115   : MVE_p<oops, iops, NoItinerary, iname, suffix, ops, vpred, cstr, vecsize, pattern> {
5116  bits<4> Qd;
5117  bits<4> Qn;
5118  bits<4> Rm;
5119
5120  let Inst{25-23} = 0b100;
5121  let Inst{22} = Qd{3};
5122  let Inst{19-17} = Qn{2-0};
5123  let Inst{15-13} = Qd{2-0};
5124  let Inst{11-9} = 0b111;
5125  let Inst{7} = Qn{3};
5126  let Inst{6} = 0b1;
5127  let Inst{4} = 0b0;
5128  let Inst{3-0} = Rm{3-0};
5129}
5130
5131class MVE_qDest_rSrc<string iname, string suffix, string cstr="", bits<2> vecsize, list<dag> pattern=[]>
5132  : MVE_qr_base<(outs MQPR:$Qd), (ins MQPR:$Qn, rGPR:$Rm),
5133                 iname, suffix, "$Qd, $Qn, $Rm", vpred_r, cstr,
5134                 vecsize, pattern>;
5135
5136class MVE_qDestSrc_rSrc<string iname, string suffix, bits<2> vecsize, list<dag> pattern=[]>
5137  : MVE_qr_base<(outs MQPR:$Qd), (ins MQPR:$Qd_src, MQPR:$Qn, rGPR:$Rm),
5138                 iname, suffix, "$Qd, $Qn, $Rm", vpred_n, "$Qd = $Qd_src",
5139                 vecsize, pattern>;
5140
5141class MVE_qDest_single_rSrc<string iname, string suffix, bits<2> vecsize, list<dag> pattern=[]>
5142  : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qd_src, rGPR:$Rm), NoItinerary, iname,
5143          suffix, "$Qd, $Rm", vpred_n, "$Qd = $Qd_src", vecsize, pattern> {
5144  bits<4> Qd;
5145  bits<4> Rm;
5146
5147  let Inst{22} = Qd{3};
5148  let Inst{15-13} = Qd{2-0};
5149  let Inst{3-0} = Rm{3-0};
5150}
5151
5152// Patterns for vector-scalar instructions with integer operands
5153multiclass MVE_vec_scalar_int_pat_m<Instruction inst, MVEVectorVTInfo VTI,
5154                                    SDPatternOperator unpred_op,
5155                                    SDPatternOperator pred_op,
5156                                    bit unpred_has_sign = 0,
5157                                    bit pred_has_sign = 0> {
5158  defvar UnpredSign = !if(unpred_has_sign, (? (i32 VTI.Unsigned)), (?));
5159  defvar PredSign = !if(pred_has_sign, (? (i32 VTI.Unsigned)), (?));
5160
5161  let Predicates = [HasMVEInt] in {
5162    // Unpredicated version
5163    def : Pat<(VTI.Vec !con((unpred_op (VTI.Vec MQPR:$Qm),
5164                                       (VTI.Vec (ARMvdup rGPR:$val))),
5165                            UnpredSign)),
5166              (VTI.Vec (inst (VTI.Vec MQPR:$Qm), (i32 rGPR:$val)))>;
5167    // Predicated version
5168    def : Pat<(VTI.Vec !con((pred_op (VTI.Vec MQPR:$Qm),
5169                                     (VTI.Vec (ARMvdup rGPR:$val))),
5170                            PredSign,
5171                            (pred_op (VTI.Pred VCCR:$mask),
5172                                     (VTI.Vec MQPR:$inactive)))),
5173              (VTI.Vec (inst (VTI.Vec MQPR:$Qm), (i32 rGPR:$val),
5174                             ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg,
5175                             (VTI.Vec MQPR:$inactive)))>;
5176  }
5177}
5178
5179class MVE_VADDSUB_qr<string iname, string suffix, bits<2> size,
5180                     bit bit_5, bit bit_12, bit bit_16, bit bit_28>
5181  : MVE_qDest_rSrc<iname, suffix, "", size> {
5182
5183  let Inst{28} = bit_28;
5184  let Inst{21-20} = size;
5185  let Inst{16} = bit_16;
5186  let Inst{12} = bit_12;
5187  let Inst{8} = 0b1;
5188  let Inst{5} = bit_5;
5189  let validForTailPredication = 1;
5190}
5191
5192// Vector-scalar add/sub
5193multiclass MVE_VADDSUB_qr_m<string iname, MVEVectorVTInfo VTI, bit subtract,
5194                            SDNode Op, Intrinsic PredInt> {
5195  def "" : MVE_VADDSUB_qr<iname, VTI.Suffix, VTI.Size, 0b0, subtract, 0b1, 0b0>;
5196  let Predicates = [HasMVEInt] in {
5197    defm : MVE_TwoOpPatternDup<VTI, Op, PredInt, (? ), !cast<Instruction>(NAME), ARMimmAllZerosV>;
5198  }
5199}
5200
5201multiclass MVE_VADD_qr_m<MVEVectorVTInfo VTI>
5202  : MVE_VADDSUB_qr_m<"vadd", VTI, 0b0, add, int_arm_mve_add_predicated>;
5203
5204multiclass MVE_VSUB_qr_m<MVEVectorVTInfo VTI>
5205  : MVE_VADDSUB_qr_m<"vsub", VTI, 0b1, sub, int_arm_mve_sub_predicated>;
5206
5207defm MVE_VADD_qr_i8  : MVE_VADD_qr_m<MVE_v16i8>;
5208defm MVE_VADD_qr_i16 : MVE_VADD_qr_m<MVE_v8i16>;
5209defm MVE_VADD_qr_i32 : MVE_VADD_qr_m<MVE_v4i32>;
5210
5211defm MVE_VSUB_qr_i8  : MVE_VSUB_qr_m<MVE_v16i8>;
5212defm MVE_VSUB_qr_i16 : MVE_VSUB_qr_m<MVE_v8i16>;
5213defm MVE_VSUB_qr_i32 : MVE_VSUB_qr_m<MVE_v4i32>;
5214
5215// Vector-scalar saturating add/sub
5216multiclass MVE_VQADDSUB_qr_m<string iname, MVEVectorVTInfo VTI, bit subtract,
5217                             SDNode Op, Intrinsic PredInt> {
5218  def "" : MVE_VADDSUB_qr<iname, VTI.Suffix, VTI.Size, 0b1, subtract,
5219                          0b0, VTI.Unsigned>;
5220
5221  let Predicates = [HasMVEInt] in {
5222    defm : MVE_TwoOpPatternDup<VTI, Op, PredInt, (? (i32 VTI.Unsigned)),
5223                               !cast<Instruction>(NAME)>;
5224  }
5225}
5226
5227multiclass MVE_VQADD_qr_m<MVEVectorVTInfo VTI, SDNode Op>
5228  : MVE_VQADDSUB_qr_m<"vqadd", VTI, 0b0, Op, int_arm_mve_qadd_predicated>;
5229
5230multiclass MVE_VQSUB_qr_m<MVEVectorVTInfo VTI, SDNode Op>
5231  : MVE_VQADDSUB_qr_m<"vqsub", VTI, 0b1, Op, int_arm_mve_qsub_predicated>;
5232
5233defm MVE_VQADD_qr_s8  : MVE_VQADD_qr_m<MVE_v16s8, saddsat>;
5234defm MVE_VQADD_qr_s16 : MVE_VQADD_qr_m<MVE_v8s16, saddsat>;
5235defm MVE_VQADD_qr_s32 : MVE_VQADD_qr_m<MVE_v4s32, saddsat>;
5236defm MVE_VQADD_qr_u8  : MVE_VQADD_qr_m<MVE_v16u8, uaddsat>;
5237defm MVE_VQADD_qr_u16 : MVE_VQADD_qr_m<MVE_v8u16, uaddsat>;
5238defm MVE_VQADD_qr_u32 : MVE_VQADD_qr_m<MVE_v4u32, uaddsat>;
5239
5240defm MVE_VQSUB_qr_s8  : MVE_VQSUB_qr_m<MVE_v16s8, ssubsat>;
5241defm MVE_VQSUB_qr_s16 : MVE_VQSUB_qr_m<MVE_v8s16, ssubsat>;
5242defm MVE_VQSUB_qr_s32 : MVE_VQSUB_qr_m<MVE_v4s32, ssubsat>;
5243defm MVE_VQSUB_qr_u8  : MVE_VQSUB_qr_m<MVE_v16u8, usubsat>;
5244defm MVE_VQSUB_qr_u16 : MVE_VQSUB_qr_m<MVE_v8u16, usubsat>;
5245defm MVE_VQSUB_qr_u32 : MVE_VQSUB_qr_m<MVE_v4u32, usubsat>;
5246
5247class MVE_VQDMULL_qr<string iname, string suffix, bit size,
5248                     bit T, string cstr="", list<dag> pattern=[]>
5249  : MVE_qDest_rSrc<iname, suffix, cstr, !if(size, 0b10, 0b01), pattern> {
5250
5251  let Inst{28} = size;
5252  let Inst{21-20} = 0b11;
5253  let Inst{16} = 0b0;
5254  let Inst{12} = T;
5255  let Inst{8} = 0b1;
5256  let Inst{5} = 0b1;
5257  let validForTailPredication = 1;
5258  let doubleWidthResult = 1;
5259}
5260
5261multiclass MVE_VQDMULL_qr_m<string iname, MVEVectorVTInfo VTI, bit size,
5262                            bit T, string cstr> {
5263  def "" : MVE_VQDMULL_qr<iname, VTI.Suffix, size, T, cstr>;
5264  defvar Inst = !cast<Instruction>(NAME);
5265
5266  let Predicates = [HasMVEInt] in {
5267    // Unpredicated saturating multiply
5268    def : Pat<(VTI.DblVec (int_arm_mve_vqdmull (VTI.Vec MQPR:$Qm),
5269                                               (VTI.Vec (ARMvdup rGPR:$val)),
5270                                               (i32 T))),
5271              (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (i32 rGPR:$val)))>;
5272    // Predicated saturating multiply
5273    def : Pat<(VTI.DblVec (int_arm_mve_vqdmull_predicated
5274                                    (VTI.Vec MQPR:$Qm),
5275                                    (VTI.Vec (ARMvdup rGPR:$val)),
5276                                    (i32 T),
5277                                    (VTI.DblPred VCCR:$mask),
5278                                    (VTI.DblVec MQPR:$inactive))),
5279              (VTI.DblVec (Inst (VTI.Vec MQPR:$Qm), (i32 rGPR:$val),
5280                             ARMVCCThen, (VTI.DblPred VCCR:$mask), zero_reg,
5281                             (VTI.DblVec MQPR:$inactive)))>;
5282  }
5283}
5284
5285multiclass MVE_VQDMULL_qr_halves<MVEVectorVTInfo VTI, bit size, string cstr=""> {
5286  defm bh : MVE_VQDMULL_qr_m<"vqdmullb", VTI, size, 0b0, cstr>;
5287  defm th : MVE_VQDMULL_qr_m<"vqdmullt", VTI, size, 0b1, cstr>;
5288}
5289
5290defm MVE_VQDMULL_qr_s16 : MVE_VQDMULL_qr_halves<MVE_v8s16, 0b0>;
5291defm MVE_VQDMULL_qr_s32 : MVE_VQDMULL_qr_halves<MVE_v4s32, 0b1, "@earlyclobber $Qd">;
5292
5293class MVE_VxADDSUB_qr<string iname, string suffix,
5294                      bit bit_28, bits<2> size, bit subtract,
5295                      bits<2> vecsize, list<dag> pattern=[]>
5296  : MVE_qDest_rSrc<iname, suffix, "", vecsize, pattern> {
5297
5298  let Inst{28} = bit_28;
5299  let Inst{21-20} = size;
5300  let Inst{16} = 0b0;
5301  let Inst{12} = subtract;
5302  let Inst{8} = 0b1;
5303  let Inst{5} = 0b0;
5304  let validForTailPredication = 1;
5305}
5306
5307multiclass MVE_VHADDSUB_qr_m<string iname, MVEVectorVTInfo VTI, bit subtract, SDPatternOperator Op,
5308                             Intrinsic unpred_int, Intrinsic pred_int, PatFrag add_op, SDNode shift_op> {
5309  def "" : MVE_VxADDSUB_qr<iname, VTI.Suffix, VTI.Unsigned, VTI.Size, subtract, VTI.Size>;
5310  defm : MVE_TwoOpPatternDup<VTI, Op, pred_int, (? (i32 VTI.Unsigned)), !cast<Instruction>(NAME)>;
5311  defm : MVE_vec_scalar_int_pat_m<!cast<Instruction>(NAME),
5312                                  VTI, unpred_int, pred_int, 1, 1>;
5313  defvar Inst = !cast<Instruction>(NAME);
5314
5315  let Predicates = [HasMVEInt] in {
5316    def : Pat<(VTI.Vec (shift_op (add_op (VTI.Vec MQPR:$Qm), (VTI.Vec (ARMvdup rGPR:$Rn))), (i32 1))),
5317              (Inst MQPR:$Qm, rGPR:$Rn)>;
5318  }
5319}
5320
5321multiclass MVE_VHADD_qr_m<MVEVectorVTInfo VTI, PatFrag add_op, SDNode shift_op, SDNode Op> :
5322  MVE_VHADDSUB_qr_m<"vhadd", VTI, 0b0, Op, int_arm_mve_vhadd,
5323                    int_arm_mve_hadd_predicated, add_op, shift_op>;
5324
5325multiclass MVE_VHSUB_qr_m<MVEVectorVTInfo VTI, PatFrag add_op, SDNode shift_op> :
5326  MVE_VHADDSUB_qr_m<"vhsub", VTI, 0b1, null_frag, int_arm_mve_vhsub,
5327                    int_arm_mve_hsub_predicated, add_op, shift_op>;
5328
5329defm MVE_VHADD_qr_s8  : MVE_VHADD_qr_m<MVE_v16s8, addnsw, ARMvshrsImm, avgfloors>;
5330defm MVE_VHADD_qr_s16 : MVE_VHADD_qr_m<MVE_v8s16, addnsw, ARMvshrsImm, avgfloors>;
5331defm MVE_VHADD_qr_s32 : MVE_VHADD_qr_m<MVE_v4s32, addnsw, ARMvshrsImm, avgfloors>;
5332defm MVE_VHADD_qr_u8  : MVE_VHADD_qr_m<MVE_v16u8, addnuw, ARMvshruImm, avgflooru>;
5333defm MVE_VHADD_qr_u16 : MVE_VHADD_qr_m<MVE_v8u16, addnuw, ARMvshruImm, avgflooru>;
5334defm MVE_VHADD_qr_u32 : MVE_VHADD_qr_m<MVE_v4u32, addnuw, ARMvshruImm, avgflooru>;
5335
5336defm MVE_VHSUB_qr_s8  : MVE_VHSUB_qr_m<MVE_v16s8, subnsw, ARMvshrsImm>;
5337defm MVE_VHSUB_qr_s16 : MVE_VHSUB_qr_m<MVE_v8s16, subnsw, ARMvshrsImm>;
5338defm MVE_VHSUB_qr_s32 : MVE_VHSUB_qr_m<MVE_v4s32, subnsw, ARMvshrsImm>;
5339defm MVE_VHSUB_qr_u8  : MVE_VHSUB_qr_m<MVE_v16u8, subnuw, ARMvshruImm>;
5340defm MVE_VHSUB_qr_u16 : MVE_VHSUB_qr_m<MVE_v8u16, subnuw, ARMvshruImm>;
5341defm MVE_VHSUB_qr_u32 : MVE_VHSUB_qr_m<MVE_v4u32, subnuw, ARMvshruImm>;
5342
5343multiclass MVE_VADDSUB_qr_f<string iname, MVEVectorVTInfo VTI, bit subtract,
5344                            SDNode Op, Intrinsic PredInt, SDPatternOperator IdentityVec> {
5345  def "" : MVE_VxADDSUB_qr<iname, VTI.Suffix, VTI.Size{0}, 0b11, subtract, VTI.Size>;
5346  defm : MVE_TwoOpPatternDup<VTI, Op, PredInt, (? ),
5347                              !cast<Instruction>(NAME), IdentityVec>;
5348}
5349
5350let Predicates = [HasMVEFloat] in {
5351  defm MVE_VADD_qr_f32 : MVE_VADDSUB_qr_f<"vadd", MVE_v4f32, 0b0, fadd,
5352                                          int_arm_mve_add_predicated, ARMimmMinusZeroF>;
5353  defm MVE_VADD_qr_f16 : MVE_VADDSUB_qr_f<"vadd", MVE_v8f16, 0b0, fadd,
5354                                          int_arm_mve_add_predicated, ARMimmMinusZeroH>;
5355
5356  defm MVE_VSUB_qr_f32 : MVE_VADDSUB_qr_f<"vsub", MVE_v4f32, 0b1, fsub,
5357                                          int_arm_mve_sub_predicated, ARMimmAllZerosV>;
5358  defm MVE_VSUB_qr_f16 : MVE_VADDSUB_qr_f<"vsub", MVE_v8f16, 0b1, fsub,
5359                                          int_arm_mve_sub_predicated, ARMimmAllZerosV>;
5360}
5361
5362class MVE_VxSHL_qr<string iname, string suffix, bit U, bits<2> size,
5363                   bit bit_7, bit bit_17, list<dag> pattern=[]>
5364  : MVE_qDest_single_rSrc<iname, suffix, size, pattern> {
5365
5366  let Inst{28} = U;
5367  let Inst{25-23} = 0b100;
5368  let Inst{21-20} = 0b11;
5369  let Inst{19-18} = size;
5370  let Inst{17} = bit_17;
5371  let Inst{16} = 0b1;
5372  let Inst{12-8} = 0b11110;
5373  let Inst{7} = bit_7;
5374  let Inst{6-4} = 0b110;
5375  let validForTailPredication = 1;
5376}
5377
5378multiclass MVE_VxSHL_qr_p<string iname, MVEVectorVTInfo VTI, bit q, bit r> {
5379  def "" : MVE_VxSHL_qr<iname, VTI.Suffix, VTI.Unsigned, VTI.Size, q, r>;
5380  defvar Inst = !cast<Instruction>(NAME);
5381
5382  def : Pat<(VTI.Vec (int_arm_mve_vshl_scalar
5383                         (VTI.Vec MQPR:$in), (i32 rGPR:$sh),
5384                         (i32 q), (i32 r), (i32 VTI.Unsigned))),
5385            (VTI.Vec (Inst (VTI.Vec MQPR:$in), (i32 rGPR:$sh)))>;
5386
5387  def : Pat<(VTI.Vec (int_arm_mve_vshl_scalar_predicated
5388                         (VTI.Vec MQPR:$in), (i32 rGPR:$sh),
5389                         (i32 q), (i32 r), (i32 VTI.Unsigned),
5390                         (VTI.Pred VCCR:$mask))),
5391            (VTI.Vec (Inst (VTI.Vec MQPR:$in), (i32 rGPR:$sh),
5392                           ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg))>;
5393}
5394
5395multiclass MVE_VxSHL_qr_types<string iname, bit bit_7, bit bit_17> {
5396  defm s8  : MVE_VxSHL_qr_p<iname, MVE_v16s8, bit_7, bit_17>;
5397  defm s16 : MVE_VxSHL_qr_p<iname, MVE_v8s16, bit_7, bit_17>;
5398  defm s32 : MVE_VxSHL_qr_p<iname, MVE_v4s32, bit_7, bit_17>;
5399  defm u8  : MVE_VxSHL_qr_p<iname, MVE_v16u8, bit_7, bit_17>;
5400  defm u16 : MVE_VxSHL_qr_p<iname, MVE_v8u16, bit_7, bit_17>;
5401  defm u32 : MVE_VxSHL_qr_p<iname, MVE_v4u32, bit_7, bit_17>;
5402}
5403
5404defm MVE_VSHL_qr   : MVE_VxSHL_qr_types<"vshl",   0b0, 0b0>;
5405defm MVE_VRSHL_qr  : MVE_VxSHL_qr_types<"vrshl",  0b0, 0b1>;
5406defm MVE_VQSHL_qr  : MVE_VxSHL_qr_types<"vqshl",  0b1, 0b0>;
5407defm MVE_VQRSHL_qr : MVE_VxSHL_qr_types<"vqrshl", 0b1, 0b1>;
5408
5409let Predicates = [HasMVEInt] in {
5410  def : Pat<(v4i32 (ARMvshlu (v4i32 MQPR:$Qm), (v4i32 (ARMvdup rGPR:$Rm)))),
5411            (v4i32 (MVE_VSHL_qru32 (v4i32 MQPR:$Qm), rGPR:$Rm))>;
5412  def : Pat<(v8i16 (ARMvshlu (v8i16 MQPR:$Qm), (v8i16 (ARMvdup rGPR:$Rm)))),
5413            (v8i16 (MVE_VSHL_qru16 (v8i16 MQPR:$Qm), rGPR:$Rm))>;
5414  def : Pat<(v16i8 (ARMvshlu (v16i8 MQPR:$Qm), (v16i8 (ARMvdup rGPR:$Rm)))),
5415            (v16i8 (MVE_VSHL_qru8 (v16i8 MQPR:$Qm), rGPR:$Rm))>;
5416
5417  def : Pat<(v4i32 (ARMvshls (v4i32 MQPR:$Qm), (v4i32 (ARMvdup rGPR:$Rm)))),
5418            (v4i32 (MVE_VSHL_qrs32 (v4i32 MQPR:$Qm), rGPR:$Rm))>;
5419  def : Pat<(v8i16 (ARMvshls (v8i16 MQPR:$Qm), (v8i16 (ARMvdup rGPR:$Rm)))),
5420            (v8i16 (MVE_VSHL_qrs16 (v8i16 MQPR:$Qm), rGPR:$Rm))>;
5421  def : Pat<(v16i8 (ARMvshls (v16i8 MQPR:$Qm), (v16i8 (ARMvdup rGPR:$Rm)))),
5422            (v16i8 (MVE_VSHL_qrs8 (v16i8 MQPR:$Qm), rGPR:$Rm))>;
5423}
5424
5425class MVE_VBRSR<string iname, string suffix, bits<2> size, list<dag> pattern=[]>
5426  : MVE_qDest_rSrc<iname, suffix, "", size, pattern> {
5427
5428  let Inst{28} = 0b1;
5429  let Inst{21-20} = size;
5430  let Inst{16} = 0b1;
5431  let Inst{12} = 0b1;
5432  let Inst{8} = 0b0;
5433  let Inst{5} = 0b1;
5434  let validForTailPredication = 1;
5435}
5436
5437def MVE_VBRSR8  : MVE_VBRSR<"vbrsr", "8", 0b00>;
5438def MVE_VBRSR16 : MVE_VBRSR<"vbrsr", "16", 0b01>;
5439def MVE_VBRSR32 : MVE_VBRSR<"vbrsr", "32", 0b10>;
5440
5441multiclass MVE_VBRSR_pat_m<MVEVectorVTInfo VTI, Instruction Inst> {
5442  // Unpredicated
5443  def : Pat<(VTI.Vec (int_arm_mve_vbrsr (VTI.Vec MQPR:$Qn), (i32 rGPR:$Rm))),
5444            (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (i32 rGPR:$Rm)))>;
5445  // Predicated
5446  def : Pat<(VTI.Vec (int_arm_mve_vbrsr_predicated
5447                          (VTI.Vec MQPR:$inactive),
5448                          (VTI.Vec MQPR:$Qn), (i32 rGPR:$Rm),
5449                          (VTI.Pred VCCR:$mask))),
5450            (VTI.Vec (Inst (VTI.Vec MQPR:$Qn), (i32 rGPR:$Rm),
5451                          ARMVCCThen, (VTI.Pred VCCR:$mask), zero_reg,
5452                          (VTI.Vec MQPR:$inactive)))>;
5453}
5454
5455let Predicates = [HasMVEInt] in {
5456  def : Pat<(v16i8 ( bitreverse (v16i8 MQPR:$val1))),
5457            (v16i8 ( MVE_VBRSR8 (v16i8 MQPR:$val1), (t2MOVi (i32 8)) ))>;
5458
5459  def : Pat<(v4i32 ( bitreverse (v4i32 MQPR:$val1))),
5460            (v4i32 ( MVE_VBRSR32 (v4i32 MQPR:$val1), (t2MOVi (i32 32)) ))>;
5461
5462  def : Pat<(v8i16 ( bitreverse (v8i16 MQPR:$val1))),
5463            (v8i16 ( MVE_VBRSR16 (v8i16 MQPR:$val1), (t2MOVi (i32 16)) ))>;
5464
5465  defm : MVE_VBRSR_pat_m<MVE_v16i8, MVE_VBRSR8>;
5466  defm : MVE_VBRSR_pat_m<MVE_v8i16, MVE_VBRSR16>;
5467  defm : MVE_VBRSR_pat_m<MVE_v4i32, MVE_VBRSR32>;
5468}
5469
5470let Predicates = [HasMVEFloat] in {
5471  defm : MVE_VBRSR_pat_m<MVE_v8f16, MVE_VBRSR16>;
5472  defm : MVE_VBRSR_pat_m<MVE_v4f32, MVE_VBRSR32>;
5473}
5474
5475class MVE_VMUL_qr_int<string iname, string suffix, bits<2> size>
5476  : MVE_qDest_rSrc<iname, suffix, "", size> {
5477
5478  let Inst{28} = 0b0;
5479  let Inst{21-20} = size;
5480  let Inst{16} = 0b1;
5481  let Inst{12} = 0b1;
5482  let Inst{8} = 0b0;
5483  let Inst{5} = 0b1;
5484  let validForTailPredication = 1;
5485}
5486
5487multiclass MVE_VMUL_qr_int_m<MVEVectorVTInfo VTI> {
5488  def "" : MVE_VMUL_qr_int<"vmul", VTI.Suffix, VTI.Size>;
5489  let Predicates = [HasMVEInt] in {
5490    defm : MVE_TwoOpPatternDup<VTI, mul, int_arm_mve_mul_predicated, (? ),
5491                               !cast<Instruction>(NAME), ARMimmOneV>;
5492  }
5493}
5494
5495defm MVE_VMUL_qr_i8  : MVE_VMUL_qr_int_m<MVE_v16i8>;
5496defm MVE_VMUL_qr_i16 : MVE_VMUL_qr_int_m<MVE_v8i16>;
5497defm MVE_VMUL_qr_i32 : MVE_VMUL_qr_int_m<MVE_v4i32>;
5498
5499class MVE_VxxMUL_qr<string iname, string suffix,
5500                    bit bit_28, bits<2> size, bits<2> vecsize, list<dag> pattern=[]>
5501  : MVE_qDest_rSrc<iname, suffix, "", vecsize, pattern> {
5502
5503  let Inst{28} = bit_28;
5504  let Inst{21-20} = size;
5505  let Inst{16} = 0b1;
5506  let Inst{12} = 0b0;
5507  let Inst{8} = 0b0;
5508  let Inst{5} = 0b1;
5509  let validForTailPredication = 1;
5510}
5511
5512multiclass MVE_VxxMUL_qr_m<string iname, MVEVectorVTInfo VTI, bit bit_28,
5513                           SDPatternOperator Op, Intrinsic int_unpred, Intrinsic int_pred> {
5514  def "" : MVE_VxxMUL_qr<iname, VTI.Suffix, bit_28, VTI.Size, VTI.Size>;
5515
5516  let Predicates = [HasMVEInt] in {
5517    defm : MVE_TwoOpPatternDup<VTI, Op, int_pred, (? ), !cast<Instruction>(NAME)>;
5518  }
5519  defm : MVE_vec_scalar_int_pat_m<!cast<Instruction>(NAME), VTI, int_unpred, int_pred>;
5520}
5521
5522multiclass MVE_VQDMULH_qr_m<MVEVectorVTInfo VTI> :
5523  MVE_VxxMUL_qr_m<"vqdmulh", VTI, 0b0, MVEvqdmulh,
5524                  int_arm_mve_vqdmulh, int_arm_mve_qdmulh_predicated>;
5525
5526multiclass MVE_VQRDMULH_qr_m<MVEVectorVTInfo VTI> :
5527  MVE_VxxMUL_qr_m<"vqrdmulh", VTI, 0b1, null_frag,
5528                  int_arm_mve_vqrdmulh, int_arm_mve_qrdmulh_predicated>;
5529
5530defm MVE_VQDMULH_qr_s8    : MVE_VQDMULH_qr_m<MVE_v16s8>;
5531defm MVE_VQDMULH_qr_s16   : MVE_VQDMULH_qr_m<MVE_v8s16>;
5532defm MVE_VQDMULH_qr_s32   : MVE_VQDMULH_qr_m<MVE_v4s32>;
5533
5534defm MVE_VQRDMULH_qr_s8   : MVE_VQRDMULH_qr_m<MVE_v16s8>;
5535defm MVE_VQRDMULH_qr_s16  : MVE_VQRDMULH_qr_m<MVE_v8s16>;
5536defm MVE_VQRDMULH_qr_s32  : MVE_VQRDMULH_qr_m<MVE_v4s32>;
5537
5538multiclass MVE_VxxMUL_qr_f_m<MVEVectorVTInfo VTI, SDPatternOperator IdentityVec> {
5539  let validForTailPredication = 1 in
5540  def "" : MVE_VxxMUL_qr<"vmul", VTI.Suffix, VTI.Size{0}, 0b11, VTI.Size>;
5541  defm : MVE_TwoOpPatternDup<VTI, fmul, int_arm_mve_mul_predicated, (? ),
5542                             !cast<Instruction>(NAME), IdentityVec>;
5543}
5544
5545let Predicates = [HasMVEFloat] in {
5546  defm MVE_VMUL_qr_f16   : MVE_VxxMUL_qr_f_m<MVE_v8f16, ARMimmOneH>;
5547  defm MVE_VMUL_qr_f32   : MVE_VxxMUL_qr_f_m<MVE_v4f32, ARMimmOneF>;
5548}
5549
5550class MVE_VFMAMLA_qr<string iname, string suffix,
5551                     bit bit_28, bits<2> bits_21_20, bit S,
5552                     bits<2> vecsize, list<dag> pattern=[]>
5553  : MVE_qDestSrc_rSrc<iname, suffix, vecsize, pattern> {
5554
5555  let Inst{28} = bit_28;
5556  let Inst{21-20} = bits_21_20;
5557  let Inst{16} = 0b1;
5558  let Inst{12} = S;
5559  let Inst{8} = 0b0;
5560  let Inst{5} = 0b0;
5561  let validForTailPredication = 1;
5562  let hasSideEffects = 0;
5563}
5564
5565multiclass MVE_VMLA_qr_multi<string iname, MVEVectorVTInfo VTI,
5566                             bit scalar_addend> {
5567  def "": MVE_VFMAMLA_qr<iname, VTI.Suffix, 0b0, VTI.Size,
5568                         scalar_addend, VTI.Size>;
5569  defvar Inst = !cast<Instruction>(NAME);
5570  defvar pred_int = !cast<Intrinsic>("int_arm_mve_" # iname # "_n_predicated");
5571  defvar v1   = (VTI.Vec MQPR:$v1);
5572  defvar v2   = (VTI.Vec MQPR:$v2);
5573  defvar vs   = (VTI.Vec (ARMvdup rGPR:$s));
5574  defvar s    = (i32 rGPR:$s);
5575  defvar pred = (VTI.Pred VCCR:$pred);
5576
5577  let Predicates = [HasMVEInt] in {
5578    if scalar_addend then {
5579      def : Pat<(VTI.Vec (add (mul v1, v2), vs)),
5580                (VTI.Vec (Inst v1, v2, s))>;
5581    } else {
5582      def : Pat<(VTI.Vec (add (mul v2, vs), v1)),
5583                (VTI.Vec (Inst v1, v2, s))>;
5584    }
5585
5586    def : Pat<(VTI.Vec (pred_int v1, v2, s, pred)),
5587              (VTI.Vec (Inst v1, v2, s, ARMVCCThen, pred, zero_reg))>;
5588  }
5589}
5590
5591defm MVE_VMLA_qr_i8   : MVE_VMLA_qr_multi<"vmla", MVE_v16i8, 0b0>;
5592defm MVE_VMLA_qr_i16  : MVE_VMLA_qr_multi<"vmla", MVE_v8i16, 0b0>;
5593defm MVE_VMLA_qr_i32  : MVE_VMLA_qr_multi<"vmla", MVE_v4i32, 0b0>;
5594
5595defm MVE_VMLAS_qr_i8  : MVE_VMLA_qr_multi<"vmlas", MVE_v16i8, 0b1>;
5596defm MVE_VMLAS_qr_i16 : MVE_VMLA_qr_multi<"vmlas", MVE_v8i16, 0b1>;
5597defm MVE_VMLAS_qr_i32 : MVE_VMLA_qr_multi<"vmlas", MVE_v4i32, 0b1>;
5598
5599multiclass MVE_VFMA_qr_multi<string iname, MVEVectorVTInfo VTI,
5600                             bit scalar_addend> {
5601  def "": MVE_VFMAMLA_qr<iname, VTI.Suffix, VTI.Size{0}, 0b11, scalar_addend, VTI.Size>;
5602  defvar Inst = !cast<Instruction>(NAME);
5603  defvar pred_int = int_arm_mve_fma_predicated;
5604  defvar v1   = (VTI.Vec MQPR:$v1);
5605  defvar v2   = (VTI.Vec MQPR:$v2);
5606  defvar vs   = (VTI.Vec (ARMvdup (i32 rGPR:$s)));
5607  defvar is   = (i32 rGPR:$s);
5608  defvar pred = (VTI.Pred VCCR:$pred);
5609
5610  let Predicates = [HasMVEFloat] in {
5611    if scalar_addend then {
5612      def : Pat<(VTI.Vec (fma v1, v2, vs)),
5613                (VTI.Vec (Inst v1, v2, is))>;
5614      def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
5615                                  (VTI.Vec (fma v1, v2, vs)),
5616                                  v1)),
5617                (VTI.Vec (Inst v1, v2, is, ARMVCCThen, $pred, zero_reg))>;
5618    } else {
5619      def : Pat<(VTI.Vec (fma v1, vs, v2)),
5620                (VTI.Vec (Inst v2, v1, is))>;
5621      def : Pat<(VTI.Vec (fma vs, v1, v2)),
5622                (VTI.Vec (Inst v2, v1, is))>;
5623      def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
5624                                  (VTI.Vec (fma vs, v2, v1)),
5625                                  v1)),
5626                (VTI.Vec (Inst v1, v2, is, ARMVCCThen, $pred, zero_reg))>;
5627      def : Pat<(VTI.Vec (vselect (VTI.Pred VCCR:$pred),
5628                                  (VTI.Vec (fma v2, vs, v1)),
5629                                  v1)),
5630                (VTI.Vec (Inst v1, v2, is, ARMVCCThen, $pred, zero_reg))>;
5631      def : Pat<(VTI.Vec (pred_int v1, vs, v2, pred)),
5632                (VTI.Vec (Inst v2, v1, is, ARMVCCThen, pred, zero_reg))>;
5633      def : Pat<(VTI.Vec (pred_int vs, v1, v2, pred)),
5634                (VTI.Vec (Inst v2, v1, is, ARMVCCThen, pred, zero_reg))>;
5635    }
5636  }
5637}
5638
5639let Predicates = [HasMVEFloat] in {
5640  defm MVE_VFMA_qr_f16  : MVE_VFMA_qr_multi<"vfma",  MVE_v8f16, 0>;
5641  defm MVE_VFMA_qr_f32  : MVE_VFMA_qr_multi<"vfma",  MVE_v4f32, 0>;
5642  defm MVE_VFMA_qr_Sf16 : MVE_VFMA_qr_multi<"vfmas", MVE_v8f16, 1>;
5643  defm MVE_VFMA_qr_Sf32 : MVE_VFMA_qr_multi<"vfmas", MVE_v4f32, 1>;
5644}
5645
5646class MVE_VQDMLAH_qr<string iname, string suffix, bit U, bits<2> size,
5647                     bit bit_5, bit bit_12, list<dag> pattern=[]>
5648  : MVE_qDestSrc_rSrc<iname, suffix, size, pattern> {
5649
5650  let Inst{28} = U;
5651  let Inst{21-20} = size;
5652  let Inst{16} = 0b0;
5653  let Inst{12} = bit_12;
5654  let Inst{8} = 0b0;
5655  let Inst{5} = bit_5;
5656}
5657
5658multiclass MVE_VQDMLAH_qr_multi<string iname, MVEVectorVTInfo VTI,
5659                                bit bit_5, bit bit_12> {
5660  def "": MVE_VQDMLAH_qr<iname, VTI.Suffix, 0b0, VTI.Size, bit_5, bit_12>;
5661  defvar Inst = !cast<Instruction>(NAME);
5662  defvar unpred_int = !cast<Intrinsic>("int_arm_mve_" # iname);
5663  defvar pred_int = !cast<Intrinsic>("int_arm_mve_" # iname # "_predicated");
5664
5665  let Predicates = [HasMVEInt] in {
5666    def : Pat<(VTI.Vec (unpred_int (VTI.Vec MQPR:$v1), (VTI.Vec MQPR:$v2),
5667                                   (i32 rGPR:$s))),
5668              (VTI.Vec (Inst       (VTI.Vec MQPR:$v1), (VTI.Vec MQPR:$v2),
5669                                   (i32 rGPR:$s)))>;
5670    def : Pat<(VTI.Vec (pred_int   (VTI.Vec MQPR:$v1), (VTI.Vec MQPR:$v2),
5671                                   (i32 rGPR:$s), (VTI.Pred VCCR:$pred))),
5672              (VTI.Vec (Inst       (VTI.Vec MQPR:$v1), (VTI.Vec MQPR:$v2),
5673                                   (i32 rGPR:$s), ARMVCCThen,
5674                                   (VTI.Pred VCCR:$pred), zero_reg))>;
5675  }
5676}
5677
5678multiclass MVE_VQDMLAH_qr_types<string iname, bit bit_5, bit bit_12> {
5679  defm s8  : MVE_VQDMLAH_qr_multi<iname, MVE_v16s8, bit_5, bit_12>;
5680  defm s16 : MVE_VQDMLAH_qr_multi<iname, MVE_v8s16, bit_5, bit_12>;
5681  defm s32 : MVE_VQDMLAH_qr_multi<iname, MVE_v4s32, bit_5, bit_12>;
5682}
5683
5684defm MVE_VQDMLAH_qr   : MVE_VQDMLAH_qr_types<"vqdmlah",   0b1, 0b0>;
5685defm MVE_VQRDMLAH_qr  : MVE_VQDMLAH_qr_types<"vqrdmlah",  0b0, 0b0>;
5686defm MVE_VQDMLASH_qr  : MVE_VQDMLAH_qr_types<"vqdmlash",  0b1, 0b1>;
5687defm MVE_VQRDMLASH_qr : MVE_VQDMLAH_qr_types<"vqrdmlash", 0b0, 0b1>;
5688
5689class MVE_VxDUP<string iname, string suffix, bits<2> size, bit bit_12,
5690                ValueType VT, SDPatternOperator vxdup>
5691  : MVE_p<(outs MQPR:$Qd, tGPREven:$Rn),
5692          (ins tGPREven:$Rn_src, MVE_VIDUP_imm:$imm), NoItinerary,
5693          iname, suffix, "$Qd, $Rn, $imm", vpred_r, "$Rn = $Rn_src", size,
5694          [(set (VT MQPR:$Qd), (i32 tGPREven:$Rn),
5695              (vxdup (i32 tGPREven:$Rn_src), (i32 imm:$imm)))]> {
5696  bits<4> Qd;
5697  bits<4> Rn;
5698  bits<2> imm;
5699
5700  let Inst{28} = 0b0;
5701  let Inst{25-23} = 0b100;
5702  let Inst{22} = Qd{3};
5703  let Inst{21-20} = size;
5704  let Inst{19-17} = Rn{3-1};
5705  let Inst{16} = 0b1;
5706  let Inst{15-13} = Qd{2-0};
5707  let Inst{12} = bit_12;
5708  let Inst{11-8} = 0b1111;
5709  let Inst{7} = imm{1};
5710  let Inst{6-1} = 0b110111;
5711  let Inst{0} = imm{0};
5712  let validForTailPredication = 1;
5713  let hasSideEffects = 0;
5714}
5715
5716def MVE_VIDUPu8  : MVE_VxDUP<"vidup", "u8",  0b00, 0b0, v16i8, ARMvidup>;
5717def MVE_VIDUPu16 : MVE_VxDUP<"vidup", "u16", 0b01, 0b0, v8i16, ARMvidup>;
5718def MVE_VIDUPu32 : MVE_VxDUP<"vidup", "u32", 0b10, 0b0, v4i32, ARMvidup>;
5719
5720def MVE_VDDUPu8  : MVE_VxDUP<"vddup", "u8",  0b00, 0b1, v16i8, null_frag>;
5721def MVE_VDDUPu16 : MVE_VxDUP<"vddup", "u16", 0b01, 0b1, v8i16, null_frag>;
5722def MVE_VDDUPu32 : MVE_VxDUP<"vddup", "u32", 0b10, 0b1, v4i32, null_frag>;
5723
5724class MVE_VxWDUP<string iname, string suffix, bits<2> size, bit bit_12,
5725                 list<dag> pattern=[]>
5726  : MVE_p<(outs MQPR:$Qd, tGPREven:$Rn),
5727          (ins tGPREven:$Rn_src, tGPROdd:$Rm, MVE_VIDUP_imm:$imm), NoItinerary,
5728          iname, suffix, "$Qd, $Rn, $Rm, $imm", vpred_r, "$Rn = $Rn_src", size,
5729          pattern> {
5730  bits<4> Qd;
5731  bits<4> Rm;
5732  bits<4> Rn;
5733  bits<2> imm;
5734
5735  let Inst{28} = 0b0;
5736  let Inst{25-23} = 0b100;
5737  let Inst{22} = Qd{3};
5738  let Inst{21-20} = size;
5739  let Inst{19-17} = Rn{3-1};
5740  let Inst{16} = 0b1;
5741  let Inst{15-13} = Qd{2-0};
5742  let Inst{12} = bit_12;
5743  let Inst{11-8} = 0b1111;
5744  let Inst{7} = imm{1};
5745  let Inst{6-4} = 0b110;
5746  let Inst{3-1} = Rm{3-1};
5747  let Inst{0} = imm{0};
5748  let validForTailPredication = 1;
5749  let hasSideEffects = 0;
5750}
5751
5752def MVE_VIWDUPu8  : MVE_VxWDUP<"viwdup", "u8",  0b00, 0b0>;
5753def MVE_VIWDUPu16 : MVE_VxWDUP<"viwdup", "u16", 0b01, 0b0>;
5754def MVE_VIWDUPu32 : MVE_VxWDUP<"viwdup", "u32", 0b10, 0b0>;
5755
5756def MVE_VDWDUPu8  : MVE_VxWDUP<"vdwdup", "u8",  0b00, 0b1>;
5757def MVE_VDWDUPu16 : MVE_VxWDUP<"vdwdup", "u16", 0b01, 0b1>;
5758def MVE_VDWDUPu32 : MVE_VxWDUP<"vdwdup", "u32", 0b10, 0b1>;
5759
5760let isReMaterializable = 1 in
5761class MVE_VCTPInst<string suffix, bits<2> size, list<dag> pattern=[]>
5762  : MVE_p<(outs VCCR:$P0), (ins rGPR:$Rn), NoItinerary, "vctp", suffix,
5763          "$Rn", vpred_n, "", size, pattern> {
5764  bits<4> Rn;
5765
5766  let Inst{28-27} = 0b10;
5767  let Inst{26-22} = 0b00000;
5768  let Inst{21-20} = size;
5769  let Inst{19-16} = Rn{3-0};
5770  let Inst{15-11} = 0b11101;
5771  let Inst{10-0}  = 0b00000000001;
5772  let Unpredictable{10-0} = 0b11111111111;
5773
5774  let Constraints = "";
5775  let DecoderMethod = "DecodeMveVCTP";
5776  let validForTailPredication = 1;
5777}
5778
5779multiclass MVE_VCTP<MVEVectorVTInfo VTI, Intrinsic intr> {
5780  def "": MVE_VCTPInst<VTI.BitsSuffix, VTI.Size>;
5781  defvar Inst = !cast<Instruction>(NAME);
5782
5783  let Predicates = [HasMVEInt] in {
5784    def : Pat<(intr rGPR:$Rn),
5785              (VTI.Pred (Inst rGPR:$Rn))>;
5786    def : Pat<(and (intr rGPR:$Rn), (VTI.Pred VCCR:$mask)),
5787              (VTI.Pred (Inst rGPR:$Rn, ARMVCCThen, VCCR:$mask, zero_reg))>;
5788  }
5789}
5790
5791defm MVE_VCTP8  : MVE_VCTP<MVE_v16i8, int_arm_mve_vctp8>;
5792defm MVE_VCTP16 : MVE_VCTP<MVE_v8i16, int_arm_mve_vctp16>;
5793defm MVE_VCTP32 : MVE_VCTP<MVE_v4i32, int_arm_mve_vctp32>;
5794defm MVE_VCTP64 : MVE_VCTP<MVE_v2i64, int_arm_mve_vctp64>;
5795
5796// end of mve_qDest_rSrc
5797
5798// start of coproc mov
5799
5800class MVE_VMOV_64bit<dag oops, dag iops, bit to_qreg, string ops, string cstr>
5801  : MVE_VMOV_lane_base<oops, !con(iops, (ins MVEPairVectorIndex2:$idx,
5802                                             MVEPairVectorIndex0:$idx2)),
5803                       NoItinerary, "vmov", "", ops, cstr, []> {
5804  bits<5> Rt;
5805  bits<5> Rt2;
5806  bits<4> Qd;
5807  bit idx;
5808  bit idx2;
5809
5810  let Inst{31-23} = 0b111011000;
5811  let Inst{22} = Qd{3};
5812  let Inst{21} = 0b0;
5813  let Inst{20} = to_qreg;
5814  let Inst{19-16} = Rt2{3-0};
5815  let Inst{15-13} = Qd{2-0};
5816  let Inst{12-5} = 0b01111000;
5817  let Inst{4} = idx2;
5818  let Inst{3-0} = Rt{3-0};
5819
5820  let VecSize = 0b10;
5821  let hasSideEffects = 0;
5822}
5823
5824// The assembly syntax for these instructions mentions the vector
5825// register name twice, e.g.
5826//
5827//    vmov q2[2], q2[0], r0, r1
5828//    vmov r0, r1, q2[2], q2[0]
5829//
5830// which needs a bit of juggling with MC operand handling.
5831//
5832// For the move _into_ a vector register, the MC operand list also has
5833// to mention the register name twice: once as the output, and once as
5834// an extra input to represent where the unchanged half of the output
5835// register comes from (when this instruction is used in code
5836// generation). So we arrange that the first mention of the vector reg
5837// in the instruction is considered by the AsmMatcher to be the output
5838// ($Qd), and the second one is the input ($QdSrc). Binding them
5839// together with the existing 'tie' constraint is enough to enforce at
5840// register allocation time that they have to be the same register.
5841//
5842// For the move _from_ a vector register, there's no way to get round
5843// the fact that both instances of that register name have to be
5844// inputs. They have to be the same register again, but this time, we
5845// can't use a tie constraint, because that has to be between an
5846// output and an input operand. So this time, we have to arrange that
5847// the q-reg appears just once in the MC operand list, in spite of
5848// being mentioned twice in the asm syntax - which needs a custom
5849// AsmMatchConverter.
5850
5851def MVE_VMOV_q_rr : MVE_VMOV_64bit<(outs MQPR:$Qd),
5852                                   (ins MQPR:$QdSrc, rGPR:$Rt, rGPR:$Rt2),
5853                                   0b1, "$Qd$idx, $QdSrc$idx2, $Rt, $Rt2",
5854                                   "$Qd = $QdSrc"> {
5855  let DecoderMethod = "DecodeMVEVMOVDRegtoQ";
5856}
5857
5858def MVE_VMOV_rr_q : MVE_VMOV_64bit<(outs rGPR:$Rt, rGPR:$Rt2), (ins MQPR:$Qd),
5859                                   0b0, "$Rt, $Rt2, $Qd$idx, $Qd$idx2", ""> {
5860  let DecoderMethod = "DecodeMVEVMOVQtoDReg";
5861  let AsmMatchConverter = "cvtMVEVMOVQtoDReg";
5862}
5863
5864let Predicates = [HasMVEInt] in {
5865  // Double lane moves. There are a number of patterns here. We know that the
5866  // insertelt's will be in descending order by index, and need to match the 5
5867  // patterns that might contain 2-0 or 3-1 pairs. These are:
5868  // 3 2 1 0    -> vmovqrr 31; vmovqrr 20
5869  // 3 2 1      -> vmovqrr 31; vmov 2
5870  // 3 1        -> vmovqrr 31
5871  // 2 1 0      -> vmovqrr 20; vmov 1
5872  // 2 0        -> vmovqrr 20
5873  // The other potential patterns will be handled by single lane inserts.
5874  def : Pat<(insertelt (insertelt (insertelt (insertelt (v4i32 MQPR:$src1),
5875                                                        rGPR:$srcA, (i32 0)),
5876                                             rGPR:$srcB, (i32 1)),
5877                                  rGPR:$srcC, (i32 2)),
5878                       rGPR:$srcD, (i32 3)),
5879            (MVE_VMOV_q_rr (MVE_VMOV_q_rr MQPR:$src1, rGPR:$srcA, rGPR:$srcC, (i32 2), (i32 0)),
5880                           rGPR:$srcB, rGPR:$srcD, (i32 3), (i32 1))>;
5881  def : Pat<(insertelt (insertelt (insertelt (v4i32 MQPR:$src1),
5882                                             rGPR:$srcB, (i32 1)),
5883                                  rGPR:$srcC, (i32 2)),
5884                       rGPR:$srcD, (i32 3)),
5885            (MVE_VMOV_q_rr (MVE_VMOV_to_lane_32 MQPR:$src1, rGPR:$srcC, (i32 2)),
5886                           rGPR:$srcB, rGPR:$srcD, (i32 3), (i32 1))>;
5887  def : Pat<(insertelt (insertelt (v4i32 MQPR:$src1), rGPR:$srcA, (i32 1)), rGPR:$srcB, (i32 3)),
5888            (MVE_VMOV_q_rr MQPR:$src1, rGPR:$srcA, rGPR:$srcB, (i32 3), (i32 1))>;
5889  def : Pat<(insertelt (insertelt (insertelt (v4i32 MQPR:$src1),
5890                                             rGPR:$srcB, (i32 0)),
5891                                  rGPR:$srcC, (i32 1)),
5892                       rGPR:$srcD, (i32 2)),
5893            (MVE_VMOV_q_rr (MVE_VMOV_to_lane_32 MQPR:$src1, rGPR:$srcC, (i32 1)),
5894                           rGPR:$srcB, rGPR:$srcD, (i32 2), (i32 0))>;
5895  def : Pat<(insertelt (insertelt (v4i32 MQPR:$src1), rGPR:$srcA, (i32 0)), rGPR:$srcB, (i32 2)),
5896            (MVE_VMOV_q_rr MQPR:$src1, rGPR:$srcA, rGPR:$srcB, (i32 2), (i32 0))>;
5897}
5898
5899// end of coproc mov
5900
5901// start of MVE interleaving load/store
5902
5903// Base class for the family of interleaving/deinterleaving
5904// load/stores with names like VLD20.8 and VST43.32.
5905class MVE_vldst24_base<bit writeback, bit fourregs, bits<2> stage, bits<2> size,
5906                       bit load, dag Oops, dag loadIops, dag wbIops,
5907                       string iname, string ops,
5908                       string cstr, list<dag> pattern=[]>
5909  : MVE_MI<Oops, !con(loadIops, wbIops), NoItinerary, iname, ops, cstr, size, pattern> {
5910  bits<4> VQd;
5911  bits<4> Rn;
5912
5913  let Inst{31-22} = 0b1111110010;
5914  let Inst{21} = writeback;
5915  let Inst{20} = load;
5916  let Inst{19-16} = Rn;
5917  let Inst{15-13} = VQd{2-0};
5918  let Inst{12-9} = 0b1111;
5919  let Inst{8-7} = size;
5920  let Inst{6-5} = stage;
5921  let Inst{4-1} = 0b0000;
5922  let Inst{0} = fourregs;
5923
5924  let mayLoad = load;
5925  let mayStore = !eq(load,0);
5926  let hasSideEffects = 0;
5927  let validForTailPredication = load;
5928}
5929
5930// A parameter class used to encapsulate all the ways the writeback
5931// variants of VLD20 and friends differ from the non-writeback ones.
5932class MVE_vldst24_writeback<bit b, dag Oo, dag Io,
5933                            string sy="", string c="", string n=""> {
5934  bit writeback = b;
5935  dag Oops = Oo;
5936  dag Iops = Io;
5937  string syntax = sy;
5938  string cstr = c;
5939  string id_suffix = n;
5940}
5941
5942// Another parameter class that encapsulates the differences between VLD2x
5943// and VLD4x.
5944class MVE_vldst24_nvecs<int n, list<int> s, bit b, RegisterOperand vl> {
5945  int nvecs = n;
5946  list<int> stages = s;
5947  bit bit0 = b;
5948  RegisterOperand VecList = vl;
5949}
5950
5951// A third parameter class that distinguishes VLDnn.8 from .16 from .32.
5952class MVE_vldst24_lanesize<int i, bits<2> b> {
5953  int lanesize = i;
5954  bits<2> sizebits = b;
5955}
5956
5957// A base class for each direction of transfer: one for load, one for
5958// store. I can't make these a fourth independent parametric tuple
5959// class, because they have to take the nvecs tuple class as a
5960// parameter, in order to find the right VecList operand type.
5961
5962class MVE_vld24_base<MVE_vldst24_nvecs n, bits<2> pat, bits<2> size,
5963                     MVE_vldst24_writeback wb, string iname,
5964                     list<dag> pattern=[]>
5965  : MVE_vldst24_base<wb.writeback, n.bit0, pat, size, 1,
5966                     !con((outs n.VecList:$VQd), wb.Oops),
5967                     (ins n.VecList:$VQdSrc), wb.Iops,
5968                     iname, "$VQd, $Rn" # wb.syntax,
5969                     wb.cstr # ",$VQdSrc = $VQd", pattern>;
5970
5971class MVE_vst24_base<MVE_vldst24_nvecs n, bits<2> pat, bits<2> size,
5972                     MVE_vldst24_writeback wb, string iname,
5973                     list<dag> pattern=[]>
5974  : MVE_vldst24_base<wb.writeback, n.bit0, pat, size, 0,
5975                     wb.Oops, (ins n.VecList:$VQd), wb.Iops,
5976                     iname, "$VQd, $Rn" # wb.syntax,
5977                     wb.cstr, pattern>;
5978
5979// Actually define all the interleaving loads and stores, by a series
5980// of nested foreaches over number of vectors (VLD2/VLD4); stage
5981// within one of those series (VLDx0/VLDx1/VLDx2/VLDx3); size of
5982// vector lane; writeback or no writeback.
5983foreach n = [MVE_vldst24_nvecs<2, [0,1],     0, VecList2Q>,
5984             MVE_vldst24_nvecs<4, [0,1,2,3], 1, VecList4Q>] in
5985foreach stage = n.stages in
5986foreach s = [MVE_vldst24_lanesize< 8, 0b00>,
5987             MVE_vldst24_lanesize<16, 0b01>,
5988             MVE_vldst24_lanesize<32, 0b10>] in
5989foreach wb = [MVE_vldst24_writeback<
5990                1, (outs rGPR:$wb), (ins t2_nosp_addr_offset_none:$Rn),
5991                "!", "$Rn.base = $wb", "_wb">,
5992              MVE_vldst24_writeback<0, (outs), (ins t2_addr_offset_none:$Rn)>] in {
5993
5994  // For each case within all of those foreaches, define the actual
5995  // instructions. The def names are made by gluing together pieces
5996  // from all the parameter classes, and will end up being things like
5997  // MVE_VLD20_8 and MVE_VST43_16_wb.
5998
5999  def "MVE_VLD" # n.nvecs # stage # "_" # s.lanesize # wb.id_suffix
6000    : MVE_vld24_base<n, stage, s.sizebits, wb,
6001                     "vld" # n.nvecs # stage # "." # s.lanesize>;
6002
6003  def "MVE_VST" # n.nvecs # stage # "_" # s.lanesize # wb.id_suffix
6004    : MVE_vst24_base<n, stage, s.sizebits, wb,
6005                     "vst" # n.nvecs # stage # "." # s.lanesize>;
6006}
6007
6008def SDTARMVST2    : SDTypeProfile<1, 5, [SDTCisPtrTy<0>, SDTCisPtrTy<1>, SDTCisVT<2, i32>, SDTCisVec<3>,
6009                                         SDTCisSameAs<3, 4>, SDTCisVT<5, i32>]>;
6010def SDTARMVST4    : SDTypeProfile<1, 7, [SDTCisPtrTy<0>, SDTCisPtrTy<1>, SDTCisVT<2, i32>, SDTCisVec<3>,
6011                                         SDTCisSameAs<3, 4>, SDTCisSameAs<3, 5>,
6012                                         SDTCisSameAs<3, 6>, SDTCisVT<7, i32>]>;
6013def MVEVST2UPD       : SDNode<"ARMISD::VST2_UPD", SDTARMVST2, [SDNPHasChain, SDNPMemOperand]>;
6014def MVEVST4UPD       : SDNode<"ARMISD::VST4_UPD", SDTARMVST4, [SDNPHasChain, SDNPMemOperand]>;
6015
6016multiclass MVE_vst24_patterns<int lanesize, ValueType VT> {
6017  foreach stage = [0,1] in
6018    def : Pat<(int_arm_mve_vst2q i32:$addr,
6019                (VT MQPR:$v0), (VT MQPR:$v1), (i32 stage)),
6020              (!cast<Instruction>("MVE_VST2"#stage#"_"#lanesize)
6021                (REG_SEQUENCE MQQPR, VT:$v0, qsub_0, VT:$v1, qsub_1),
6022                t2_addr_offset_none:$addr)>;
6023  foreach stage = [0,1] in
6024    def : Pat<(i32 (MVEVST2UPD i32:$addr, (i32 32),
6025                (VT MQPR:$v0), (VT MQPR:$v1), (i32 stage))),
6026              (i32 (!cast<Instruction>("MVE_VST2"#stage#"_"#lanesize#_wb)
6027                (REG_SEQUENCE MQQPR, VT:$v0, qsub_0, VT:$v1, qsub_1),
6028                t2_addr_offset_none:$addr))>;
6029
6030  foreach stage = [0,1,2,3] in
6031    def : Pat<(int_arm_mve_vst4q i32:$addr,
6032                (VT MQPR:$v0), (VT MQPR:$v1),
6033                (VT MQPR:$v2), (VT MQPR:$v3), (i32 stage)),
6034              (!cast<Instruction>("MVE_VST4"#stage#"_"#lanesize)
6035                (REG_SEQUENCE MQQQQPR, VT:$v0, qsub_0, VT:$v1, qsub_1,
6036                                       VT:$v2, qsub_2, VT:$v3, qsub_3),
6037                t2_addr_offset_none:$addr)>;
6038  foreach stage = [0,1,2,3] in
6039    def : Pat<(i32 (MVEVST4UPD i32:$addr, (i32 64),
6040                (VT MQPR:$v0), (VT MQPR:$v1),
6041                (VT MQPR:$v2), (VT MQPR:$v3), (i32 stage))),
6042              (i32 (!cast<Instruction>("MVE_VST4"#stage#"_"#lanesize#_wb)
6043                (REG_SEQUENCE MQQQQPR, VT:$v0, qsub_0, VT:$v1, qsub_1,
6044                                       VT:$v2, qsub_2, VT:$v3, qsub_3),
6045                t2_addr_offset_none:$addr))>;
6046}
6047defm : MVE_vst24_patterns<8, v16i8>;
6048defm : MVE_vst24_patterns<16, v8i16>;
6049defm : MVE_vst24_patterns<32, v4i32>;
6050defm : MVE_vst24_patterns<16, v8f16>;
6051defm : MVE_vst24_patterns<32, v4f32>;
6052
6053// end of MVE interleaving load/store
6054
6055// start of MVE predicable load/store
6056
6057// A parameter class for the direction of transfer.
6058class MVE_ldst_direction<bit b, dag Oo, dag Io, string c=""> {
6059  bit load = b;
6060  dag Oops = Oo;
6061  dag Iops = Io;
6062  string cstr = c;
6063}
6064def MVE_ld: MVE_ldst_direction<1, (outs MQPR:$Qd), (ins), ",@earlyclobber $Qd">;
6065def MVE_st: MVE_ldst_direction<0, (outs), (ins MQPR:$Qd)>;
6066
6067// A parameter class for the size of memory access in a load.
6068class MVE_memsz<bits<2> e, int s, AddrMode m, string mn, list<string> types> {
6069  bits<2> encoding = e;         // opcode bit(s) for encoding
6070  int shift = s;                // shift applied to immediate load offset
6071  AddrMode AM = m;
6072
6073  // For instruction aliases: define the complete list of type
6074  // suffixes at this size, and the canonical ones for loads and
6075  // stores.
6076  string MnemonicLetter = mn;
6077  int TypeBits = !shl(8, s);
6078  string CanonLoadSuffix = ".u" # TypeBits;
6079  string CanonStoreSuffix = "." # TypeBits;
6080  list<string> suffixes = !foreach(letter, types, "." # letter # TypeBits);
6081}
6082
6083// Instances of MVE_memsz.
6084//
6085// (memD doesn't need an AddrMode, because those are only for
6086// contiguous loads, and memD is only used by gather/scatters.)
6087def MVE_memB: MVE_memsz<0b00, 0, AddrModeT2_i7,   "b", ["", "u", "s"]>;
6088def MVE_memH: MVE_memsz<0b01, 1, AddrModeT2_i7s2, "h", ["", "u", "s", "f"]>;
6089def MVE_memW: MVE_memsz<0b10, 2, AddrModeT2_i7s4, "w", ["", "u", "s", "f"]>;
6090def MVE_memD: MVE_memsz<0b11, 3, ?,               "d", ["", "u", "s", "f"]>;
6091
6092// This is the base class for all the MVE loads and stores other than
6093// the interleaving ones. All the non-interleaving loads/stores share
6094// the characteristic that they operate on just one vector register,
6095// so they are VPT-predicable.
6096//
6097// The predication operand is vpred_n, for both loads and stores. For
6098// store instructions, the reason is obvious: if there is no output
6099// register, there can't be a need for an input parameter giving the
6100// output register's previous value. Load instructions also don't need
6101// that input parameter, because unlike MVE data processing
6102// instructions, predicated loads are defined to set the inactive
6103// lanes of the output register to zero, instead of preserving their
6104// input values.
6105class MVE_VLDRSTR_base<MVE_ldst_direction dir, bit U, bit P, bit W, bit opc,
6106                       dag oops, dag iops, string asm, string suffix,
6107                       string ops, string cstr, bits<2> vecsize, list<dag> pattern=[]>
6108 : MVE_p<oops, iops, NoItinerary, asm, suffix, ops, vpred_n, cstr, vecsize, pattern> {
6109  bits<3> Qd;
6110
6111  let Inst{28} = U;
6112  let Inst{25} = 0b0;
6113  let Inst{24} = P;
6114  let Inst{22} = 0b0;
6115  let Inst{21} = W;
6116  let Inst{20} = dir.load;
6117  let Inst{15-13} = Qd{2-0};
6118  let Inst{12} = opc;
6119  let Inst{11-9} = 0b111;
6120
6121  let mayLoad = dir.load;
6122  let mayStore = !eq(dir.load,0);
6123  let hasSideEffects = 0;
6124  let validForTailPredication = 1;
6125}
6126
6127// Contiguous load and store instructions. These come in two main
6128// categories: same-size loads/stores in which 128 bits of vector
6129// register is transferred to or from 128 bits of memory in the most
6130// obvious way, and widening loads / narrowing stores, in which the
6131// size of memory accessed is less than the size of a vector register,
6132// so the load instructions sign- or zero-extend each memory value
6133// into a wider vector lane, and the store instructions truncate
6134// correspondingly.
6135//
6136// The instruction mnemonics for these two classes look reasonably
6137// similar, but the actual encodings are different enough to need two
6138// separate base classes.
6139
6140// Contiguous, same size
6141class MVE_VLDRSTR_cs<MVE_ldst_direction dir, MVE_memsz memsz, bit P, bit W,
6142                     dag oops, dag iops, string asm, string suffix,
6143                     IndexMode im, string ops, string cstr>
6144  : MVE_VLDRSTR_base<dir, 0, P, W, 1, oops, iops, asm, suffix, ops, cstr, memsz.encoding> {
6145  bits<12> addr;
6146  let Inst{23} = addr{7};
6147  let Inst{19-16} = addr{11-8};
6148  let Inst{8-7} = memsz.encoding;
6149  let Inst{6-0} = addr{6-0};
6150
6151  let IM = im;
6152}
6153
6154// Contiguous, widening/narrowing
6155class MVE_VLDRSTR_cw<MVE_ldst_direction dir, MVE_memsz memsz, bit U,
6156                     bit P, bit W, bits<2> size, dag oops, dag iops,
6157                     string asm, string suffix, IndexMode im,
6158                     string ops, string cstr>
6159  : MVE_VLDRSTR_base<dir, U, P, W, 0, oops, iops, asm, suffix, ops, cstr, size> {
6160  bits<11> addr;
6161  let Inst{23} = addr{7};
6162  let Inst{19} = memsz.encoding{0}; // enough to tell 16- from 32-bit
6163  let Inst{18-16} = addr{10-8};
6164  let Inst{8-7} = size;
6165  let Inst{6-0} = addr{6-0};
6166
6167  let IM = im;
6168}
6169
6170// Multiclass wrapper on each of the _cw and _cs base classes, to
6171// generate three writeback modes (none, preindex, postindex).
6172
6173multiclass MVE_VLDRSTR_cw_m<MVE_ldst_direction dir, MVE_memsz memsz,
6174                            string asm, string suffix, bit U, bits<2> size> {
6175  let AM = memsz.AM in {
6176    def "" : MVE_VLDRSTR_cw<
6177        dir, memsz, U, 1, 0, size,
6178        dir.Oops, !con(dir.Iops, (ins taddrmode_imm7<memsz.shift>:$addr)),
6179        asm, suffix, IndexModeNone, "$Qd, $addr", "">;
6180
6181    def _pre : MVE_VLDRSTR_cw<
6182        dir, memsz, U, 1, 1, size,
6183        !con((outs tGPR:$wb), dir.Oops),
6184        !con(dir.Iops, (ins taddrmode_imm7<memsz.shift>:$addr)),
6185        asm, suffix, IndexModePre, "$Qd, $addr!", "$addr.base = $wb"> {
6186      let DecoderMethod = "DecodeMVE_MEM_1_pre<"#memsz.shift#">";
6187    }
6188
6189    def _post : MVE_VLDRSTR_cw<
6190        dir, memsz, U, 0, 1, size,
6191        !con((outs tGPR:$wb), dir.Oops),
6192        !con(dir.Iops, (ins t_addr_offset_none:$Rn,
6193                            t2am_imm7_offset<memsz.shift>:$addr)),
6194        asm, suffix, IndexModePost, "$Qd, $Rn$addr", "$Rn.base = $wb"> {
6195      bits<4> Rn;
6196      let Inst{18-16} = Rn{2-0};
6197    }
6198  }
6199}
6200
6201multiclass MVE_VLDRSTR_cs_m<MVE_ldst_direction dir, MVE_memsz memsz,
6202                            string asm, string suffix> {
6203  let AM = memsz.AM in {
6204    def "" : MVE_VLDRSTR_cs<
6205        dir, memsz, 1, 0,
6206        dir.Oops, !con(dir.Iops, (ins t2addrmode_imm7<memsz.shift>:$addr)),
6207        asm, suffix, IndexModeNone, "$Qd, $addr", "">;
6208
6209    def _pre : MVE_VLDRSTR_cs<
6210        dir, memsz, 1, 1,
6211        !con((outs rGPR:$wb), dir.Oops),
6212        !con(dir.Iops, (ins t2addrmode_imm7_pre<memsz.shift>:$addr)),
6213        asm, suffix, IndexModePre, "$Qd, $addr!", "$addr.base = $wb"> {
6214      let DecoderMethod = "DecodeMVE_MEM_2_pre<"#memsz.shift#">";
6215    }
6216
6217    def _post : MVE_VLDRSTR_cs<
6218        dir, memsz, 0, 1,
6219        !con((outs rGPR:$wb), dir.Oops),
6220        !con(dir.Iops, (ins t2_nosp_addr_offset_none:$Rn,
6221                            t2am_imm7_offset<memsz.shift>:$addr)),
6222        asm, suffix, IndexModePost, "$Qd, $Rn$addr", "$Rn.base = $wb"> {
6223      bits<4> Rn;
6224      let Inst{19-16} = Rn{3-0};
6225    }
6226  }
6227}
6228
6229// Now actually declare all the contiguous load/stores, via those
6230// multiclasses. The instruction ids coming out of this are the bare
6231// names shown in the defm, with _pre or _post appended for writeback,
6232// e.g. MVE_VLDRBS16, MVE_VSTRB16_pre, MVE_VSTRHU16_post.
6233
6234defm MVE_VLDRBS16: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memB, "vldrb", "s16", 0, 0b01>;
6235defm MVE_VLDRBS32: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memB, "vldrb", "s32", 0, 0b10>;
6236defm MVE_VLDRBU16: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memB, "vldrb", "u16", 1, 0b01>;
6237defm MVE_VLDRBU32: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memB, "vldrb", "u32", 1, 0b10>;
6238defm MVE_VLDRHS32: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memH, "vldrh", "s32", 0, 0b10>;
6239defm MVE_VLDRHU32: MVE_VLDRSTR_cw_m<MVE_ld, MVE_memH, "vldrh", "u32", 1, 0b10>;
6240
6241defm MVE_VLDRBU8:  MVE_VLDRSTR_cs_m<MVE_ld, MVE_memB, "vldrb", "u8">;
6242defm MVE_VLDRHU16: MVE_VLDRSTR_cs_m<MVE_ld, MVE_memH, "vldrh", "u16">;
6243defm MVE_VLDRWU32: MVE_VLDRSTR_cs_m<MVE_ld, MVE_memW, "vldrw", "u32">;
6244
6245defm MVE_VSTRB16:  MVE_VLDRSTR_cw_m<MVE_st, MVE_memB, "vstrb", "16",  0, 0b01>;
6246defm MVE_VSTRB32:  MVE_VLDRSTR_cw_m<MVE_st, MVE_memB, "vstrb", "32",  0, 0b10>;
6247defm MVE_VSTRH32:  MVE_VLDRSTR_cw_m<MVE_st, MVE_memH, "vstrh", "32",  0, 0b10>;
6248
6249defm MVE_VSTRBU8 : MVE_VLDRSTR_cs_m<MVE_st, MVE_memB, "vstrb", "8">;
6250defm MVE_VSTRHU16: MVE_VLDRSTR_cs_m<MVE_st, MVE_memH, "vstrh", "16">;
6251defm MVE_VSTRWU32: MVE_VLDRSTR_cs_m<MVE_st, MVE_memW, "vstrw", "32">;
6252
6253// Gather loads / scatter stores whose address operand is of the form
6254// [Rn,Qm], i.e. a single GPR as the common base address, plus a
6255// vector of offset from it. ('Load/store this sequence of elements of
6256// the same array.')
6257//
6258// Like the contiguous family, these loads and stores can widen the
6259// loaded values / truncate the stored ones, or they can just
6260// load/store the same size of memory and vector lane. But unlike the
6261// contiguous family, there's no particular difference in encoding
6262// between those two cases.
6263//
6264// This family also comes with the option to scale the offset values
6265// in Qm by the size of the loaded memory (i.e. to treat them as array
6266// indices), or not to scale them (to treat them as plain byte offsets
6267// in memory, so that perhaps the loaded values are unaligned). The
6268// scaled instructions' address operand in assembly looks like
6269// [Rn,Qm,UXTW #2] or similar.
6270
6271// Base class.
6272class MVE_VLDRSTR_rq<MVE_ldst_direction dir, MVE_memsz memsz, bit U,
6273                     bits<2> size, bit os, string asm, string suffix, int shift>
6274  : MVE_VLDRSTR_base<dir, U, 0b0, 0b0, 0, dir.Oops,
6275                     !con(dir.Iops, (ins mve_addr_rq_shift<shift>:$addr)),
6276                     asm, suffix, "$Qd, $addr", dir.cstr, size> {
6277  bits<7> addr;
6278  let Inst{23} = 0b1;
6279  let Inst{19-16} = addr{6-3};
6280  let Inst{8-7} = size;
6281  let Inst{6} = memsz.encoding{1};
6282  let Inst{5} = 0;
6283  let Inst{4} = memsz.encoding{0};
6284  let Inst{3-1} = addr{2-0};
6285  let Inst{0} = os;
6286}
6287
6288// Multiclass that defines the scaled and unscaled versions of an
6289// instruction, when the memory size is wider than a byte. The scaled
6290// version gets the default name like MVE_VLDRBU16_rq; the unscaled /
6291// potentially unaligned version gets a "_u" suffix, e.g.
6292// MVE_VLDRBU16_rq_u.
6293multiclass MVE_VLDRSTR_rq_w<MVE_ldst_direction dir, MVE_memsz memsz,
6294                            string asm, string suffix, bit U, bits<2> size> {
6295  def _u : MVE_VLDRSTR_rq<dir, memsz, U, size, 0, asm, suffix, 0>;
6296  def "" : MVE_VLDRSTR_rq<dir, memsz, U, size, 1, asm, suffix, memsz.shift>;
6297}
6298
6299// Subclass of MVE_VLDRSTR_rq with the same API as that multiclass,
6300// for use when the memory size is one byte, so there's no 'scaled'
6301// version of the instruction at all. (This is encoded as if it were
6302// unscaled, but named in the default way with no _u suffix.)
6303class MVE_VLDRSTR_rq_b<MVE_ldst_direction dir, MVE_memsz memsz,
6304                       string asm, string suffix, bit U, bits<2> size>
6305  : MVE_VLDRSTR_rq<dir, memsz, U, size, 0, asm, suffix, 0>;
6306
6307// Multiclasses wrapping that to add ISel patterns for intrinsics.
6308multiclass MVE_VLDR_rq_w<MVE_memsz memsz, list<MVEVectorVTInfo> VTIs> {
6309  defm "": MVE_VLDRSTR_rq_w<MVE_ld, memsz, "vldr" # memsz.MnemonicLetter,
6310                            VTIs[0].Suffix, VTIs[0].Unsigned, VTIs[0].Size>;
6311  defvar Inst = !cast<Instruction>(NAME);
6312  defvar InstU = !cast<Instruction>(NAME # "_u");
6313
6314  foreach VTI = VTIs in
6315  foreach UnsignedFlag = !if(!eq(VTI.Size, memsz.encoding),
6316                             [0,1], [VTI.Unsigned]) in {
6317    def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), memsz.TypeBits, 0, UnsignedFlag)),
6318              (VTI.Vec (InstU GPR:$base, MQPR:$offsets))>;
6319    def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), memsz.TypeBits, memsz.shift, UnsignedFlag)),
6320              (VTI.Vec (Inst GPR:$base, MQPR:$offsets))>;
6321    def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), memsz.TypeBits, 0, UnsignedFlag, (VTI.Pred VCCR:$pred))),
6322              (VTI.Vec (InstU GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred, zero_reg))>;
6323    def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), memsz.TypeBits, memsz.shift, UnsignedFlag, (VTI.Pred VCCR:$pred))),
6324              (VTI.Vec (Inst GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred, zero_reg))>;
6325  }
6326}
6327multiclass MVE_VLDR_rq_b<list<MVEVectorVTInfo> VTIs> {
6328  def "": MVE_VLDRSTR_rq_b<MVE_ld, MVE_memB, "vldrb",
6329                           VTIs[0].Suffix, VTIs[0].Unsigned, VTIs[0].Size>;
6330  defvar Inst = !cast<Instruction>(NAME);
6331
6332  foreach VTI = VTIs in {
6333    def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), 8, 0, VTI.Unsigned)),
6334              (VTI.Vec (Inst GPR:$base, MQPR:$offsets))>;
6335    def : Pat<(VTI.Vec (int_arm_mve_vldr_gather_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), 8, 0, VTI.Unsigned, (VTI.Pred VCCR:$pred))),
6336              (VTI.Vec (Inst GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred, zero_reg))>;
6337  }
6338}
6339multiclass MVE_VSTR_rq_w<MVE_memsz memsz, list<MVEVectorVTInfo> VTIs> {
6340  defm "": MVE_VLDRSTR_rq_w<MVE_st, memsz, "vstr" # memsz.MnemonicLetter,
6341                            VTIs[0].BitsSuffix, 0, VTIs[0].Size>;
6342  defvar Inst = !cast<Instruction>(NAME);
6343  defvar InstU = !cast<Instruction>(NAME # "_u");
6344
6345  foreach VTI = VTIs in {
6346    def : Pat<(int_arm_mve_vstr_scatter_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), memsz.TypeBits, 0),
6347              (InstU MQPR:$data, GPR:$base, MQPR:$offsets)>;
6348    def : Pat<(int_arm_mve_vstr_scatter_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), memsz.TypeBits, memsz.shift),
6349              (Inst MQPR:$data, GPR:$base, MQPR:$offsets)>;
6350    def : Pat<(int_arm_mve_vstr_scatter_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), memsz.TypeBits, 0, (VTI.Pred VCCR:$pred)),
6351              (InstU MQPR:$data, GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred, zero_reg)>;
6352    def : Pat<(int_arm_mve_vstr_scatter_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), memsz.TypeBits, memsz.shift, (VTI.Pred VCCR:$pred)),
6353              (Inst MQPR:$data, GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred, zero_reg)>;
6354  }
6355}
6356multiclass MVE_VSTR_rq_b<list<MVEVectorVTInfo> VTIs> {
6357  def "": MVE_VLDRSTR_rq_b<MVE_st, MVE_memB, "vstrb",
6358                           VTIs[0].BitsSuffix, 0, VTIs[0].Size>;
6359  defvar Inst = !cast<Instruction>(NAME);
6360
6361  foreach VTI = VTIs in {
6362    def : Pat<(int_arm_mve_vstr_scatter_offset GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), 8, 0),
6363              (Inst MQPR:$data, GPR:$base, MQPR:$offsets)>;
6364    def : Pat<(int_arm_mve_vstr_scatter_offset_predicated GPR:$base, (VTIs[0].Vec MQPR:$offsets), (VTI.Vec MQPR:$data), 8, 0, (VTI.Pred VCCR:$pred)),
6365              (Inst MQPR:$data, GPR:$base, MQPR:$offsets, ARMVCCThen, VCCR:$pred, zero_reg)>;
6366  }
6367}
6368
6369// Actually define all the loads and stores in this family.
6370
6371defm MVE_VLDRBU8_rq : MVE_VLDR_rq_b<[MVE_v16u8,MVE_v16s8]>;
6372defm MVE_VLDRBU16_rq: MVE_VLDR_rq_b<[MVE_v8u16]>;
6373defm MVE_VLDRBS16_rq: MVE_VLDR_rq_b<[MVE_v8s16]>;
6374defm MVE_VLDRBU32_rq: MVE_VLDR_rq_b<[MVE_v4u32]>;
6375defm MVE_VLDRBS32_rq: MVE_VLDR_rq_b<[MVE_v4s32]>;
6376
6377defm MVE_VLDRHU16_rq: MVE_VLDR_rq_w<MVE_memH, [MVE_v8u16,MVE_v8s16,MVE_v8f16]>;
6378defm MVE_VLDRHU32_rq: MVE_VLDR_rq_w<MVE_memH, [MVE_v4u32]>;
6379defm MVE_VLDRHS32_rq: MVE_VLDR_rq_w<MVE_memH, [MVE_v4s32]>;
6380defm MVE_VLDRWU32_rq: MVE_VLDR_rq_w<MVE_memW, [MVE_v4u32,MVE_v4s32,MVE_v4f32]>;
6381defm MVE_VLDRDU64_rq: MVE_VLDR_rq_w<MVE_memD, [MVE_v2u64,MVE_v2s64]>;
6382
6383defm MVE_VSTRB8_rq  : MVE_VSTR_rq_b<[MVE_v16i8]>;
6384defm MVE_VSTRB16_rq : MVE_VSTR_rq_b<[MVE_v8i16]>;
6385defm MVE_VSTRB32_rq : MVE_VSTR_rq_b<[MVE_v4i32]>;
6386
6387defm MVE_VSTRH16_rq : MVE_VSTR_rq_w<MVE_memH, [MVE_v8i16,MVE_v8f16]>;
6388defm MVE_VSTRH32_rq : MVE_VSTR_rq_w<MVE_memH, [MVE_v4i32]>;
6389defm MVE_VSTRW32_rq : MVE_VSTR_rq_w<MVE_memW, [MVE_v4i32,MVE_v4f32]>;
6390defm MVE_VSTRD64_rq : MVE_VSTR_rq_w<MVE_memD, [MVE_v2i64]>;
6391
6392// Gather loads / scatter stores whose address operand is of the form
6393// [Qm,#imm], i.e. a vector containing a full base address for each
6394// loaded item, plus an immediate offset applied consistently to all
6395// of them. ('Load/store the same field from this vector of pointers
6396// to a structure type.')
6397//
6398// This family requires the vector lane size to be at least 32 bits
6399// (so there's room for an address in each lane at all). It has no
6400// widening/narrowing variants. But it does support preindex
6401// writeback, in which the address vector is updated to hold the
6402// addresses actually loaded from.
6403
6404// Base class.
6405class MVE_VLDRSTR_qi<MVE_ldst_direction dir, MVE_memsz memsz, bit W, dag wbops,
6406                     string asm, string wbAsm, string suffix, string cstr = "">
6407  : MVE_VLDRSTR_base<dir, 1, 1, W, 1, !con(wbops, dir.Oops),
6408                     !con(dir.Iops, (ins mve_addr_q_shift<memsz.shift>:$addr)),
6409                     asm, suffix, "$Qd, $addr" # wbAsm, cstr # dir.cstr, memsz.encoding> {
6410  bits<11> addr;
6411  let Inst{23} = addr{7};
6412  let Inst{19-17} = addr{10-8};
6413  let Inst{16} = 0;
6414  let Inst{8} = memsz.encoding{0}; // enough to distinguish 32- from 64-bit
6415  let Inst{7} = 0;
6416  let Inst{6-0} = addr{6-0};
6417}
6418
6419// Multiclass that generates the non-writeback and writeback variants.
6420multiclass MVE_VLDRSTR_qi_m<MVE_ldst_direction dir, MVE_memsz memsz,
6421                            string asm, string suffix> {
6422  def ""   : MVE_VLDRSTR_qi<dir, memsz, 0, (outs),          asm, "",  suffix>;
6423  def _pre : MVE_VLDRSTR_qi<dir, memsz, 1, (outs MQPR:$wb), asm, "!", suffix,
6424                            "$addr.base = $wb"> {
6425    let DecoderMethod="DecodeMVE_MEM_3_pre<"#memsz.shift#">";
6426  }
6427}
6428
6429// Multiclasses wrapping that one, adding selection patterns for the
6430// non-writeback loads and all the stores. (The writeback loads must
6431// deliver multiple output values, so they have to be selected by C++
6432// code.)
6433multiclass MVE_VLDR_qi<MVE_memsz memsz, MVEVectorVTInfo AVTI,
6434                       list<MVEVectorVTInfo> DVTIs> {
6435  defm "" : MVE_VLDRSTR_qi_m<MVE_ld, memsz, "vldr" # memsz.MnemonicLetter,
6436                             "u" # memsz.TypeBits>;
6437  defvar Inst = !cast<Instruction>(NAME);
6438
6439  foreach DVTI = DVTIs in {
6440    def : Pat<(DVTI.Vec (int_arm_mve_vldr_gather_base
6441                 (AVTI.Vec MQPR:$addr), (i32 imm:$offset))),
6442              (DVTI.Vec (Inst (AVTI.Vec MQPR:$addr), (i32 imm:$offset)))>;
6443    def : Pat<(DVTI.Vec (int_arm_mve_vldr_gather_base_predicated
6444                 (AVTI.Vec MQPR:$addr), (i32 imm:$offset), (AVTI.Pred VCCR:$pred))),
6445              (DVTI.Vec (Inst (AVTI.Vec MQPR:$addr), (i32 imm:$offset),
6446                        ARMVCCThen, VCCR:$pred, zero_reg))>;
6447  }
6448}
6449multiclass MVE_VSTR_qi<MVE_memsz memsz, MVEVectorVTInfo AVTI,
6450                       list<MVEVectorVTInfo> DVTIs> {
6451  defm "" : MVE_VLDRSTR_qi_m<MVE_st, memsz, "vstr" # memsz.MnemonicLetter,
6452                             !cast<string>(memsz.TypeBits)>;
6453  defvar Inst = !cast<Instruction>(NAME);
6454  defvar InstPre = !cast<Instruction>(NAME # "_pre");
6455
6456  foreach DVTI = DVTIs in {
6457    def : Pat<(int_arm_mve_vstr_scatter_base
6458                (AVTI.Vec MQPR:$addr), (i32 imm:$offset), (DVTI.Vec MQPR:$data)),
6459              (Inst (DVTI.Vec MQPR:$data), (AVTI.Vec MQPR:$addr),
6460                    (i32 imm:$offset))>;
6461    def : Pat<(int_arm_mve_vstr_scatter_base_predicated
6462                (AVTI.Vec MQPR:$addr), (i32 imm:$offset), (DVTI.Vec MQPR:$data), (AVTI.Pred VCCR:$pred)),
6463              (Inst (DVTI.Vec MQPR:$data), (AVTI.Vec MQPR:$addr),
6464                    (i32 imm:$offset), ARMVCCThen, VCCR:$pred, zero_reg)>;
6465    def : Pat<(AVTI.Vec (int_arm_mve_vstr_scatter_base_wb
6466                (AVTI.Vec MQPR:$addr), (i32 imm:$offset), (DVTI.Vec MQPR:$data))),
6467              (AVTI.Vec (InstPre (DVTI.Vec MQPR:$data), (AVTI.Vec MQPR:$addr),
6468                                 (i32 imm:$offset)))>;
6469    def : Pat<(AVTI.Vec (int_arm_mve_vstr_scatter_base_wb_predicated
6470                (AVTI.Vec MQPR:$addr), (i32 imm:$offset), (DVTI.Vec MQPR:$data), (AVTI.Pred VCCR:$pred))),
6471              (AVTI.Vec (InstPre (DVTI.Vec MQPR:$data), (AVTI.Vec MQPR:$addr),
6472                                 (i32 imm:$offset), ARMVCCThen, VCCR:$pred, zero_reg))>;
6473  }
6474}
6475
6476// Actual instruction definitions.
6477defm MVE_VLDRWU32_qi: MVE_VLDR_qi<MVE_memW, MVE_v4i32, [MVE_v4i32,MVE_v4f32]>;
6478defm MVE_VLDRDU64_qi: MVE_VLDR_qi<MVE_memD, MVE_v2i64, [MVE_v2i64,MVE_v2f64]>;
6479defm MVE_VSTRW32_qi:  MVE_VSTR_qi<MVE_memW, MVE_v4i32, [MVE_v4i32,MVE_v4f32]>;
6480defm MVE_VSTRD64_qi:  MVE_VSTR_qi<MVE_memD, MVE_v2i64, [MVE_v2i64,MVE_v2f64]>;
6481
6482// Define aliases for all the instructions where memory size and
6483// vector lane size are the same. These are mnemonic aliases, so they
6484// apply consistently across all of the above families - contiguous
6485// loads, and both the rq and qi types of gather/scatter.
6486//
6487// Rationale: As long as you're loading (for example) 16-bit memory
6488// values into 16-bit vector lanes, you can think of them as signed or
6489// unsigned integers, fp16 or just raw 16-bit blobs and it makes no
6490// difference. So we permit all of vldrh.16, vldrh.u16, vldrh.s16,
6491// vldrh.f16 and treat them all as equivalent to the canonical
6492// spelling (which happens to be .u16 for loads, and just .16 for
6493// stores).
6494
6495foreach vpt_cond = ["", "t", "e"] in
6496foreach memsz = [MVE_memB, MVE_memH, MVE_memW, MVE_memD] in
6497foreach suffix = memsz.suffixes in {
6498  // Define an alias with every suffix in the list, except for the one
6499  // used by the real Instruction record (i.e. the one that all the
6500  // rest are aliases *for*).
6501
6502  if !ne(suffix, memsz.CanonLoadSuffix) then {
6503    def : MnemonicAlias<
6504      "vldr" # memsz.MnemonicLetter # vpt_cond # suffix,
6505      "vldr" # memsz.MnemonicLetter # vpt_cond # memsz.CanonLoadSuffix>;
6506  }
6507
6508  if !ne(suffix, memsz.CanonStoreSuffix) then {
6509    def : MnemonicAlias<
6510      "vstr" # memsz.MnemonicLetter # vpt_cond # suffix,
6511      "vstr" # memsz.MnemonicLetter # vpt_cond # memsz.CanonStoreSuffix>;
6512  }
6513}
6514
6515// end of MVE predicable load/store
6516
6517class MVE_VPT<string suffix, bits<2> size, dag iops, string asm, list<dag> pattern=[]>
6518  : MVE_MI<(outs ), iops, NoItinerary, !strconcat("vpt", "${Mk}", ".", suffix), asm, "", size, pattern> {
6519  bits<3> fc;
6520  bits<4> Mk;
6521  bits<3> Qn;
6522
6523  let Inst{31-23} = 0b111111100;
6524  let Inst{22} = Mk{3};
6525  let Inst{21-20} = size;
6526  let Inst{19-17} = Qn{2-0};
6527  let Inst{16} = 0b1;
6528  let Inst{15-13} = Mk{2-0};
6529  let Inst{12} = fc{2};
6530  let Inst{11-8} = 0b1111;
6531  let Inst{7} = fc{0};
6532  let Inst{4} = 0b0;
6533
6534  let Defs = [VPR];
6535  let validForTailPredication=1;
6536}
6537
6538class MVE_VPTt1<string suffix, bits<2> size, dag iops>
6539  : MVE_VPT<suffix, size, iops, "$fc, $Qn, $Qm"> {
6540  bits<4> Qm;
6541  bits<4> Mk;
6542
6543  let Inst{6} = 0b0;
6544  let Inst{5} = Qm{3};
6545  let Inst{3-1} = Qm{2-0};
6546  let Inst{0} = fc{1};
6547}
6548
6549class MVE_VPTt1i<string suffix, bits<2> size>
6550 : MVE_VPTt1<suffix, size,
6551           (ins vpt_mask:$Mk, MQPR:$Qn, MQPR:$Qm, pred_basic_i:$fc)> {
6552  let Inst{12} = 0b0;
6553  let Inst{0} = 0b0;
6554}
6555
6556def MVE_VPTv4i32 : MVE_VPTt1i<"i32", 0b10>;
6557def MVE_VPTv8i16 : MVE_VPTt1i<"i16", 0b01>;
6558def MVE_VPTv16i8 : MVE_VPTt1i<"i8", 0b00>;
6559
6560class MVE_VPTt1u<string suffix, bits<2> size>
6561 : MVE_VPTt1<suffix, size,
6562           (ins vpt_mask:$Mk, MQPR:$Qn, MQPR:$Qm, pred_basic_u:$fc)> {
6563  let Inst{12} = 0b0;
6564  let Inst{0} = 0b1;
6565}
6566
6567def MVE_VPTv4u32 : MVE_VPTt1u<"u32", 0b10>;
6568def MVE_VPTv8u16 : MVE_VPTt1u<"u16", 0b01>;
6569def MVE_VPTv16u8 : MVE_VPTt1u<"u8", 0b00>;
6570
6571class MVE_VPTt1s<string suffix, bits<2> size>
6572 : MVE_VPTt1<suffix, size,
6573           (ins vpt_mask:$Mk, MQPR:$Qn, MQPR:$Qm, pred_basic_s:$fc)> {
6574  let Inst{12} = 0b1;
6575}
6576
6577def MVE_VPTv4s32 : MVE_VPTt1s<"s32", 0b10>;
6578def MVE_VPTv8s16 : MVE_VPTt1s<"s16", 0b01>;
6579def MVE_VPTv16s8 : MVE_VPTt1s<"s8", 0b00>;
6580
6581class MVE_VPTt2<string suffix, bits<2> size, dag iops>
6582  : MVE_VPT<suffix, size, iops,
6583          "$fc, $Qn, $Rm"> {
6584  bits<4> Rm;
6585  bits<3> fc;
6586  bits<4> Mk;
6587
6588  let Inst{6} = 0b1;
6589  let Inst{5} = fc{1};
6590  let Inst{3-0} = Rm{3-0};
6591}
6592
6593class MVE_VPTt2i<string suffix, bits<2> size>
6594  : MVE_VPTt2<suffix, size,
6595            (ins vpt_mask:$Mk, MQPR:$Qn, GPRwithZR:$Rm, pred_basic_i:$fc)> {
6596  let Inst{12} = 0b0;
6597  let Inst{5} = 0b0;
6598}
6599
6600def MVE_VPTv4i32r : MVE_VPTt2i<"i32", 0b10>;
6601def MVE_VPTv8i16r : MVE_VPTt2i<"i16", 0b01>;
6602def MVE_VPTv16i8r : MVE_VPTt2i<"i8", 0b00>;
6603
6604class MVE_VPTt2u<string suffix, bits<2> size>
6605  : MVE_VPTt2<suffix, size,
6606            (ins vpt_mask:$Mk, MQPR:$Qn, GPRwithZR:$Rm, pred_basic_u:$fc)> {
6607  let Inst{12} = 0b0;
6608  let Inst{5} = 0b1;
6609}
6610
6611def MVE_VPTv4u32r : MVE_VPTt2u<"u32", 0b10>;
6612def MVE_VPTv8u16r : MVE_VPTt2u<"u16", 0b01>;
6613def MVE_VPTv16u8r : MVE_VPTt2u<"u8", 0b00>;
6614
6615class MVE_VPTt2s<string suffix, bits<2> size>
6616  : MVE_VPTt2<suffix, size,
6617            (ins vpt_mask:$Mk, MQPR:$Qn, GPRwithZR:$Rm, pred_basic_s:$fc)> {
6618  let Inst{12} = 0b1;
6619}
6620
6621def MVE_VPTv4s32r : MVE_VPTt2s<"s32", 0b10>;
6622def MVE_VPTv8s16r : MVE_VPTt2s<"s16", 0b01>;
6623def MVE_VPTv16s8r : MVE_VPTt2s<"s8", 0b00>;
6624
6625
6626class MVE_VPTf<string suffix, bit size, dag iops, string asm, list<dag> pattern=[]>
6627  : MVE_MI<(outs ), iops, NoItinerary, !strconcat("vpt", "${Mk}", ".", suffix), asm,
6628            "", !if(size, 0b01, 0b10), pattern> {
6629  bits<3> fc;
6630  bits<4> Mk;
6631  bits<3> Qn;
6632
6633  let Inst{31-29} = 0b111;
6634  let Inst{28} = size;
6635  let Inst{27-23} = 0b11100;
6636  let Inst{22} = Mk{3};
6637  let Inst{21-20} = 0b11;
6638  let Inst{19-17} = Qn{2-0};
6639  let Inst{16} = 0b1;
6640  let Inst{15-13} = Mk{2-0};
6641  let Inst{12} = fc{2};
6642  let Inst{11-8} = 0b1111;
6643  let Inst{7} = fc{0};
6644  let Inst{4} = 0b0;
6645
6646  let Defs = [VPR];
6647  let Predicates = [HasMVEFloat];
6648  let validForTailPredication=1;
6649}
6650
6651class MVE_VPTft1<string suffix, bit size>
6652  : MVE_VPTf<suffix, size, (ins vpt_mask:$Mk, MQPR:$Qn, MQPR:$Qm, pred_basic_fp:$fc),
6653          "$fc, $Qn, $Qm"> {
6654  bits<3> fc;
6655  bits<4> Qm;
6656
6657  let Inst{6} = 0b0;
6658  let Inst{5} = Qm{3};
6659  let Inst{3-1} = Qm{2-0};
6660  let Inst{0} = fc{1};
6661}
6662
6663def MVE_VPTv4f32         : MVE_VPTft1<"f32", 0b0>;
6664def MVE_VPTv8f16         : MVE_VPTft1<"f16", 0b1>;
6665
6666class MVE_VPTft2<string suffix, bit size>
6667  : MVE_VPTf<suffix, size, (ins vpt_mask:$Mk, MQPR:$Qn, GPRwithZR:$Rm, pred_basic_fp:$fc),
6668          "$fc, $Qn, $Rm"> {
6669  bits<3> fc;
6670  bits<4> Rm;
6671
6672  let Inst{6} = 0b1;
6673  let Inst{5} = fc{1};
6674  let Inst{3-0} = Rm{3-0};
6675}
6676
6677def MVE_VPTv4f32r        : MVE_VPTft2<"f32", 0b0>;
6678def MVE_VPTv8f16r        : MVE_VPTft2<"f16", 0b1>;
6679
6680def MVE_VPST : MVE_MI<(outs ), (ins vpt_mask:$Mk), NoItinerary,
6681       !strconcat("vpst", "${Mk}"), "", "", 0b00, []> {
6682  bits<4> Mk;
6683
6684  let Inst{31-23} = 0b111111100;
6685  let Inst{22} = Mk{3};
6686  let Inst{21-16} = 0b110001;
6687  let Inst{15-13} = Mk{2-0};
6688  let Inst{12-0} = 0b0111101001101;
6689  let Unpredictable{12} = 0b1;
6690  let Unpredictable{7} = 0b1;
6691  let Unpredictable{5} = 0b1;
6692
6693  let Uses = [VPR];
6694  let validForTailPredication = 1;
6695}
6696
6697def MVE_VPSEL : MVE_p<(outs MQPR:$Qd), (ins MQPR:$Qn, MQPR:$Qm), NoItinerary,
6698                      "vpsel", "", "$Qd, $Qn, $Qm", vpred_n, "", 0b00, []> {
6699  bits<4> Qn;
6700  bits<4> Qd;
6701  bits<4> Qm;
6702
6703  let Inst{28} = 0b1;
6704  let Inst{25-23} = 0b100;
6705  let Inst{22} = Qd{3};
6706  let Inst{21-20} = 0b11;
6707  let Inst{19-17} = Qn{2-0};
6708  let Inst{16} = 0b1;
6709  let Inst{15-13} = Qd{2-0};
6710  let Inst{12-9} = 0b0111;
6711  let Inst{8} = 0b1;
6712  let Inst{7} = Qn{3};
6713  let Inst{6} = 0b0;
6714  let Inst{5} = Qm{3};
6715  let Inst{4} = 0b0;
6716  let Inst{3-1} = Qm{2-0};
6717  let Inst{0} = 0b1;
6718}
6719
6720foreach suffix = ["s8", "s16", "s32", "u8", "u16", "u32",
6721                  "i8", "i16", "i32",       "f16", "f32"] in
6722def : MVEInstAlias<"vpsel${vp}." # suffix # "\t$Qd, $Qn, $Qm",
6723                   (MVE_VPSEL MQPR:$Qd, MQPR:$Qn, MQPR:$Qm, vpred_n:$vp)>;
6724
6725let Predicates = [HasMVEInt] in {
6726  def : Pat<(v16i8 (vselect (v16i1 VCCR:$pred), (v16i8 MQPR:$v1), (v16i8 MQPR:$v2))),
6727            (v16i8 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone, VCCR:$pred, zero_reg))>;
6728  def : Pat<(v8i16 (vselect (v8i1 VCCR:$pred), (v8i16 MQPR:$v1), (v8i16 MQPR:$v2))),
6729            (v8i16 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone, VCCR:$pred, zero_reg))>;
6730  def : Pat<(v4i32 (vselect (v4i1 VCCR:$pred), (v4i32 MQPR:$v1), (v4i32 MQPR:$v2))),
6731            (v4i32 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone, VCCR:$pred, zero_reg))>;
6732  def : Pat<(v2i64 (vselect (v2i1 VCCR:$pred), (v2i64 MQPR:$v1), (v2i64 MQPR:$v2))),
6733            (v2i64 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone, VCCR:$pred, zero_reg))>;
6734
6735  def : Pat<(v8f16 (vselect (v8i1 VCCR:$pred), (v8f16 MQPR:$v1), (v8f16 MQPR:$v2))),
6736            (v8f16 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone, VCCR:$pred, zero_reg))>;
6737  def : Pat<(v4f32 (vselect (v4i1 VCCR:$pred), (v4f32 MQPR:$v1), (v4f32 MQPR:$v2))),
6738            (v4f32 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone, VCCR:$pred, zero_reg))>;
6739  def : Pat<(v2f64 (vselect (v2i1 VCCR:$pred), (v2f64 MQPR:$v1), (v2f64 MQPR:$v2))),
6740            (v2f64 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone, VCCR:$pred, zero_reg))>;
6741
6742  def : Pat<(v16i8 (vselect (v16i8 MQPR:$pred), (v16i8 MQPR:$v1), (v16i8 MQPR:$v2))),
6743            (v16i8 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone,
6744                              (MVE_VCMPi8 (v16i8 MQPR:$pred), (MVE_VMOVimmi8 0), ARMCCne), zero_reg))>;
6745  def : Pat<(v8i16 (vselect (v8i16 MQPR:$pred), (v8i16 MQPR:$v1), (v8i16 MQPR:$v2))),
6746            (v8i16 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone,
6747                              (MVE_VCMPi16 (v8i16 MQPR:$pred), (MVE_VMOVimmi16 0), ARMCCne), zero_reg))>;
6748  def : Pat<(v4i32 (vselect (v4i32 MQPR:$pred), (v4i32 MQPR:$v1), (v4i32 MQPR:$v2))),
6749            (v4i32 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone,
6750                              (MVE_VCMPi32 (v4i32 MQPR:$pred), (MVE_VMOVimmi32 0), ARMCCne), zero_reg))>;
6751
6752  def : Pat<(v8f16 (vselect (v8i16 MQPR:$pred), (v8f16 MQPR:$v1), (v8f16 MQPR:$v2))),
6753            (v8f16 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone,
6754                              (MVE_VCMPi16 (v8i16 MQPR:$pred), (MVE_VMOVimmi16 0), ARMCCne), zero_reg))>;
6755  def : Pat<(v4f32 (vselect (v4i32 MQPR:$pred), (v4f32 MQPR:$v1), (v4f32 MQPR:$v2))),
6756            (v4f32 (MVE_VPSEL MQPR:$v1, MQPR:$v2, ARMVCCNone,
6757                              (MVE_VCMPi32 (v4i32 MQPR:$pred), (MVE_VMOVimmi32 0), ARMCCne), zero_reg))>;
6758
6759  // Pred <-> Int
6760  def : Pat<(v16i8 (zext  (v16i1 VCCR:$pred))),
6761            (v16i8 (MVE_VPSEL (MVE_VMOVimmi8 1), (MVE_VMOVimmi8 0), ARMVCCNone, VCCR:$pred, zero_reg))>;
6762  def : Pat<(v8i16 (zext  (v8i1  VCCR:$pred))),
6763            (v8i16 (MVE_VPSEL (MVE_VMOVimmi16 1), (MVE_VMOVimmi16 0), ARMVCCNone, VCCR:$pred, zero_reg))>;
6764  def : Pat<(v4i32 (zext  (v4i1  VCCR:$pred))),
6765            (v4i32 (MVE_VPSEL (MVE_VMOVimmi32 1), (MVE_VMOVimmi32 0), ARMVCCNone, VCCR:$pred, zero_reg))>;
6766  def : Pat<(v2i64 (zext  (v2i1  VCCR:$pred))),
6767            (v2i64 (MVE_VPSEL (MVE_VMOVimmi64 1), (MVE_VMOVimmi32 0), ARMVCCNone, VCCR:$pred, zero_reg))>;
6768
6769  def : Pat<(v16i8 (sext  (v16i1 VCCR:$pred))),
6770            (v16i8 (MVE_VPSEL (MVE_VMOVimmi8 255), (MVE_VMOVimmi8 0), ARMVCCNone, VCCR:$pred, zero_reg))>;
6771  def : Pat<(v8i16 (sext  (v8i1  VCCR:$pred))),
6772            (v8i16 (MVE_VPSEL (MVE_VMOVimmi8 255), (MVE_VMOVimmi16 0), ARMVCCNone, VCCR:$pred, zero_reg))>;
6773  def : Pat<(v4i32 (sext  (v4i1  VCCR:$pred))),
6774            (v4i32 (MVE_VPSEL (MVE_VMOVimmi8 255), (MVE_VMOVimmi32 0), ARMVCCNone, VCCR:$pred, zero_reg))>;
6775  def : Pat<(v2i64 (sext  (v2i1  VCCR:$pred))),
6776            (v2i64 (MVE_VPSEL (MVE_VMOVimmi8 255), (MVE_VMOVimmi32 0), ARMVCCNone, VCCR:$pred, zero_reg))>;
6777
6778  def : Pat<(v16i8 (anyext  (v16i1 VCCR:$pred))),
6779            (v16i8 (MVE_VPSEL (MVE_VMOVimmi8 1), (MVE_VMOVimmi8 0), ARMVCCNone, VCCR:$pred, zero_reg))>;
6780  def : Pat<(v8i16 (anyext  (v8i1  VCCR:$pred))),
6781            (v8i16 (MVE_VPSEL (MVE_VMOVimmi16 1), (MVE_VMOVimmi16 0), ARMVCCNone, VCCR:$pred, zero_reg))>;
6782  def : Pat<(v4i32 (anyext  (v4i1  VCCR:$pred))),
6783            (v4i32 (MVE_VPSEL (MVE_VMOVimmi32 1), (MVE_VMOVimmi32 0), ARMVCCNone, VCCR:$pred, zero_reg))>;
6784  def : Pat<(v2i64 (anyext  (v2i1  VCCR:$pred))),
6785            (v2i64 (MVE_VPSEL (MVE_VMOVimmi64 1), (MVE_VMOVimmi32 0), ARMVCCNone, VCCR:$pred, zero_reg))>;
6786}
6787
6788let Predicates = [HasMVEFloat] in {
6789  // Pred <-> Float
6790  // 112 is 1.0 in float
6791  def : Pat<(v4f32 (uint_to_fp (v4i1 VCCR:$pred))),
6792            (v4f32 (MVE_VPSEL (v4f32 (MVE_VMOVimmf32 112)), (v4f32 (MVE_VMOVimmi32 0)), ARMVCCNone, VCCR:$pred, zero_reg))>;
6793  // 2620 in 1.0 in half
6794  def : Pat<(v8f16 (uint_to_fp (v8i1 VCCR:$pred))),
6795            (v8f16 (MVE_VPSEL (v8f16 (MVE_VMOVimmi16 2620)), (v8f16 (MVE_VMOVimmi16 0)), ARMVCCNone, VCCR:$pred, zero_reg))>;
6796  // 240 is -1.0 in float
6797  def : Pat<(v4f32 (sint_to_fp (v4i1 VCCR:$pred))),
6798            (v4f32 (MVE_VPSEL (v4f32 (MVE_VMOVimmf32 240)), (v4f32 (MVE_VMOVimmi32 0)), ARMVCCNone, VCCR:$pred, zero_reg))>;
6799  // 2748 is -1.0 in half
6800  def : Pat<(v8f16 (sint_to_fp (v8i1 VCCR:$pred))),
6801            (v8f16 (MVE_VPSEL (v8f16 (MVE_VMOVimmi16 2748)), (v8f16 (MVE_VMOVimmi16 0)), ARMVCCNone, VCCR:$pred, zero_reg))>;
6802
6803  def : Pat<(v4i1 (fp_to_uint (v4f32 MQPR:$v1))),
6804            (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), ZR, ARMCCne))>;
6805  def : Pat<(v8i1 (fp_to_uint (v8f16 MQPR:$v1))),
6806            (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), ZR, ARMCCne))>;
6807  def : Pat<(v4i1 (fp_to_sint (v4f32 MQPR:$v1))),
6808            (v4i1 (MVE_VCMPf32r (v4f32 MQPR:$v1), ZR, ARMCCne))>;
6809  def : Pat<(v8i1 (fp_to_sint (v8f16 MQPR:$v1))),
6810            (v8i1 (MVE_VCMPf16r (v8f16 MQPR:$v1), ZR, ARMCCne))>;
6811}
6812
6813def MVE_VPNOT : MVE_p<(outs VCCR:$P0), (ins VCCR:$P0_in), NoItinerary,
6814                      "vpnot", "", "", vpred_n, "", 0b00, []> {
6815  let Inst{31-0} = 0b11111110001100010000111101001101;
6816  let Unpredictable{19-17} = 0b111;
6817  let Unpredictable{12} = 0b1;
6818  let Unpredictable{7} = 0b1;
6819  let Unpredictable{5} = 0b1;
6820
6821  let Constraints = "";
6822  let DecoderMethod = "DecodeMVEVPNOT";
6823}
6824
6825let Predicates = [HasMVEInt] in {
6826  def : Pat<(v2i1 (xor (v2i1 VCCR:$pred), (v2i1 (predicate_cast (i32 65535))))),
6827            (v2i1 (MVE_VPNOT (v2i1 VCCR:$pred)))>;
6828  def : Pat<(v4i1 (xor (v4i1 VCCR:$pred), (v4i1 (predicate_cast (i32 65535))))),
6829            (v4i1 (MVE_VPNOT (v4i1 VCCR:$pred)))>;
6830  def : Pat<(v8i1 (xor (v8i1 VCCR:$pred), (v8i1 (predicate_cast (i32 65535))))),
6831            (v8i1 (MVE_VPNOT (v8i1 VCCR:$pred)))>;
6832  def : Pat<(v16i1 (xor (v16i1 VCCR:$pred), (v16i1 (predicate_cast (i32 65535))))),
6833            (v16i1 (MVE_VPNOT (v16i1 VCCR:$pred)))>;
6834}
6835
6836
6837class MVE_loltp_start<dag iops, string asm, string ops, bits<2> size>
6838  : t2LOL<(outs GPRlr:$LR), iops, asm, ops> {
6839  bits<4> Rn;
6840  let Predicates = [HasMVEInt];
6841  let Inst{22} = 0b0;
6842  let Inst{21-20} = size;
6843  let Inst{19-16} = Rn{3-0};
6844  let Inst{12} = 0b0;
6845}
6846
6847class MVE_DLSTP<string asm, bits<2> size>
6848  : MVE_loltp_start<(ins rGPR:$Rn), asm, "$LR, $Rn", size> {
6849  let Inst{13} = 0b1;
6850  let Inst{11-1} = 0b00000000000;
6851  let Unpredictable{10-1} = 0b1111111111;
6852}
6853
6854class MVE_WLSTP<string asm, bits<2> size>
6855  : MVE_loltp_start<(ins rGPR:$Rn, wlslabel_u11:$label),
6856                    asm, "$LR, $Rn, $label", size> {
6857  bits<11> label;
6858  let Inst{13} = 0b0;
6859  let Inst{11} = label{0};
6860  let Inst{10-1} = label{10-1};
6861  let isBranch = 1;
6862  let isTerminator = 1;
6863}
6864
6865def SDT_MVEMEMCPYLOOPNODE
6866    : SDTypeProfile<0, 3, [SDTCisPtrTy<0>, SDTCisPtrTy<1>, SDTCisVT<2, i32>]>;
6867def MVE_MEMCPYLOOPNODE : SDNode<"ARMISD::MEMCPYLOOP", SDT_MVEMEMCPYLOOPNODE,
6868                                [SDNPHasChain, SDNPMayStore, SDNPMayLoad]>;
6869
6870let usesCustomInserter = 1, hasNoSchedulingInfo = 1, Defs = [CPSR] in {
6871  def MVE_MEMCPYLOOPINST : PseudoInst<(outs),
6872        (ins rGPR:$dst, rGPR:$src, rGPR:$sz),
6873        NoItinerary,
6874        [(MVE_MEMCPYLOOPNODE rGPR:$dst, rGPR:$src, rGPR:$sz)]>;
6875}
6876
6877def SDT_MVEMEMSETLOOPNODE
6878    : SDTypeProfile<0, 3, [SDTCisPtrTy<0>, SDTCisVT<1, v16i8>, SDTCisVT<2, i32>]>;
6879def MVE_MEMSETLOOPNODE : SDNode<"ARMISD::MEMSETLOOP", SDT_MVEMEMSETLOOPNODE,
6880                                [SDNPHasChain, SDNPMayStore, SDNPMayLoad]>;
6881
6882let usesCustomInserter = 1, hasNoSchedulingInfo = 1, Defs = [CPSR] in {
6883  def MVE_MEMSETLOOPINST : PseudoInst<(outs),
6884        (ins rGPR:$dst, MQPR:$src, rGPR:$sz),
6885        NoItinerary,
6886        [(MVE_MEMSETLOOPNODE rGPR:$dst, MQPR:$src, rGPR:$sz)]>;
6887}
6888
6889def MVE_DLSTP_8  : MVE_DLSTP<"dlstp.8",  0b00>;
6890def MVE_DLSTP_16 : MVE_DLSTP<"dlstp.16", 0b01>;
6891def MVE_DLSTP_32 : MVE_DLSTP<"dlstp.32", 0b10>;
6892def MVE_DLSTP_64 : MVE_DLSTP<"dlstp.64", 0b11>;
6893
6894def MVE_WLSTP_8  : MVE_WLSTP<"wlstp.8",  0b00>;
6895def MVE_WLSTP_16 : MVE_WLSTP<"wlstp.16", 0b01>;
6896def MVE_WLSTP_32 : MVE_WLSTP<"wlstp.32", 0b10>;
6897def MVE_WLSTP_64 : MVE_WLSTP<"wlstp.64", 0b11>;
6898
6899class MVE_loltp_end<dag oops, dag iops, string asm, string ops>
6900  : t2LOL<oops, iops, asm, ops> {
6901  let Predicates = [HasMVEInt];
6902  let Inst{22-21} = 0b00;
6903  let Inst{19-16} = 0b1111;
6904  let Inst{12} = 0b0;
6905}
6906
6907def MVE_LETP : MVE_loltp_end<(outs GPRlr:$LRout),
6908                             (ins GPRlr:$LRin, lelabel_u11:$label),
6909                             "letp", "$LRin, $label"> {
6910  bits<11> label;
6911  let Inst{20} = 0b1;
6912  let Inst{13} = 0b0;
6913  let Inst{11} = label{0};
6914  let Inst{10-1} = label{10-1};
6915  let isBranch = 1;
6916  let isTerminator = 1;
6917}
6918
6919def MVE_LCTP : MVE_loltp_end<(outs), (ins pred:$p), "lctp${p}", ""> {
6920  let Inst{20} = 0b0;
6921  let Inst{13} = 0b1;
6922  let Inst{11-1} = 0b00000000000;
6923  let Unpredictable{21-20} = 0b11;
6924  let Unpredictable{11-1} = 0b11111111111;
6925}
6926
6927
6928// Pseudo instructions for lowering MQQPR and MQQQQPR stack spills and reloads.
6929// They are equivalent to VLDMDIA/VSTMDIA with a single reg, as opposed to multiple
6930// dreg subregs.
6931
6932let Predicates = [HasMVEInt], AM = AddrMode4 in {
6933let mayStore = 1, hasSideEffects = 0 in {
6934  def MQQPRStore : t2PseudoInst<(outs), (ins MQQPR:$val, GPRnopc:$ptr),
6935                                4, NoItinerary, []>;
6936  def MQQQQPRStore : t2PseudoInst<(outs), (ins MQQQQPR:$val, GPRnopc:$ptr),
6937                                  4, NoItinerary, []>;
6938}
6939let mayLoad = 1, hasSideEffects = 0 in {
6940  def MQQPRLoad : t2PseudoInst<(outs MQQPR:$val), (ins GPRnopc:$ptr),
6941                               4, NoItinerary, []>;
6942  def MQQQQPRLoad : t2PseudoInst<(outs MQQQQPR:$val), (ins GPRnopc:$ptr),
6943                                 4, NoItinerary, []>;
6944}
6945}
6946
6947// Pseudo for lowering MVE Q register COPYs. These will usually get converted
6948// to a "MVE_VORR dst, src, src", but may behave differently in tail predicated
6949// loops to ensure the whole register is copied, not a subset from a
6950// tail-predicated MVE_VORR. In the event we cannot prove a MVE_VORR is valid,
6951// it will become a pair of VMOVD instructions for each half of the Q register.
6952let Predicates = [HasMVEInt], hasSideEffects = 0, isMoveReg = 1,
6953    D = MVEDomain in {
6954  def MQPRCopy : t2PseudoInst<(outs MQPR:$dst), (ins MQPR:$src),
6955                              8, NoItinerary, []>;
6956}
6957
6958
6959//===----------------------------------------------------------------------===//
6960// Patterns
6961//===----------------------------------------------------------------------===//
6962
6963// PatFrags for loads and stores. Often trying to keep semi-consistent names.
6964
6965def aligned32_pre_store : PatFrag<(ops node:$val, node:$ptr, node:$offset),
6966                                  (pre_store node:$val, node:$ptr, node:$offset), [{
6967  return cast<StoreSDNode>(N)->getAlign() >= 4;
6968}]>;
6969def aligned32_post_store : PatFrag<(ops node:$val, node:$ptr, node:$offset),
6970                                   (post_store node:$val, node:$ptr, node:$offset), [{
6971  return cast<StoreSDNode>(N)->getAlign() >= 4;
6972}]>;
6973def aligned16_pre_store : PatFrag<(ops node:$val, node:$ptr, node:$offset),
6974                                  (pre_store node:$val, node:$ptr, node:$offset), [{
6975  return cast<StoreSDNode>(N)->getAlign() >= 2;
6976}]>;
6977def aligned16_post_store : PatFrag<(ops node:$val, node:$ptr, node:$offset),
6978                                   (post_store node:$val, node:$ptr, node:$offset), [{
6979  return cast<StoreSDNode>(N)->getAlign() >= 2;
6980}]>;
6981
6982
6983def aligned_maskedloadvi8 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
6984                                    (masked_ld node:$ptr, undef, node:$pred, node:$passthru), [{
6985  auto *Ld = cast<MaskedLoadSDNode>(N);
6986  return Ld->getMemoryVT().getScalarType() == MVT::i8;
6987}]>;
6988def aligned_sextmaskedloadvi8 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
6989                                        (aligned_maskedloadvi8 node:$ptr, node:$pred, node:$passthru), [{
6990  return cast<MaskedLoadSDNode>(N)->getExtensionType() == ISD::SEXTLOAD;
6991}]>;
6992def aligned_zextmaskedloadvi8 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
6993                                        (aligned_maskedloadvi8 node:$ptr, node:$pred, node:$passthru), [{
6994  return cast<MaskedLoadSDNode>(N)->getExtensionType() == ISD::ZEXTLOAD;
6995}]>;
6996def aligned_extmaskedloadvi8 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
6997                                       (aligned_maskedloadvi8 node:$ptr, node:$pred, node:$passthru), [{
6998  auto *Ld = cast<MaskedLoadSDNode>(N);
6999  EVT ScalarVT = Ld->getMemoryVT().getScalarType();
7000  return ScalarVT.isInteger() && Ld->getExtensionType() == ISD::EXTLOAD;
7001}]>;
7002def aligned_maskedloadvi16: PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
7003                                    (masked_ld node:$ptr, undef, node:$pred, node:$passthru), [{
7004  auto *Ld = cast<MaskedLoadSDNode>(N);
7005  EVT ScalarVT = Ld->getMemoryVT().getScalarType();
7006  return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && Ld->getAlign() >= 2;
7007}]>;
7008def aligned_sextmaskedloadvi16 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
7009                                         (aligned_maskedloadvi16 node:$ptr, node:$pred, node:$passthru), [{
7010  return cast<MaskedLoadSDNode>(N)->getExtensionType() == ISD::SEXTLOAD;
7011}]>;
7012def aligned_zextmaskedloadvi16 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
7013                                         (aligned_maskedloadvi16 node:$ptr, node:$pred, node:$passthru), [{
7014  return cast<MaskedLoadSDNode>(N)->getExtensionType() == ISD::ZEXTLOAD;
7015}]>;
7016def aligned_extmaskedloadvi16 : PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
7017                                        (aligned_maskedloadvi16 node:$ptr, node:$pred, node:$passthru), [{
7018  auto *Ld = cast<MaskedLoadSDNode>(N);
7019  EVT ScalarVT = Ld->getMemoryVT().getScalarType();
7020  return ScalarVT.isInteger() && Ld->getExtensionType() == ISD::EXTLOAD;
7021}]>;
7022def aligned_maskedloadvi32: PatFrag<(ops node:$ptr, node:$pred, node:$passthru),
7023                                    (masked_ld node:$ptr, undef, node:$pred, node:$passthru), [{
7024  auto *Ld = cast<MaskedLoadSDNode>(N);
7025  EVT ScalarVT = Ld->getMemoryVT().getScalarType();
7026  return (ScalarVT == MVT::i32 || ScalarVT == MVT::f32) && Ld->getAlign() >= 4;
7027}]>;
7028
7029def aligned_maskedstvi8 : PatFrag<(ops node:$val, node:$ptr, node:$pred),
7030                                  (masked_st node:$val, node:$ptr, undef, node:$pred), [{
7031  return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
7032}]>;
7033def aligned_maskedstvi16 : PatFrag<(ops node:$val, node:$ptr, node:$pred),
7034                                   (masked_st node:$val, node:$ptr, undef, node:$pred), [{
7035  auto *St = cast<MaskedStoreSDNode>(N);
7036  EVT ScalarVT = St->getMemoryVT().getScalarType();
7037  return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlign() >= 2;
7038}]>;
7039def aligned_maskedstvi32 : PatFrag<(ops node:$val, node:$ptr, node:$pred),
7040                                   (masked_st node:$val, node:$ptr, undef, node:$pred), [{
7041  auto *St = cast<MaskedStoreSDNode>(N);
7042  EVT ScalarVT = St->getMemoryVT().getScalarType();
7043  return (ScalarVT == MVT::i32 || ScalarVT == MVT::f32) && St->getAlign() >= 4;
7044}]>;
7045
7046def pre_maskedstore : PatFrag<(ops node:$val, node:$base, node:$offset, node:$mask),
7047                              (masked_st node:$val, node:$base, node:$offset, node:$mask), [{
7048  ISD::MemIndexedMode AM = cast<MaskedStoreSDNode>(N)->getAddressingMode();
7049  return AM == ISD::PRE_INC || AM == ISD::PRE_DEC;
7050}]>;
7051def post_maskedstore : PatFrag<(ops node:$val, node:$base, node:$offset, node:$mask),
7052                               (masked_st node:$val, node:$base, node:$offset, node:$mask), [{
7053  ISD::MemIndexedMode AM = cast<MaskedStoreSDNode>(N)->getAddressingMode();
7054  return AM == ISD::POST_INC || AM == ISD::POST_DEC;
7055}]>;
7056def aligned_pre_maskedstorevi8 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
7057                                         (pre_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
7058  return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
7059}]>;
7060def aligned_post_maskedstorevi8 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
7061                                          (post_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
7062  return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
7063}]>;
7064def aligned_pre_maskedstorevi16 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
7065                                          (pre_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
7066  auto *St = cast<MaskedStoreSDNode>(N);
7067  EVT ScalarVT = St->getMemoryVT().getScalarType();
7068  return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlign() >= 2;
7069}]>;
7070def aligned_post_maskedstorevi16 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
7071                                           (post_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
7072  auto *St = cast<MaskedStoreSDNode>(N);
7073  EVT ScalarVT = St->getMemoryVT().getScalarType();
7074  return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlign() >= 2;
7075}]>;
7076def aligned_pre_maskedstorevi32 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
7077                                          (pre_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
7078  auto *St = cast<MaskedStoreSDNode>(N);
7079  EVT ScalarVT = St->getMemoryVT().getScalarType();
7080  return (ScalarVT == MVT::i32 || ScalarVT == MVT::f32) && St->getAlign() >= 4;
7081}]>;
7082def aligned_post_maskedstorevi32 : PatFrag<(ops node:$val, node:$ptr, node:$offset, node:$mask),
7083                                           (post_maskedstore node:$val, node:$ptr, node:$offset, node:$mask), [{
7084  auto *St = cast<MaskedStoreSDNode>(N);
7085  EVT ScalarVT = St->getMemoryVT().getScalarType();
7086  return (ScalarVT == MVT::i32 || ScalarVT == MVT::f32) && St->getAlign() >= 4;
7087}]>;
7088
7089
7090// PatFrags for "Aligned" extending / truncating
7091
7092def aligned_extloadvi8  : PatFrag<(ops node:$ptr), (extloadvi8 node:$ptr)>;
7093def aligned_sextloadvi8 : PatFrag<(ops node:$ptr), (sextloadvi8 node:$ptr)>;
7094def aligned_zextloadvi8 : PatFrag<(ops node:$ptr), (zextloadvi8 node:$ptr)>;
7095
7096def aligned_truncstvi8 : PatFrag<(ops node:$val, node:$ptr),
7097                                 (truncstorevi8 node:$val, node:$ptr)>;
7098def aligned_post_truncstvi8 : PatFrag<(ops node:$val, node:$base, node:$offset),
7099                                      (post_truncstvi8 node:$val, node:$base, node:$offset)>;
7100def aligned_pre_truncstvi8 : PatFrag<(ops node:$val, node:$base, node:$offset),
7101                                     (pre_truncstvi8 node:$val, node:$base, node:$offset)>;
7102
7103let MinAlignment = 2 in {
7104  def aligned_extloadvi16  : PatFrag<(ops node:$ptr), (extloadvi16 node:$ptr)>;
7105  def aligned_sextloadvi16 : PatFrag<(ops node:$ptr), (sextloadvi16 node:$ptr)>;
7106  def aligned_zextloadvi16 : PatFrag<(ops node:$ptr), (zextloadvi16 node:$ptr)>;
7107
7108  def aligned_truncstvi16 : PatFrag<(ops node:$val, node:$ptr),
7109                                    (truncstorevi16 node:$val, node:$ptr)>;
7110  def aligned_post_truncstvi16 : PatFrag<(ops node:$val, node:$base, node:$offset),
7111                                         (post_truncstvi16 node:$val, node:$base, node:$offset)>;
7112  def aligned_pre_truncstvi16 : PatFrag<(ops node:$val, node:$base, node:$offset),
7113                                        (pre_truncstvi16 node:$val, node:$base, node:$offset)>;
7114}
7115
7116def truncmaskedst : PatFrag<(ops node:$val, node:$base, node:$pred),
7117                            (masked_st node:$val, node:$base, undef, node:$pred), [{
7118  return cast<MaskedStoreSDNode>(N)->isTruncatingStore();
7119}]>;
7120def aligned_truncmaskedstvi8 : PatFrag<(ops node:$val, node:$base, node:$pred),
7121                                       (truncmaskedst node:$val, node:$base, node:$pred), [{
7122  return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
7123}]>;
7124def aligned_truncmaskedstvi16 : PatFrag<(ops node:$val, node:$base, node:$pred),
7125                                        (truncmaskedst node:$val, node:$base, node:$pred), [{
7126  auto *St = cast<MaskedStoreSDNode>(N);
7127  EVT ScalarVT = St->getMemoryVT().getScalarType();
7128  return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlign() >= 2;
7129}]>;
7130def pre_truncmaskedst : PatFrag<(ops node:$val, node:$base, node:$offset, node:$pred),
7131                                (masked_st node:$val, node:$base, node:$offset, node:$pred), [{
7132  ISD::MemIndexedMode AM = cast<MaskedStoreSDNode>(N)->getAddressingMode();
7133  return cast<MaskedStoreSDNode>(N)->isTruncatingStore() && (AM == ISD::PRE_INC || AM == ISD::PRE_DEC);
7134}]>;
7135def aligned_pre_truncmaskedstvi8 : PatFrag<(ops node:$val, node:$base, node:$offset, node:$pred),
7136                                           (pre_truncmaskedst node:$val, node:$base, node:$offset, node:$pred), [{
7137  return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
7138}]>;
7139def aligned_pre_truncmaskedstvi16 : PatFrag<(ops node:$val, node:$base, node:$offset, node:$pred),
7140                                            (pre_truncmaskedst node:$val, node:$base, node:$offset, node:$pred), [{
7141  auto *St = cast<MaskedStoreSDNode>(N);
7142  EVT ScalarVT = St->getMemoryVT().getScalarType();
7143  return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlign() >= 2;
7144}]>;
7145def post_truncmaskedst : PatFrag<(ops node:$val, node:$base, node:$offset, node:$postd),
7146                                 (masked_st node:$val, node:$base, node:$offset, node:$postd), [{
7147  ISD::MemIndexedMode AM = cast<MaskedStoreSDNode>(N)->getAddressingMode();
7148  return cast<MaskedStoreSDNode>(N)->isTruncatingStore() && (AM == ISD::POST_INC || AM == ISD::POST_DEC);
7149}]>;
7150def aligned_post_truncmaskedstvi8 : PatFrag<(ops node:$val, node:$base, node:$offset, node:$postd),
7151                                            (post_truncmaskedst node:$val, node:$base, node:$offset, node:$postd), [{
7152  return cast<MaskedStoreSDNode>(N)->getMemoryVT().getScalarType() == MVT::i8;
7153}]>;
7154def aligned_post_truncmaskedstvi16 : PatFrag<(ops node:$val, node:$base, node:$offset, node:$postd),
7155                                             (post_truncmaskedst node:$val, node:$base, node:$offset, node:$postd), [{
7156  auto *St = cast<MaskedStoreSDNode>(N);
7157  EVT ScalarVT = St->getMemoryVT().getScalarType();
7158  return (ScalarVT == MVT::i16 || ScalarVT == MVT::f16) && St->getAlign() >= 2;
7159}]>;
7160
7161// Load/store patterns
7162
7163class MVE_vector_store_typed<ValueType Ty, Instruction RegImmInst,
7164                             PatFrag StoreKind, int shift>
7165  : Pat<(StoreKind (Ty MQPR:$val), t2addrmode_imm7<shift>:$addr),
7166        (RegImmInst (Ty MQPR:$val), t2addrmode_imm7<shift>:$addr)>;
7167
7168class MVE_vector_maskedstore_typed<ValueType Ty, Instruction RegImmInst,
7169                                   PatFrag StoreKind, int shift>
7170  : Pat<(StoreKind (Ty MQPR:$val), t2addrmode_imm7<shift>:$addr, VCCR:$pred),
7171        (RegImmInst (Ty MQPR:$val), t2addrmode_imm7<shift>:$addr, ARMVCCThen, VCCR:$pred, zero_reg)>;
7172
7173multiclass MVE_vector_store<Instruction RegImmInst, PatFrag StoreKind,
7174                            int shift> {
7175  def : MVE_vector_store_typed<v16i8, RegImmInst, StoreKind, shift>;
7176  def : MVE_vector_store_typed<v8i16, RegImmInst, StoreKind, shift>;
7177  def : MVE_vector_store_typed<v8f16, RegImmInst, StoreKind, shift>;
7178  def : MVE_vector_store_typed<v4i32, RegImmInst, StoreKind, shift>;
7179  def : MVE_vector_store_typed<v4f32, RegImmInst, StoreKind, shift>;
7180  def : MVE_vector_store_typed<v2i64, RegImmInst, StoreKind, shift>;
7181  def : MVE_vector_store_typed<v2f64, RegImmInst, StoreKind, shift>;
7182}
7183
7184class MVE_vector_load_typed<ValueType Ty, Instruction RegImmInst,
7185                            PatFrag LoadKind, int shift>
7186  : Pat<(Ty (LoadKind t2addrmode_imm7<shift>:$addr)),
7187        (Ty (RegImmInst t2addrmode_imm7<shift>:$addr))>;
7188
7189class MVE_vector_maskedload_typed<ValueType Ty, Instruction RegImmInst,
7190                                  PatFrag LoadKind, int shift>
7191  : Pat<(Ty (LoadKind t2addrmode_imm7<shift>:$addr, VCCR:$pred, (Ty (ARMvmovImm (i32 0))))),
7192        (Ty (RegImmInst t2addrmode_imm7<shift>:$addr, ARMVCCThen, VCCR:$pred, zero_reg))>;
7193
7194multiclass MVE_vector_load<Instruction RegImmInst, PatFrag LoadKind,
7195                           int shift> {
7196  def : MVE_vector_load_typed<v16i8, RegImmInst, LoadKind, shift>;
7197  def : MVE_vector_load_typed<v8i16, RegImmInst, LoadKind, shift>;
7198  def : MVE_vector_load_typed<v8f16, RegImmInst, LoadKind, shift>;
7199  def : MVE_vector_load_typed<v4i32, RegImmInst, LoadKind, shift>;
7200  def : MVE_vector_load_typed<v4f32, RegImmInst, LoadKind, shift>;
7201  def : MVE_vector_load_typed<v2i64, RegImmInst, LoadKind, shift>;
7202  def : MVE_vector_load_typed<v2f64, RegImmInst, LoadKind, shift>;
7203}
7204
7205class MVE_vector_offset_store_typed<ValueType Ty, Instruction Opcode,
7206                                    PatFrag StoreKind, int shift>
7207  : Pat<(StoreKind (Ty MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<shift>:$addr),
7208        (Opcode MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<shift>:$addr)>;
7209
7210class MVE_vector_offset_maskedstore_typed<ValueType Ty, Instruction Opcode,
7211                                          PatFrag StoreKind, int shift>
7212  : Pat<(StoreKind (Ty MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<shift>:$addr, VCCR:$pred),
7213        (Opcode MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<shift>:$addr, ARMVCCThen, VCCR:$pred, zero_reg)>;
7214
7215multiclass MVE_vector_offset_store<Instruction RegImmInst, PatFrag StoreKind,
7216                                   int shift> {
7217  def : MVE_vector_offset_store_typed<v16i8, RegImmInst, StoreKind, shift>;
7218  def : MVE_vector_offset_store_typed<v8i16, RegImmInst, StoreKind, shift>;
7219  def : MVE_vector_offset_store_typed<v8f16, RegImmInst, StoreKind, shift>;
7220  def : MVE_vector_offset_store_typed<v4i32, RegImmInst, StoreKind, shift>;
7221  def : MVE_vector_offset_store_typed<v4f32, RegImmInst, StoreKind, shift>;
7222  def : MVE_vector_offset_store_typed<v2i64, RegImmInst, StoreKind, shift>;
7223  def : MVE_vector_offset_store_typed<v2f64, RegImmInst, StoreKind, shift>;
7224}
7225
7226
7227let Predicates = [HasMVEInt, IsLE] in {
7228  // Stores
7229  defm : MVE_vector_store<MVE_VSTRBU8, byte_alignedstore, 0>;
7230  defm : MVE_vector_store<MVE_VSTRHU16, hword_alignedstore, 1>;
7231  defm : MVE_vector_store<MVE_VSTRWU32, alignedstore32, 2>;
7232
7233  // Loads
7234  defm : MVE_vector_load<MVE_VLDRBU8, byte_alignedload, 0>;
7235  defm : MVE_vector_load<MVE_VLDRHU16, hword_alignedload, 1>;
7236  defm : MVE_vector_load<MVE_VLDRWU32, alignedload32, 2>;
7237
7238  // Pre/post inc stores
7239  defm : MVE_vector_offset_store<MVE_VSTRBU8_pre, pre_store, 0>;
7240  defm : MVE_vector_offset_store<MVE_VSTRBU8_post, post_store, 0>;
7241  defm : MVE_vector_offset_store<MVE_VSTRHU16_pre, aligned16_pre_store, 1>;
7242  defm : MVE_vector_offset_store<MVE_VSTRHU16_post, aligned16_post_store, 1>;
7243  defm : MVE_vector_offset_store<MVE_VSTRWU32_pre, aligned32_pre_store, 2>;
7244  defm : MVE_vector_offset_store<MVE_VSTRWU32_post, aligned32_post_store, 2>;
7245}
7246
7247let Predicates = [HasMVEInt, IsBE] in {
7248  // Aligned Stores
7249  def : MVE_vector_store_typed<v16i8, MVE_VSTRBU8, store, 0>;
7250  def : MVE_vector_store_typed<v8i16, MVE_VSTRHU16, alignedstore16, 1>;
7251  def : MVE_vector_store_typed<v8f16, MVE_VSTRHU16, alignedstore16, 1>;
7252  def : MVE_vector_store_typed<v4i32, MVE_VSTRWU32, alignedstore32, 2>;
7253  def : MVE_vector_store_typed<v4f32, MVE_VSTRWU32, alignedstore32, 2>;
7254
7255  // Aligned Loads
7256  def : MVE_vector_load_typed<v16i8, MVE_VLDRBU8, load, 0>;
7257  def : MVE_vector_load_typed<v8i16, MVE_VLDRHU16, alignedload16, 1>;
7258  def : MVE_vector_load_typed<v8f16, MVE_VLDRHU16, alignedload16, 1>;
7259  def : MVE_vector_load_typed<v4i32, MVE_VLDRWU32, alignedload32, 2>;
7260  def : MVE_vector_load_typed<v4f32, MVE_VLDRWU32, alignedload32, 2>;
7261
7262  // Other unaligned loads/stores need to go though a VREV
7263  def : Pat<(v2f64 (load t2addrmode_imm7<0>:$addr)),
7264            (v2f64 (MVE_VREV64_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
7265  def : Pat<(v2i64 (load t2addrmode_imm7<0>:$addr)),
7266            (v2i64 (MVE_VREV64_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
7267  def : Pat<(v4i32 (load t2addrmode_imm7<0>:$addr)),
7268            (v4i32 (MVE_VREV32_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
7269  def : Pat<(v4f32 (load t2addrmode_imm7<0>:$addr)),
7270            (v4f32 (MVE_VREV32_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
7271  def : Pat<(v8i16 (load t2addrmode_imm7<0>:$addr)),
7272            (v8i16 (MVE_VREV16_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
7273  def : Pat<(v8f16 (load t2addrmode_imm7<0>:$addr)),
7274            (v8f16 (MVE_VREV16_8 (MVE_VLDRBU8 t2addrmode_imm7<0>:$addr)))>;
7275  def : Pat<(store (v2f64 MQPR:$val), t2addrmode_imm7<0>:$addr),
7276            (MVE_VSTRBU8 (MVE_VREV64_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
7277  def : Pat<(store (v2i64 MQPR:$val), t2addrmode_imm7<0>:$addr),
7278            (MVE_VSTRBU8 (MVE_VREV64_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
7279  def : Pat<(store (v4i32 MQPR:$val), t2addrmode_imm7<0>:$addr),
7280            (MVE_VSTRBU8 (MVE_VREV32_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
7281  def : Pat<(store (v4f32 MQPR:$val), t2addrmode_imm7<0>:$addr),
7282            (MVE_VSTRBU8 (MVE_VREV32_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
7283  def : Pat<(store (v8i16 MQPR:$val), t2addrmode_imm7<0>:$addr),
7284            (MVE_VSTRBU8 (MVE_VREV16_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
7285  def : Pat<(store (v8f16 MQPR:$val), t2addrmode_imm7<0>:$addr),
7286            (MVE_VSTRBU8 (MVE_VREV16_8 MQPR:$val), t2addrmode_imm7<0>:$addr)>;
7287
7288  // Pre/Post inc stores
7289  def : MVE_vector_offset_store_typed<v16i8, MVE_VSTRBU8_pre, pre_store, 0>;
7290  def : MVE_vector_offset_store_typed<v16i8, MVE_VSTRBU8_post, post_store, 0>;
7291  def : MVE_vector_offset_store_typed<v8i16, MVE_VSTRHU16_pre, aligned16_pre_store, 1>;
7292  def : MVE_vector_offset_store_typed<v8i16, MVE_VSTRHU16_post, aligned16_post_store, 1>;
7293  def : MVE_vector_offset_store_typed<v8f16, MVE_VSTRHU16_pre, aligned16_pre_store, 1>;
7294  def : MVE_vector_offset_store_typed<v8f16, MVE_VSTRHU16_post, aligned16_post_store, 1>;
7295  def : MVE_vector_offset_store_typed<v4i32, MVE_VSTRWU32_pre, aligned32_pre_store, 2>;
7296  def : MVE_vector_offset_store_typed<v4i32, MVE_VSTRWU32_post, aligned32_post_store, 2>;
7297  def : MVE_vector_offset_store_typed<v4f32, MVE_VSTRWU32_pre, aligned32_pre_store, 2>;
7298  def : MVE_vector_offset_store_typed<v4f32, MVE_VSTRWU32_post, aligned32_post_store, 2>;
7299}
7300
7301let Predicates = [HasMVEInt] in {
7302  // Aligned masked store, shared between LE and BE
7303  def : MVE_vector_maskedstore_typed<v16i8, MVE_VSTRBU8, aligned_maskedstvi8, 0>;
7304  def : MVE_vector_maskedstore_typed<v8i16, MVE_VSTRHU16, aligned_maskedstvi16, 1>;
7305  def : MVE_vector_maskedstore_typed<v8f16, MVE_VSTRHU16, aligned_maskedstvi16, 1>;
7306  def : MVE_vector_maskedstore_typed<v4i32, MVE_VSTRWU32, aligned_maskedstvi32, 2>;
7307  def : MVE_vector_maskedstore_typed<v4f32, MVE_VSTRWU32, aligned_maskedstvi32, 2>;
7308
7309  // Pre/Post inc masked stores
7310  def : MVE_vector_offset_maskedstore_typed<v16i8, MVE_VSTRBU8_pre, aligned_pre_maskedstorevi8, 0>;
7311  def : MVE_vector_offset_maskedstore_typed<v16i8, MVE_VSTRBU8_post, aligned_post_maskedstorevi8, 0>;
7312  def : MVE_vector_offset_maskedstore_typed<v8i16, MVE_VSTRHU16_pre, aligned_pre_maskedstorevi16, 1>;
7313  def : MVE_vector_offset_maskedstore_typed<v8i16, MVE_VSTRHU16_post, aligned_post_maskedstorevi16, 1>;
7314  def : MVE_vector_offset_maskedstore_typed<v8f16, MVE_VSTRHU16_pre, aligned_pre_maskedstorevi16, 1>;
7315  def : MVE_vector_offset_maskedstore_typed<v8f16, MVE_VSTRHU16_post, aligned_post_maskedstorevi16, 1>;
7316  def : MVE_vector_offset_maskedstore_typed<v4i32, MVE_VSTRWU32_pre, aligned_pre_maskedstorevi32, 2>;
7317  def : MVE_vector_offset_maskedstore_typed<v4i32, MVE_VSTRWU32_post, aligned_post_maskedstorevi32, 2>;
7318  def : MVE_vector_offset_maskedstore_typed<v4f32, MVE_VSTRWU32_pre, aligned_pre_maskedstorevi32, 2>;
7319  def : MVE_vector_offset_maskedstore_typed<v4f32, MVE_VSTRWU32_post, aligned_post_maskedstorevi32, 2>;
7320
7321  // Aligned masked loads
7322  def : MVE_vector_maskedload_typed<v16i8, MVE_VLDRBU8, aligned_maskedloadvi8, 0>;
7323  def : MVE_vector_maskedload_typed<v8i16, MVE_VLDRHU16, aligned_maskedloadvi16, 1>;
7324  def : MVE_vector_maskedload_typed<v8f16, MVE_VLDRHU16, aligned_maskedloadvi16, 1>;
7325  def : MVE_vector_maskedload_typed<v4i32, MVE_VLDRWU32, aligned_maskedloadvi32, 2>;
7326  def : MVE_vector_maskedload_typed<v4f32, MVE_VLDRWU32, aligned_maskedloadvi32, 2>;
7327}
7328
7329// Widening/Narrowing Loads/Stores
7330
7331multiclass MVEExtLoadStore<Instruction LoadSInst, Instruction LoadUInst, string StoreInst,
7332                         string Amble, ValueType VT, int Shift> {
7333  // Trunc stores
7334  def : Pat<(!cast<PatFrag>("aligned_truncst"#Amble) (VT MQPR:$val), taddrmode_imm7<Shift>:$addr),
7335            (!cast<Instruction>(StoreInst) MQPR:$val, taddrmode_imm7<Shift>:$addr)>;
7336  def : Pat<(!cast<PatFrag>("aligned_post_truncst"#Amble) (VT MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<Shift>:$addr),
7337            (!cast<Instruction>(StoreInst#"_post") MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<Shift>:$addr)>;
7338  def : Pat<(!cast<PatFrag>("aligned_pre_truncst"#Amble) (VT MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<Shift>:$addr),
7339            (!cast<Instruction>(StoreInst#"_pre") MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<Shift>:$addr)>;
7340
7341  // Masked trunc stores
7342  def : Pat<(!cast<PatFrag>("aligned_truncmaskedst"#Amble) (VT MQPR:$val), taddrmode_imm7<Shift>:$addr, VCCR:$pred),
7343            (!cast<Instruction>(StoreInst) MQPR:$val, taddrmode_imm7<Shift>:$addr, ARMVCCThen, VCCR:$pred, zero_reg)>;
7344  def : Pat<(!cast<PatFrag>("aligned_post_truncmaskedst"#Amble) (VT MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<Shift>:$addr, VCCR:$pred),
7345            (!cast<Instruction>(StoreInst#"_post") MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<Shift>:$addr, ARMVCCThen, VCCR:$pred, zero_reg)>;
7346  def : Pat<(!cast<PatFrag>("aligned_pre_truncmaskedst"#Amble) (VT MQPR:$Rt), tGPR:$Rn, t2am_imm7_offset<Shift>:$addr, VCCR:$pred),
7347            (!cast<Instruction>(StoreInst#"_pre") MQPR:$Rt, tGPR:$Rn, t2am_imm7_offset<Shift>:$addr, ARMVCCThen, VCCR:$pred, zero_reg)>;
7348
7349  // Ext loads
7350  def : Pat<(VT (!cast<PatFrag>("aligned_extload"#Amble) taddrmode_imm7<Shift>:$addr)),
7351            (VT (LoadUInst taddrmode_imm7<Shift>:$addr))>;
7352  def : Pat<(VT (!cast<PatFrag>("aligned_sextload"#Amble) taddrmode_imm7<Shift>:$addr)),
7353            (VT (LoadSInst taddrmode_imm7<Shift>:$addr))>;
7354  def : Pat<(VT (!cast<PatFrag>("aligned_zextload"#Amble) taddrmode_imm7<Shift>:$addr)),
7355            (VT (LoadUInst taddrmode_imm7<Shift>:$addr))>;
7356
7357  // Masked ext loads
7358  def : Pat<(VT (!cast<PatFrag>("aligned_extmaskedload"#Amble) taddrmode_imm7<Shift>:$addr, VCCR:$pred, (VT (ARMvmovImm (i32 0))))),
7359            (VT (LoadUInst taddrmode_imm7<Shift>:$addr, ARMVCCThen, VCCR:$pred, zero_reg))>;
7360  def : Pat<(VT (!cast<PatFrag>("aligned_sextmaskedload"#Amble) taddrmode_imm7<Shift>:$addr, VCCR:$pred, (VT (ARMvmovImm (i32 0))))),
7361            (VT (LoadSInst taddrmode_imm7<Shift>:$addr, ARMVCCThen, VCCR:$pred, zero_reg))>;
7362  def : Pat<(VT (!cast<PatFrag>("aligned_zextmaskedload"#Amble) taddrmode_imm7<Shift>:$addr, VCCR:$pred, (VT (ARMvmovImm (i32 0))))),
7363            (VT (LoadUInst taddrmode_imm7<Shift>:$addr, ARMVCCThen, VCCR:$pred, zero_reg))>;
7364}
7365
7366let Predicates = [HasMVEInt] in {
7367  defm : MVEExtLoadStore<MVE_VLDRBS16, MVE_VLDRBU16, "MVE_VSTRB16", "vi8", v8i16, 0>;
7368  defm : MVEExtLoadStore<MVE_VLDRBS32, MVE_VLDRBU32, "MVE_VSTRB32", "vi8", v4i32, 0>;
7369  defm : MVEExtLoadStore<MVE_VLDRHS32, MVE_VLDRHU32, "MVE_VSTRH32", "vi16", v4i32, 1>;
7370}
7371
7372
7373// Bit convert patterns
7374
7375let Predicates = [HasMVEInt] in {
7376  def : Pat<(v2f64 (bitconvert (v2i64 MQPR:$src))), (v2f64 MQPR:$src)>;
7377  def : Pat<(v2i64 (bitconvert (v2f64 MQPR:$src))), (v2i64 MQPR:$src)>;
7378
7379  def : Pat<(v4i32 (bitconvert (v4f32 MQPR:$src))), (v4i32 MQPR:$src)>;
7380  def : Pat<(v4f32 (bitconvert (v4i32 MQPR:$src))), (v4f32 MQPR:$src)>;
7381
7382  def : Pat<(v8i16 (bitconvert (v8f16 MQPR:$src))), (v8i16  MQPR:$src)>;
7383  def : Pat<(v8f16 (bitconvert (v8i16 MQPR:$src))), (v8f16  MQPR:$src)>;
7384}
7385
7386let Predicates = [IsLE,HasMVEInt] in {
7387  def : Pat<(v2f64 (bitconvert (v4f32 MQPR:$src))), (v2f64 MQPR:$src)>;
7388  def : Pat<(v2f64 (bitconvert (v4i32 MQPR:$src))), (v2f64 MQPR:$src)>;
7389  def : Pat<(v2f64 (bitconvert (v8f16 MQPR:$src))), (v2f64 MQPR:$src)>;
7390  def : Pat<(v2f64 (bitconvert (v8i16 MQPR:$src))), (v2f64 MQPR:$src)>;
7391  def : Pat<(v2f64 (bitconvert (v16i8 MQPR:$src))), (v2f64 MQPR:$src)>;
7392
7393  def : Pat<(v2i64 (bitconvert (v4f32 MQPR:$src))), (v2i64 MQPR:$src)>;
7394  def : Pat<(v2i64 (bitconvert (v4i32 MQPR:$src))), (v2i64 MQPR:$src)>;
7395  def : Pat<(v2i64 (bitconvert (v8f16 MQPR:$src))), (v2i64 MQPR:$src)>;
7396  def : Pat<(v2i64 (bitconvert (v8i16 MQPR:$src))), (v2i64 MQPR:$src)>;
7397  def : Pat<(v2i64 (bitconvert (v16i8 MQPR:$src))), (v2i64 MQPR:$src)>;
7398
7399  def : Pat<(v4f32 (bitconvert (v2f64 MQPR:$src))), (v4f32 MQPR:$src)>;
7400  def : Pat<(v4f32 (bitconvert (v2i64 MQPR:$src))), (v4f32 MQPR:$src)>;
7401  def : Pat<(v4f32 (bitconvert (v8f16 MQPR:$src))), (v4f32 MQPR:$src)>;
7402  def : Pat<(v4f32 (bitconvert (v8i16 MQPR:$src))), (v4f32 MQPR:$src)>;
7403  def : Pat<(v4f32 (bitconvert (v16i8 MQPR:$src))), (v4f32 MQPR:$src)>;
7404
7405  def : Pat<(v4i32 (bitconvert (v2f64 MQPR:$src))), (v4i32 MQPR:$src)>;
7406  def : Pat<(v4i32 (bitconvert (v2i64 MQPR:$src))), (v4i32 MQPR:$src)>;
7407  def : Pat<(v4i32 (bitconvert (v8f16 MQPR:$src))), (v4i32 MQPR:$src)>;
7408  def : Pat<(v4i32 (bitconvert (v8i16 MQPR:$src))), (v4i32 MQPR:$src)>;
7409  def : Pat<(v4i32 (bitconvert (v16i8 MQPR:$src))), (v4i32 MQPR:$src)>;
7410
7411  def : Pat<(v8f16 (bitconvert (v2f64 MQPR:$src))), (v8f16 MQPR:$src)>;
7412  def : Pat<(v8f16 (bitconvert (v2i64 MQPR:$src))), (v8f16 MQPR:$src)>;
7413  def : Pat<(v8f16 (bitconvert (v4f32 MQPR:$src))), (v8f16 MQPR:$src)>;
7414  def : Pat<(v8f16 (bitconvert (v4i32 MQPR:$src))), (v8f16 MQPR:$src)>;
7415  def : Pat<(v8f16 (bitconvert (v16i8 MQPR:$src))), (v8f16 MQPR:$src)>;
7416
7417  def : Pat<(v8i16 (bitconvert (v2f64 MQPR:$src))), (v8i16 MQPR:$src)>;
7418  def : Pat<(v8i16 (bitconvert (v2i64 MQPR:$src))), (v8i16 MQPR:$src)>;
7419  def : Pat<(v8i16 (bitconvert (v4f32 MQPR:$src))), (v8i16 MQPR:$src)>;
7420  def : Pat<(v8i16 (bitconvert (v4i32 MQPR:$src))), (v8i16 MQPR:$src)>;
7421  def : Pat<(v8i16 (bitconvert (v16i8 MQPR:$src))), (v8i16 MQPR:$src)>;
7422
7423  def : Pat<(v16i8 (bitconvert (v2f64 MQPR:$src))), (v16i8 MQPR:$src)>;
7424  def : Pat<(v16i8 (bitconvert (v2i64 MQPR:$src))), (v16i8 MQPR:$src)>;
7425  def : Pat<(v16i8 (bitconvert (v4f32 MQPR:$src))), (v16i8 MQPR:$src)>;
7426  def : Pat<(v16i8 (bitconvert (v4i32 MQPR:$src))), (v16i8 MQPR:$src)>;
7427  def : Pat<(v16i8 (bitconvert (v8f16 MQPR:$src))), (v16i8 MQPR:$src)>;
7428  def : Pat<(v16i8 (bitconvert (v8i16 MQPR:$src))), (v16i8 MQPR:$src)>;
7429}
7430
7431let Predicates = [IsBE,HasMVEInt] in {
7432  def : Pat<(v2f64 (bitconvert (v4f32 MQPR:$src))), (v2f64 (MVE_VREV64_32 MQPR:$src))>;
7433  def : Pat<(v2f64 (bitconvert (v4i32 MQPR:$src))), (v2f64 (MVE_VREV64_32 MQPR:$src))>;
7434  def : Pat<(v2f64 (bitconvert (v8f16 MQPR:$src))), (v2f64 (MVE_VREV64_16 MQPR:$src))>;
7435  def : Pat<(v2f64 (bitconvert (v8i16 MQPR:$src))), (v2f64 (MVE_VREV64_16 MQPR:$src))>;
7436  def : Pat<(v2f64 (bitconvert (v16i8 MQPR:$src))), (v2f64 (MVE_VREV64_8 MQPR:$src))>;
7437
7438  def : Pat<(v2i64 (bitconvert (v4f32 MQPR:$src))), (v2i64 (MVE_VREV64_32 MQPR:$src))>;
7439  def : Pat<(v2i64 (bitconvert (v4i32 MQPR:$src))), (v2i64 (MVE_VREV64_32 MQPR:$src))>;
7440  def : Pat<(v2i64 (bitconvert (v8f16 MQPR:$src))), (v2i64 (MVE_VREV64_16 MQPR:$src))>;
7441  def : Pat<(v2i64 (bitconvert (v8i16 MQPR:$src))), (v2i64 (MVE_VREV64_16 MQPR:$src))>;
7442  def : Pat<(v2i64 (bitconvert (v16i8 MQPR:$src))), (v2i64 (MVE_VREV64_8 MQPR:$src))>;
7443
7444  def : Pat<(v4f32 (bitconvert (v2f64 MQPR:$src))), (v4f32 (MVE_VREV64_32 MQPR:$src))>;
7445  def : Pat<(v4f32 (bitconvert (v2i64 MQPR:$src))), (v4f32 (MVE_VREV64_32 MQPR:$src))>;
7446  def : Pat<(v4f32 (bitconvert (v8f16 MQPR:$src))), (v4f32 (MVE_VREV32_16 MQPR:$src))>;
7447  def : Pat<(v4f32 (bitconvert (v8i16 MQPR:$src))), (v4f32 (MVE_VREV32_16 MQPR:$src))>;
7448  def : Pat<(v4f32 (bitconvert (v16i8 MQPR:$src))), (v4f32 (MVE_VREV32_8 MQPR:$src))>;
7449
7450  def : Pat<(v4i32 (bitconvert (v2f64 MQPR:$src))), (v4i32 (MVE_VREV64_32 MQPR:$src))>;
7451  def : Pat<(v4i32 (bitconvert (v2i64 MQPR:$src))), (v4i32 (MVE_VREV64_32 MQPR:$src))>;
7452  def : Pat<(v4i32 (bitconvert (v8f16 MQPR:$src))), (v4i32 (MVE_VREV32_16 MQPR:$src))>;
7453  def : Pat<(v4i32 (bitconvert (v8i16 MQPR:$src))), (v4i32 (MVE_VREV32_16 MQPR:$src))>;
7454  def : Pat<(v4i32 (bitconvert (v16i8 MQPR:$src))), (v4i32 (MVE_VREV32_8 MQPR:$src))>;
7455
7456  def : Pat<(v8f16 (bitconvert (v2f64 MQPR:$src))), (v8f16 (MVE_VREV64_16 MQPR:$src))>;
7457  def : Pat<(v8f16 (bitconvert (v2i64 MQPR:$src))), (v8f16 (MVE_VREV64_16 MQPR:$src))>;
7458  def : Pat<(v8f16 (bitconvert (v4f32 MQPR:$src))), (v8f16 (MVE_VREV32_16 MQPR:$src))>;
7459  def : Pat<(v8f16 (bitconvert (v4i32 MQPR:$src))), (v8f16 (MVE_VREV32_16 MQPR:$src))>;
7460  def : Pat<(v8f16 (bitconvert (v16i8 MQPR:$src))), (v8f16 (MVE_VREV16_8 MQPR:$src))>;
7461
7462  def : Pat<(v8i16 (bitconvert (v2f64 MQPR:$src))), (v8i16 (MVE_VREV64_16 MQPR:$src))>;
7463  def : Pat<(v8i16 (bitconvert (v2i64 MQPR:$src))), (v8i16 (MVE_VREV64_16 MQPR:$src))>;
7464  def : Pat<(v8i16 (bitconvert (v4f32 MQPR:$src))), (v8i16 (MVE_VREV32_16 MQPR:$src))>;
7465  def : Pat<(v8i16 (bitconvert (v4i32 MQPR:$src))), (v8i16 (MVE_VREV32_16 MQPR:$src))>;
7466  def : Pat<(v8i16 (bitconvert (v16i8 MQPR:$src))), (v8i16 (MVE_VREV16_8 MQPR:$src))>;
7467
7468  def : Pat<(v16i8 (bitconvert (v2f64 MQPR:$src))), (v16i8 (MVE_VREV64_8 MQPR:$src))>;
7469  def : Pat<(v16i8 (bitconvert (v2i64 MQPR:$src))), (v16i8 (MVE_VREV64_8 MQPR:$src))>;
7470  def : Pat<(v16i8 (bitconvert (v4f32 MQPR:$src))), (v16i8 (MVE_VREV32_8 MQPR:$src))>;
7471  def : Pat<(v16i8 (bitconvert (v4i32 MQPR:$src))), (v16i8 (MVE_VREV32_8 MQPR:$src))>;
7472  def : Pat<(v16i8 (bitconvert (v8f16 MQPR:$src))), (v16i8 (MVE_VREV16_8 MQPR:$src))>;
7473  def : Pat<(v16i8 (bitconvert (v8i16 MQPR:$src))), (v16i8 (MVE_VREV16_8 MQPR:$src))>;
7474}
7475