1 //===-- SystemZISelLowering.h - SystemZ DAG lowering interface --*- C++ -*-===// 2 // 3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. 4 // See https://llvm.org/LICENSE.txt for license information. 5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception 6 // 7 //===----------------------------------------------------------------------===// 8 // 9 // This file defines the interfaces that SystemZ uses to lower LLVM code into a 10 // selection DAG. 11 // 12 //===----------------------------------------------------------------------===// 13 14 #ifndef LLVM_LIB_TARGET_SYSTEMZ_SYSTEMZISELLOWERING_H 15 #define LLVM_LIB_TARGET_SYSTEMZ_SYSTEMZISELLOWERING_H 16 17 #include "SystemZ.h" 18 #include "SystemZInstrInfo.h" 19 #include "llvm/CodeGen/MachineBasicBlock.h" 20 #include "llvm/CodeGen/SelectionDAG.h" 21 #include "llvm/CodeGen/TargetLowering.h" 22 #include <optional> 23 24 namespace llvm { 25 namespace SystemZISD { 26 enum NodeType : unsigned { 27 FIRST_NUMBER = ISD::BUILTIN_OP_END, 28 29 // Return with a glue operand. Operand 0 is the chain operand. 30 RET_GLUE, 31 32 // Calls a function. Operand 0 is the chain operand and operand 1 33 // is the target address. The arguments start at operand 2. 34 // There is an optional glue operand at the end. 35 CALL, 36 SIBCALL, 37 38 // TLS calls. Like regular calls, except operand 1 is the TLS symbol. 39 // (The call target is implicitly __tls_get_offset.) 40 TLS_GDCALL, 41 TLS_LDCALL, 42 43 // Wraps a TargetGlobalAddress that should be loaded using PC-relative 44 // accesses (LARL). Operand 0 is the address. 45 PCREL_WRAPPER, 46 47 // Used in cases where an offset is applied to a TargetGlobalAddress. 48 // Operand 0 is the full TargetGlobalAddress and operand 1 is a 49 // PCREL_WRAPPER for an anchor point. This is used so that we can 50 // cheaply refer to either the full address or the anchor point 51 // as a register base. 52 PCREL_OFFSET, 53 54 // Integer comparisons. There are three operands: the two values 55 // to compare, and an integer of type SystemZICMP. 56 ICMP, 57 58 // Floating-point comparisons. The two operands are the values to compare. 59 FCMP, 60 61 // Test under mask. The first operand is ANDed with the second operand 62 // and the condition codes are set on the result. The third operand is 63 // a boolean that is true if the condition codes need to distinguish 64 // between CCMASK_TM_MIXED_MSB_0 and CCMASK_TM_MIXED_MSB_1 (which the 65 // register forms do but the memory forms don't). 66 TM, 67 68 // Branches if a condition is true. Operand 0 is the chain operand; 69 // operand 1 is the 4-bit condition-code mask, with bit N in 70 // big-endian order meaning "branch if CC=N"; operand 2 is the 71 // target block and operand 3 is the flag operand. 72 BR_CCMASK, 73 74 // Selects between operand 0 and operand 1. Operand 2 is the 75 // mask of condition-code values for which operand 0 should be 76 // chosen over operand 1; it has the same form as BR_CCMASK. 77 // Operand 3 is the flag operand. 78 SELECT_CCMASK, 79 80 // Evaluates to the gap between the stack pointer and the 81 // base of the dynamically-allocatable area. 82 ADJDYNALLOC, 83 84 // For allocating stack space when using stack clash protector. 85 // Allocation is performed by block, and each block is probed. 86 PROBED_ALLOCA, 87 88 // Count number of bits set in operand 0 per byte. 89 POPCNT, 90 91 // Wrappers around the ISD opcodes of the same name. The output is GR128. 92 // Input operands may be GR64 or GR32, depending on the instruction. 93 SMUL_LOHI, 94 UMUL_LOHI, 95 SDIVREM, 96 UDIVREM, 97 98 // Add/subtract with overflow/carry. These have the same operands as 99 // the corresponding standard operations, except with the carry flag 100 // replaced by a condition code value. 101 SADDO, SSUBO, UADDO, USUBO, ADDCARRY, SUBCARRY, 102 103 // Set the condition code from a boolean value in operand 0. 104 // Operand 1 is a mask of all condition-code values that may result of this 105 // operation, operand 2 is a mask of condition-code values that may result 106 // if the boolean is true. 107 // Note that this operation is always optimized away, we will never 108 // generate any code for it. 109 GET_CCMASK, 110 111 // Use a series of MVCs to copy bytes from one memory location to another. 112 // The operands are: 113 // - the target address 114 // - the source address 115 // - the constant length 116 // 117 // This isn't a memory opcode because we'd need to attach two 118 // MachineMemOperands rather than one. 119 MVC, 120 121 // Similar to MVC, but for logic operations (AND, OR, XOR). 122 NC, 123 OC, 124 XC, 125 126 // Use CLC to compare two blocks of memory, with the same comments 127 // as for MVC. 128 CLC, 129 130 // Use MVC to set a block of memory after storing the first byte. 131 MEMSET_MVC, 132 133 // Use an MVST-based sequence to implement stpcpy(). 134 STPCPY, 135 136 // Use a CLST-based sequence to implement strcmp(). The two input operands 137 // are the addresses of the strings to compare. 138 STRCMP, 139 140 // Use an SRST-based sequence to search a block of memory. The first 141 // operand is the end address, the second is the start, and the third 142 // is the character to search for. CC is set to 1 on success and 2 143 // on failure. 144 SEARCH_STRING, 145 146 // Store the CC value in bits 29 and 28 of an integer. 147 IPM, 148 149 // Transaction begin. The first operand is the chain, the second 150 // the TDB pointer, and the third the immediate control field. 151 // Returns CC value and chain. 152 TBEGIN, 153 TBEGIN_NOFLOAT, 154 155 // Transaction end. Just the chain operand. Returns CC value and chain. 156 TEND, 157 158 // Create a vector constant by filling byte N of the result with bit 159 // 15-N of the single operand. 160 BYTE_MASK, 161 162 // Create a vector constant by replicating an element-sized RISBG-style mask. 163 // The first operand specifies the starting set bit and the second operand 164 // specifies the ending set bit. Both operands count from the MSB of the 165 // element. 166 ROTATE_MASK, 167 168 // Replicate a GPR scalar value into all elements of a vector. 169 REPLICATE, 170 171 // Create a vector from two i64 GPRs. 172 JOIN_DWORDS, 173 174 // Replicate one element of a vector into all elements. The first operand 175 // is the vector and the second is the index of the element to replicate. 176 SPLAT, 177 178 // Interleave elements from the high half of operand 0 and the high half 179 // of operand 1. 180 MERGE_HIGH, 181 182 // Likewise for the low halves. 183 MERGE_LOW, 184 185 // Concatenate the vectors in the first two operands, shift them left 186 // by the third operand, and take the first half of the result. 187 SHL_DOUBLE, 188 189 // Take one element of the first v2i64 operand and the one element of 190 // the second v2i64 operand and concatenate them to form a v2i64 result. 191 // The third operand is a 4-bit value of the form 0A0B, where A and B 192 // are the element selectors for the first operand and second operands 193 // respectively. 194 PERMUTE_DWORDS, 195 196 // Perform a general vector permute on vector operands 0 and 1. 197 // Each byte of operand 2 controls the corresponding byte of the result, 198 // in the same way as a byte-level VECTOR_SHUFFLE mask. 199 PERMUTE, 200 201 // Pack vector operands 0 and 1 into a single vector with half-sized elements. 202 PACK, 203 204 // Likewise, but saturate the result and set CC. PACKS_CC does signed 205 // saturation and PACKLS_CC does unsigned saturation. 206 PACKS_CC, 207 PACKLS_CC, 208 209 // Unpack the first half of vector operand 0 into double-sized elements. 210 // UNPACK_HIGH sign-extends and UNPACKL_HIGH zero-extends. 211 UNPACK_HIGH, 212 UNPACKL_HIGH, 213 214 // Likewise for the second half. 215 UNPACK_LOW, 216 UNPACKL_LOW, 217 218 // Shift/rotate each element of vector operand 0 by the number of bits 219 // specified by scalar operand 1. 220 VSHL_BY_SCALAR, 221 VSRL_BY_SCALAR, 222 VSRA_BY_SCALAR, 223 VROTL_BY_SCALAR, 224 225 // For each element of the output type, sum across all sub-elements of 226 // operand 0 belonging to the corresponding element, and add in the 227 // rightmost sub-element of the corresponding element of operand 1. 228 VSUM, 229 230 // Compute carry/borrow indication for add/subtract. 231 VACC, VSCBI, 232 // Add/subtract with carry/borrow. 233 VAC, VSBI, 234 // Compute carry/borrow indication for add/subtract with carry/borrow. 235 VACCC, VSBCBI, 236 237 // Compare integer vector operands 0 and 1 to produce the usual 0/-1 238 // vector result. VICMPE is for equality, VICMPH for "signed greater than" 239 // and VICMPHL for "unsigned greater than". 240 VICMPE, 241 VICMPH, 242 VICMPHL, 243 244 // Likewise, but also set the condition codes on the result. 245 VICMPES, 246 VICMPHS, 247 VICMPHLS, 248 249 // Compare floating-point vector operands 0 and 1 to produce the usual 0/-1 250 // vector result. VFCMPE is for "ordered and equal", VFCMPH for "ordered and 251 // greater than" and VFCMPHE for "ordered and greater than or equal to". 252 VFCMPE, 253 VFCMPH, 254 VFCMPHE, 255 256 // Likewise, but also set the condition codes on the result. 257 VFCMPES, 258 VFCMPHS, 259 VFCMPHES, 260 261 // Test floating-point data class for vectors. 262 VFTCI, 263 264 // Extend the even f32 elements of vector operand 0 to produce a vector 265 // of f64 elements. 266 VEXTEND, 267 268 // Round the f64 elements of vector operand 0 to f32s and store them in the 269 // even elements of the result. 270 VROUND, 271 272 // AND the two vector operands together and set CC based on the result. 273 VTM, 274 275 // i128 high integer comparisons. 276 SCMP128HI, 277 UCMP128HI, 278 279 // String operations that set CC as a side-effect. 280 VFAE_CC, 281 VFAEZ_CC, 282 VFEE_CC, 283 VFEEZ_CC, 284 VFENE_CC, 285 VFENEZ_CC, 286 VISTR_CC, 287 VSTRC_CC, 288 VSTRCZ_CC, 289 VSTRS_CC, 290 VSTRSZ_CC, 291 292 // Test Data Class. 293 // 294 // Operand 0: the value to test 295 // Operand 1: the bit mask 296 TDC, 297 298 // z/OS XPLINK ADA Entry 299 // Wraps a TargetGlobalAddress that should be loaded from a function's 300 // AssociatedData Area (ADA). Tha ADA is passed to the function by the 301 // caller in the XPLink ABI defined register R5. 302 // Operand 0: the GlobalValue/External Symbol 303 // Operand 1: the ADA register 304 // Operand 2: the offset (0 for the first and 8 for the second element in the 305 // function descriptor) 306 ADA_ENTRY, 307 308 // Strict variants of scalar floating-point comparisons. 309 // Quiet and signaling versions. 310 FIRST_STRICTFP_OPCODE, 311 STRICT_FCMP = FIRST_STRICTFP_OPCODE, 312 STRICT_FCMPS, 313 314 // Strict variants of vector floating-point comparisons. 315 // Quiet and signaling versions. 316 STRICT_VFCMPE, 317 STRICT_VFCMPH, 318 STRICT_VFCMPHE, 319 STRICT_VFCMPES, 320 STRICT_VFCMPHS, 321 STRICT_VFCMPHES, 322 323 // Strict variants of VEXTEND and VROUND. 324 STRICT_VEXTEND, 325 STRICT_VROUND, 326 LAST_STRICTFP_OPCODE = STRICT_VROUND, 327 328 // Wrappers around the inner loop of an 8- or 16-bit ATOMIC_SWAP or 329 // ATOMIC_LOAD_<op>. 330 // 331 // Operand 0: the address of the containing 32-bit-aligned field 332 // Operand 1: the second operand of <op>, in the high bits of an i32 333 // for everything except ATOMIC_SWAPW 334 // Operand 2: how many bits to rotate the i32 left to bring the first 335 // operand into the high bits 336 // Operand 3: the negative of operand 2, for rotating the other way 337 // Operand 4: the width of the field in bits (8 or 16) 338 FIRST_MEMORY_OPCODE, 339 ATOMIC_SWAPW = FIRST_MEMORY_OPCODE, 340 ATOMIC_LOADW_ADD, 341 ATOMIC_LOADW_SUB, 342 ATOMIC_LOADW_AND, 343 ATOMIC_LOADW_OR, 344 ATOMIC_LOADW_XOR, 345 ATOMIC_LOADW_NAND, 346 ATOMIC_LOADW_MIN, 347 ATOMIC_LOADW_MAX, 348 ATOMIC_LOADW_UMIN, 349 ATOMIC_LOADW_UMAX, 350 351 // A wrapper around the inner loop of an ATOMIC_CMP_SWAP. 352 // 353 // Operand 0: the address of the containing 32-bit-aligned field 354 // Operand 1: the compare value, in the low bits of an i32 355 // Operand 2: the swap value, in the low bits of an i32 356 // Operand 3: how many bits to rotate the i32 left to bring the first 357 // operand into the high bits 358 // Operand 4: the negative of operand 2, for rotating the other way 359 // Operand 5: the width of the field in bits (8 or 16) 360 ATOMIC_CMP_SWAPW, 361 362 // Atomic compare-and-swap returning CC value. 363 // Val, CC, OUTCHAIN = ATOMIC_CMP_SWAP(INCHAIN, ptr, cmp, swap) 364 ATOMIC_CMP_SWAP, 365 366 // 128-bit atomic load. 367 // Val, OUTCHAIN = ATOMIC_LOAD_128(INCHAIN, ptr) 368 ATOMIC_LOAD_128, 369 370 // 128-bit atomic store. 371 // OUTCHAIN = ATOMIC_STORE_128(INCHAIN, val, ptr) 372 ATOMIC_STORE_128, 373 374 // 128-bit atomic compare-and-swap. 375 // Val, CC, OUTCHAIN = ATOMIC_CMP_SWAP(INCHAIN, ptr, cmp, swap) 376 ATOMIC_CMP_SWAP_128, 377 378 // Byte swapping load/store. Same operands as regular load/store. 379 LRV, STRV, 380 381 // Element swapping load/store. Same operands as regular load/store. 382 VLER, VSTER, 383 384 // Use STORE CLOCK FAST to store current TOD clock value. 385 STCKF, 386 387 // Prefetch from the second operand using the 4-bit control code in 388 // the first operand. The code is 1 for a load prefetch and 2 for 389 // a store prefetch. 390 PREFETCH, 391 LAST_MEMORY_OPCODE = PREFETCH, 392 }; 393 394 // Return true if OPCODE is some kind of PC-relative address. 395 inline bool isPCREL(unsigned Opcode) { 396 return Opcode == PCREL_WRAPPER || Opcode == PCREL_OFFSET; 397 } 398 } // end namespace SystemZISD 399 400 namespace SystemZICMP { 401 // Describes whether an integer comparison needs to be signed or unsigned, 402 // or whether either type is OK. 403 enum { 404 Any, 405 UnsignedOnly, 406 SignedOnly 407 }; 408 } // end namespace SystemZICMP 409 410 class SystemZSubtarget; 411 412 class SystemZTargetLowering : public TargetLowering { 413 public: 414 explicit SystemZTargetLowering(const TargetMachine &TM, 415 const SystemZSubtarget &STI); 416 417 bool useSoftFloat() const override; 418 419 // Override TargetLowering. 420 MVT getScalarShiftAmountTy(const DataLayout &, EVT) const override { 421 return MVT::i32; 422 } 423 MVT getVectorIdxTy(const DataLayout &DL) const override { 424 // Only the lower 12 bits of an element index are used, so we don't 425 // want to clobber the upper 32 bits of a GPR unnecessarily. 426 return MVT::i32; 427 } 428 TargetLoweringBase::LegalizeTypeAction getPreferredVectorAction(MVT VT) 429 const override { 430 // Widen subvectors to the full width rather than promoting integer 431 // elements. This is better because: 432 // 433 // (a) it means that we can handle the ABI for passing and returning 434 // sub-128 vectors without having to handle them as legal types. 435 // 436 // (b) we don't have instructions to extend on load and truncate on store, 437 // so promoting the integers is less efficient. 438 // 439 // (c) there are no multiplication instructions for the widest integer 440 // type (v2i64). 441 if (VT.getScalarSizeInBits() % 8 == 0) 442 return TypeWidenVector; 443 return TargetLoweringBase::getPreferredVectorAction(VT); 444 } 445 unsigned 446 getNumRegisters(LLVMContext &Context, EVT VT, 447 std::optional<MVT> RegisterVT) const override { 448 // i128 inline assembly operand. 449 if (VT == MVT::i128 && RegisterVT && *RegisterVT == MVT::Untyped) 450 return 1; 451 return TargetLowering::getNumRegisters(Context, VT); 452 } 453 MVT getRegisterTypeForCallingConv(LLVMContext &Context, CallingConv::ID CC, 454 EVT VT) const override { 455 // 128-bit single-element vector types are passed like other vectors, 456 // not like their element type. 457 if (VT.isVector() && VT.getSizeInBits() == 128 && 458 VT.getVectorNumElements() == 1) 459 return MVT::v16i8; 460 return TargetLowering::getRegisterTypeForCallingConv(Context, CC, VT); 461 } 462 bool isCheapToSpeculateCtlz(Type *) const override { return true; } 463 bool isCheapToSpeculateCttz(Type *) const override { return true; } 464 bool preferZeroCompareBranch() const override { return true; } 465 bool isMaskAndCmp0FoldingBeneficial(const Instruction &AndI) const override { 466 ConstantInt* Mask = dyn_cast<ConstantInt>(AndI.getOperand(1)); 467 return Mask && Mask->getValue().isIntN(16); 468 } 469 bool convertSetCCLogicToBitwiseLogic(EVT VT) const override { 470 return VT.isScalarInteger(); 471 } 472 EVT getSetCCResultType(const DataLayout &DL, LLVMContext &, 473 EVT) const override; 474 bool isFMAFasterThanFMulAndFAdd(const MachineFunction &MF, 475 EVT VT) const override; 476 bool isFPImmLegal(const APFloat &Imm, EVT VT, 477 bool ForCodeSize) const override; 478 bool ShouldShrinkFPConstant(EVT VT) const override { 479 // Do not shrink 64-bit FP constpool entries since LDEB is slower than 480 // LD, and having the full constant in memory enables reg/mem opcodes. 481 return VT != MVT::f64; 482 } 483 MachineBasicBlock *emitEHSjLjSetJmp(MachineInstr &MI, 484 MachineBasicBlock *MBB) const; 485 486 MachineBasicBlock *emitEHSjLjLongJmp(MachineInstr &MI, 487 MachineBasicBlock *MBB) const; 488 489 bool hasInlineStackProbe(const MachineFunction &MF) const override; 490 AtomicExpansionKind shouldCastAtomicLoadInIR(LoadInst *LI) const override; 491 AtomicExpansionKind shouldCastAtomicStoreInIR(StoreInst *SI) const override; 492 AtomicExpansionKind 493 shouldExpandAtomicRMWInIR(AtomicRMWInst *RMW) const override; 494 bool isLegalICmpImmediate(int64_t Imm) const override; 495 bool isLegalAddImmediate(int64_t Imm) const override; 496 bool isLegalAddressingMode(const DataLayout &DL, const AddrMode &AM, Type *Ty, 497 unsigned AS, 498 Instruction *I = nullptr) const override; 499 bool allowsMisalignedMemoryAccesses(EVT VT, unsigned AS, Align Alignment, 500 MachineMemOperand::Flags Flags, 501 unsigned *Fast) const override; 502 bool 503 findOptimalMemOpLowering(std::vector<EVT> &MemOps, unsigned Limit, 504 const MemOp &Op, unsigned DstAS, unsigned SrcAS, 505 const AttributeList &FuncAttributes) const override; 506 EVT getOptimalMemOpType(const MemOp &Op, 507 const AttributeList &FuncAttributes) const override; 508 bool isTruncateFree(Type *, Type *) const override; 509 bool isTruncateFree(EVT, EVT) const override; 510 511 bool shouldFormOverflowOp(unsigned Opcode, EVT VT, 512 bool MathUsed) const override { 513 // Form add and sub with overflow intrinsics regardless of any extra 514 // users of the math result. 515 return VT == MVT::i32 || VT == MVT::i64; 516 } 517 518 bool shouldConsiderGEPOffsetSplit() const override { return true; } 519 520 bool shouldExpandCmpUsingSelects(EVT VT) const override { return true; } 521 522 const char *getTargetNodeName(unsigned Opcode) const override; 523 std::pair<unsigned, const TargetRegisterClass *> 524 getRegForInlineAsmConstraint(const TargetRegisterInfo *TRI, 525 StringRef Constraint, MVT VT) const override; 526 TargetLowering::ConstraintType 527 getConstraintType(StringRef Constraint) const override; 528 TargetLowering::ConstraintWeight 529 getSingleConstraintMatchWeight(AsmOperandInfo &info, 530 const char *constraint) const override; 531 void LowerAsmOperandForConstraint(SDValue Op, StringRef Constraint, 532 std::vector<SDValue> &Ops, 533 SelectionDAG &DAG) const override; 534 535 InlineAsm::ConstraintCode 536 getInlineAsmMemConstraint(StringRef ConstraintCode) const override { 537 if (ConstraintCode.size() == 1) { 538 switch(ConstraintCode[0]) { 539 default: 540 break; 541 case 'o': 542 return InlineAsm::ConstraintCode::o; 543 case 'Q': 544 return InlineAsm::ConstraintCode::Q; 545 case 'R': 546 return InlineAsm::ConstraintCode::R; 547 case 'S': 548 return InlineAsm::ConstraintCode::S; 549 case 'T': 550 return InlineAsm::ConstraintCode::T; 551 } 552 } else if (ConstraintCode.size() == 2 && ConstraintCode[0] == 'Z') { 553 switch (ConstraintCode[1]) { 554 default: 555 break; 556 case 'Q': 557 return InlineAsm::ConstraintCode::ZQ; 558 case 'R': 559 return InlineAsm::ConstraintCode::ZR; 560 case 'S': 561 return InlineAsm::ConstraintCode::ZS; 562 case 'T': 563 return InlineAsm::ConstraintCode::ZT; 564 } 565 } 566 return TargetLowering::getInlineAsmMemConstraint(ConstraintCode); 567 } 568 569 Register getRegisterByName(const char *RegName, LLT VT, 570 const MachineFunction &MF) const override; 571 572 /// If a physical register, this returns the register that receives the 573 /// exception address on entry to an EH pad. 574 Register 575 getExceptionPointerRegister(const Constant *PersonalityFn) const override; 576 577 /// If a physical register, this returns the register that receives the 578 /// exception typeid on entry to a landing pad. 579 Register 580 getExceptionSelectorRegister(const Constant *PersonalityFn) const override; 581 582 /// Override to support customized stack guard loading. 583 bool useLoadStackGuardNode(const Module &M) const override { return true; } 584 void insertSSPDeclarations(Module &M) const override { 585 } 586 587 MachineBasicBlock * 588 EmitInstrWithCustomInserter(MachineInstr &MI, 589 MachineBasicBlock *BB) const override; 590 SDValue LowerOperation(SDValue Op, SelectionDAG &DAG) const override; 591 void LowerOperationWrapper(SDNode *N, SmallVectorImpl<SDValue> &Results, 592 SelectionDAG &DAG) const override; 593 void ReplaceNodeResults(SDNode *N, SmallVectorImpl<SDValue>&Results, 594 SelectionDAG &DAG) const override; 595 const MCPhysReg *getScratchRegisters(CallingConv::ID CC) const override; 596 bool allowTruncateForTailCall(Type *, Type *) const override; 597 bool mayBeEmittedAsTailCall(const CallInst *CI) const override; 598 bool splitValueIntoRegisterParts( 599 SelectionDAG & DAG, const SDLoc &DL, SDValue Val, SDValue *Parts, 600 unsigned NumParts, MVT PartVT, std::optional<CallingConv::ID> CC) 601 const override; 602 SDValue joinRegisterPartsIntoValue( 603 SelectionDAG & DAG, const SDLoc &DL, const SDValue *Parts, 604 unsigned NumParts, MVT PartVT, EVT ValueVT, 605 std::optional<CallingConv::ID> CC) const override; 606 SDValue LowerFormalArguments(SDValue Chain, CallingConv::ID CallConv, 607 bool isVarArg, 608 const SmallVectorImpl<ISD::InputArg> &Ins, 609 const SDLoc &DL, SelectionDAG &DAG, 610 SmallVectorImpl<SDValue> &InVals) const override; 611 SDValue LowerCall(CallLoweringInfo &CLI, 612 SmallVectorImpl<SDValue> &InVals) const override; 613 614 std::pair<SDValue, SDValue> 615 makeExternalCall(SDValue Chain, SelectionDAG &DAG, const char *CalleeName, 616 EVT RetVT, ArrayRef<SDValue> Ops, CallingConv::ID CallConv, 617 bool IsSigned, SDLoc DL, bool DoesNotReturn, 618 bool IsReturnValueUsed) const; 619 620 bool CanLowerReturn(CallingConv::ID CallConv, MachineFunction &MF, 621 bool isVarArg, 622 const SmallVectorImpl<ISD::OutputArg> &Outs, 623 LLVMContext &Context, 624 const Type *RetTy) const override; 625 SDValue LowerReturn(SDValue Chain, CallingConv::ID CallConv, bool IsVarArg, 626 const SmallVectorImpl<ISD::OutputArg> &Outs, 627 const SmallVectorImpl<SDValue> &OutVals, const SDLoc &DL, 628 SelectionDAG &DAG) const override; 629 SDValue PerformDAGCombine(SDNode *N, DAGCombinerInfo &DCI) const override; 630 631 /// Determine which of the bits specified in Mask are known to be either 632 /// zero or one and return them in the KnownZero/KnownOne bitsets. 633 void computeKnownBitsForTargetNode(const SDValue Op, 634 KnownBits &Known, 635 const APInt &DemandedElts, 636 const SelectionDAG &DAG, 637 unsigned Depth = 0) const override; 638 639 /// Determine the number of bits in the operation that are sign bits. 640 unsigned ComputeNumSignBitsForTargetNode(SDValue Op, 641 const APInt &DemandedElts, 642 const SelectionDAG &DAG, 643 unsigned Depth) const override; 644 645 bool isGuaranteedNotToBeUndefOrPoisonForTargetNode( 646 SDValue Op, const APInt &DemandedElts, const SelectionDAG &DAG, 647 bool PoisonOnly, unsigned Depth) const override; 648 649 ISD::NodeType getExtendForAtomicOps() const override { 650 return ISD::ANY_EXTEND; 651 } 652 ISD::NodeType getExtendForAtomicCmpSwapArg() const override { 653 return ISD::ZERO_EXTEND; 654 } 655 656 bool supportSwiftError() const override { 657 return true; 658 } 659 660 unsigned getStackProbeSize(const MachineFunction &MF) const; 661 662 private: 663 const SystemZSubtarget &Subtarget; 664 665 // Implement LowerOperation for individual opcodes. 666 SDValue getVectorCmp(SelectionDAG &DAG, unsigned Opcode, 667 const SDLoc &DL, EVT VT, 668 SDValue CmpOp0, SDValue CmpOp1, SDValue Chain) const; 669 SDValue lowerVectorSETCC(SelectionDAG &DAG, const SDLoc &DL, 670 EVT VT, ISD::CondCode CC, 671 SDValue CmpOp0, SDValue CmpOp1, 672 SDValue Chain = SDValue(), 673 bool IsSignaling = false) const; 674 SDValue lowerSETCC(SDValue Op, SelectionDAG &DAG) const; 675 SDValue lowerSTRICT_FSETCC(SDValue Op, SelectionDAG &DAG, 676 bool IsSignaling) const; 677 SDValue lowerBR_CC(SDValue Op, SelectionDAG &DAG) const; 678 SDValue lowerSELECT_CC(SDValue Op, SelectionDAG &DAG) const; 679 SDValue lowerGlobalAddress(GlobalAddressSDNode *Node, 680 SelectionDAG &DAG) const; 681 SDValue lowerTLSGetOffset(GlobalAddressSDNode *Node, 682 SelectionDAG &DAG, unsigned Opcode, 683 SDValue GOTOffset) const; 684 SDValue lowerThreadPointer(const SDLoc &DL, SelectionDAG &DAG) const; 685 SDValue lowerGlobalTLSAddress(GlobalAddressSDNode *Node, 686 SelectionDAG &DAG) const; 687 SDValue lowerBlockAddress(BlockAddressSDNode *Node, 688 SelectionDAG &DAG) const; 689 SDValue lowerJumpTable(JumpTableSDNode *JT, SelectionDAG &DAG) const; 690 SDValue lowerConstantPool(ConstantPoolSDNode *CP, SelectionDAG &DAG) const; 691 SDValue lowerFRAMEADDR(SDValue Op, SelectionDAG &DAG) const; 692 SDValue lowerRETURNADDR(SDValue Op, SelectionDAG &DAG) const; 693 SDValue lowerVASTART(SDValue Op, SelectionDAG &DAG) const; 694 SDValue lowerVASTART_ELF(SDValue Op, SelectionDAG &DAG) const; 695 SDValue lowerVASTART_XPLINK(SDValue Op, SelectionDAG &DAG) const; 696 SDValue lowerVACOPY(SDValue Op, SelectionDAG &DAG) const; 697 SDValue lowerDYNAMIC_STACKALLOC(SDValue Op, SelectionDAG &DAG) const; 698 SDValue lowerDYNAMIC_STACKALLOC_ELF(SDValue Op, SelectionDAG &DAG) const; 699 SDValue lowerDYNAMIC_STACKALLOC_XPLINK(SDValue Op, SelectionDAG &DAG) const; 700 SDValue lowerGET_DYNAMIC_AREA_OFFSET(SDValue Op, SelectionDAG &DAG) const; 701 SDValue lowerMULH(SDValue Op, SelectionDAG &DAG, unsigned Opcode) const; 702 SDValue lowerSMUL_LOHI(SDValue Op, SelectionDAG &DAG) const; 703 SDValue lowerUMUL_LOHI(SDValue Op, SelectionDAG &DAG) const; 704 SDValue lowerSDIVREM(SDValue Op, SelectionDAG &DAG) const; 705 SDValue lowerUDIVREM(SDValue Op, SelectionDAG &DAG) const; 706 SDValue lowerXALUO(SDValue Op, SelectionDAG &DAG) const; 707 SDValue lowerUADDSUBO_CARRY(SDValue Op, SelectionDAG &DAG) const; 708 SDValue lowerBITCAST(SDValue Op, SelectionDAG &DAG) const; 709 SDValue lowerOR(SDValue Op, SelectionDAG &DAG) const; 710 SDValue lowerCTPOP(SDValue Op, SelectionDAG &DAG) const; 711 SDValue lowerVECREDUCE_ADD(SDValue Op, SelectionDAG &DAG) const; 712 SDValue lowerATOMIC_FENCE(SDValue Op, SelectionDAG &DAG) const; 713 SDValue lowerATOMIC_LDST_I128(SDValue Op, SelectionDAG &DAG) const; 714 SDValue lowerATOMIC_LOAD_OP(SDValue Op, SelectionDAG &DAG, 715 unsigned Opcode) const; 716 SDValue lowerATOMIC_LOAD_SUB(SDValue Op, SelectionDAG &DAG) const; 717 SDValue lowerATOMIC_CMP_SWAP(SDValue Op, SelectionDAG &DAG) const; 718 SDValue lowerSTACKSAVE(SDValue Op, SelectionDAG &DAG) const; 719 SDValue lowerSTACKRESTORE(SDValue Op, SelectionDAG &DAG) const; 720 SDValue lowerPREFETCH(SDValue Op, SelectionDAG &DAG) const; 721 SDValue lowerINTRINSIC_W_CHAIN(SDValue Op, SelectionDAG &DAG) const; 722 SDValue lowerINTRINSIC_WO_CHAIN(SDValue Op, SelectionDAG &DAG) const; 723 bool isVectorElementLoad(SDValue Op) const; 724 SDValue buildVector(SelectionDAG &DAG, const SDLoc &DL, EVT VT, 725 SmallVectorImpl<SDValue> &Elems) const; 726 SDValue lowerBUILD_VECTOR(SDValue Op, SelectionDAG &DAG) const; 727 SDValue lowerVECTOR_SHUFFLE(SDValue Op, SelectionDAG &DAG) const; 728 SDValue lowerSCALAR_TO_VECTOR(SDValue Op, SelectionDAG &DAG) const; 729 SDValue lowerINSERT_VECTOR_ELT(SDValue Op, SelectionDAG &DAG) const; 730 SDValue lowerEXTRACT_VECTOR_ELT(SDValue Op, SelectionDAG &DAG) const; 731 SDValue lowerSIGN_EXTEND_VECTOR_INREG(SDValue Op, SelectionDAG &DAG) const; 732 SDValue lowerZERO_EXTEND_VECTOR_INREG(SDValue Op, SelectionDAG &DAG) const; 733 SDValue lowerShift(SDValue Op, SelectionDAG &DAG, unsigned ByScalar) const; 734 SDValue lowerIS_FPCLASS(SDValue Op, SelectionDAG &DAG) const; 735 SDValue lowerGET_ROUNDING(SDValue Op, SelectionDAG &DAG) const; 736 SDValue lowerREADCYCLECOUNTER(SDValue Op, SelectionDAG &DAG) const; 737 738 bool canTreatAsByteVector(EVT VT) const; 739 SDValue combineExtract(const SDLoc &DL, EVT ElemVT, EVT VecVT, SDValue OrigOp, 740 unsigned Index, DAGCombinerInfo &DCI, 741 bool Force) const; 742 SDValue combineTruncateExtract(const SDLoc &DL, EVT TruncVT, SDValue Op, 743 DAGCombinerInfo &DCI) const; 744 SDValue combineZERO_EXTEND(SDNode *N, DAGCombinerInfo &DCI) const; 745 SDValue combineSIGN_EXTEND(SDNode *N, DAGCombinerInfo &DCI) const; 746 SDValue combineSIGN_EXTEND_INREG(SDNode *N, DAGCombinerInfo &DCI) const; 747 SDValue combineMERGE(SDNode *N, DAGCombinerInfo &DCI) const; 748 bool canLoadStoreByteSwapped(EVT VT) const; 749 SDValue combineLOAD(SDNode *N, DAGCombinerInfo &DCI) const; 750 SDValue combineSTORE(SDNode *N, DAGCombinerInfo &DCI) const; 751 SDValue combineVECTOR_SHUFFLE(SDNode *N, DAGCombinerInfo &DCI) const; 752 SDValue combineEXTRACT_VECTOR_ELT(SDNode *N, DAGCombinerInfo &DCI) const; 753 SDValue combineJOIN_DWORDS(SDNode *N, DAGCombinerInfo &DCI) const; 754 SDValue combineFP_ROUND(SDNode *N, DAGCombinerInfo &DCI) const; 755 SDValue combineFP_EXTEND(SDNode *N, DAGCombinerInfo &DCI) const; 756 SDValue combineINT_TO_FP(SDNode *N, DAGCombinerInfo &DCI) const; 757 SDValue combineBSWAP(SDNode *N, DAGCombinerInfo &DCI) const; 758 SDValue combineBR_CCMASK(SDNode *N, DAGCombinerInfo &DCI) const; 759 SDValue combineSELECT_CCMASK(SDNode *N, DAGCombinerInfo &DCI) const; 760 SDValue combineGET_CCMASK(SDNode *N, DAGCombinerInfo &DCI) const; 761 SDValue combineIntDIVREM(SDNode *N, DAGCombinerInfo &DCI) const; 762 SDValue combineINTRINSIC(SDNode *N, DAGCombinerInfo &DCI) const; 763 764 SDValue unwrapAddress(SDValue N) const override; 765 766 // If the last instruction before MBBI in MBB was some form of COMPARE, 767 // try to replace it with a COMPARE AND BRANCH just before MBBI. 768 // CCMask and Target are the BRC-like operands for the branch. 769 // Return true if the change was made. 770 bool convertPrevCompareToBranch(MachineBasicBlock *MBB, 771 MachineBasicBlock::iterator MBBI, 772 unsigned CCMask, 773 MachineBasicBlock *Target) const; 774 775 // Implement EmitInstrWithCustomInserter for individual operation types. 776 MachineBasicBlock *emitAdjCallStack(MachineInstr &MI, 777 MachineBasicBlock *BB) const; 778 MachineBasicBlock *emitSelect(MachineInstr &MI, MachineBasicBlock *BB) const; 779 MachineBasicBlock *emitCondStore(MachineInstr &MI, MachineBasicBlock *BB, 780 unsigned StoreOpcode, unsigned STOCOpcode, 781 bool Invert) const; 782 MachineBasicBlock *emitICmp128Hi(MachineInstr &MI, MachineBasicBlock *BB, 783 bool Unsigned) const; 784 MachineBasicBlock *emitPair128(MachineInstr &MI, 785 MachineBasicBlock *MBB) const; 786 MachineBasicBlock *emitExt128(MachineInstr &MI, MachineBasicBlock *MBB, 787 bool ClearEven) const; 788 MachineBasicBlock *emitAtomicLoadBinary(MachineInstr &MI, 789 MachineBasicBlock *BB, 790 unsigned BinOpcode, 791 bool Invert = false) const; 792 MachineBasicBlock *emitAtomicLoadMinMax(MachineInstr &MI, 793 MachineBasicBlock *MBB, 794 unsigned CompareOpcode, 795 unsigned KeepOldMask) const; 796 MachineBasicBlock *emitAtomicCmpSwapW(MachineInstr &MI, 797 MachineBasicBlock *BB) const; 798 MachineBasicBlock *emitMemMemWrapper(MachineInstr &MI, MachineBasicBlock *BB, 799 unsigned Opcode, 800 bool IsMemset = false) const; 801 MachineBasicBlock *emitStringWrapper(MachineInstr &MI, MachineBasicBlock *BB, 802 unsigned Opcode) const; 803 MachineBasicBlock *emitTransactionBegin(MachineInstr &MI, 804 MachineBasicBlock *MBB, 805 unsigned Opcode, bool NoFloat) const; 806 MachineBasicBlock *emitLoadAndTestCmp0(MachineInstr &MI, 807 MachineBasicBlock *MBB, 808 unsigned Opcode) const; 809 MachineBasicBlock *emitProbedAlloca(MachineInstr &MI, 810 MachineBasicBlock *MBB) const; 811 812 SDValue getBackchainAddress(SDValue SP, SelectionDAG &DAG) const; 813 814 MachineMemOperand::Flags 815 getTargetMMOFlags(const Instruction &I) const override; 816 const TargetRegisterClass *getRepRegClassFor(MVT VT) const override; 817 818 bool isFullyInternal(const Function *Fn) const; 819 void verifyNarrowIntegerArgs_Call(const SmallVectorImpl<ISD::OutputArg> &Outs, 820 const Function *F, SDValue Callee) const; 821 void verifyNarrowIntegerArgs_Ret(const SmallVectorImpl<ISD::OutputArg> &Outs, 822 const Function *F) const; 823 bool verifyNarrowIntegerArgs(const SmallVectorImpl<ISD::OutputArg> &Outs, 824 bool IsInternal) const; 825 }; 826 827 struct SystemZVectorConstantInfo { 828 private: 829 APInt IntBits; // The 128 bits as an integer. 830 APInt SplatBits; // Smallest splat value. 831 APInt SplatUndef; // Bits correspoding to undef operands of the BVN. 832 unsigned SplatBitSize = 0; 833 bool isFP128 = false; 834 public: 835 unsigned Opcode = 0; 836 SmallVector<unsigned, 2> OpVals; 837 MVT VecVT; 838 SystemZVectorConstantInfo(APInt IntImm); 839 SystemZVectorConstantInfo(APFloat FPImm) 840 : SystemZVectorConstantInfo(FPImm.bitcastToAPInt()) { 841 isFP128 = (&FPImm.getSemantics() == &APFloat::IEEEquad()); 842 } 843 SystemZVectorConstantInfo(BuildVectorSDNode *BVN); 844 bool isVectorConstantLegal(const SystemZSubtarget &Subtarget); 845 }; 846 847 } // end namespace llvm 848 849 #endif 850