| /netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/Support/ |
| H A D | ScaledNumber.h | 52 inline std::pair<DigitsT, int16_t> getRounded(DigitsT Digits, int16_t Scale, in getRounded() argument 59 return std::make_pair(DigitsT(1) << (getWidth<DigitsT>() - 1), Scale + 1); in getRounded() 60 return std::make_pair(Digits, Scale); in getRounded() 64 inline std::pair<uint32_t, int16_t> getRounded32(uint32_t Digits, int16_t Scale, in getRounded32() argument 66 return getRounded(Digits, Scale, ShouldRound); in getRounded32() 70 inline std::pair<uint64_t, int16_t> getRounded64(uint64_t Digits, int16_t Scale, in getRounded64() argument 72 return getRounded(Digits, Scale, ShouldRound); in getRounded64() 80 int16_t Scale = 0) { 85 return std::make_pair(Digits, Scale); 89 return getRounded<DigitsT>(Digits >> Shift, Scale + Shift, [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/X86/ |
| H A D | X86InstrBuilder.h | 53 unsigned Scale; member 60 : BaseType(RegBase), Scale(1), IndexReg(0), Disp(0), GV(nullptr), in X86AddressMode() 66 assert(Scale == 1 || Scale == 2 || Scale == 4 || Scale == 8); in getFullAddress() 76 MO.push_back(MachineOperand::CreateImm(Scale)); in getFullAddress() 105 AM.Scale = Op1.getImm(); in getAddressFromInstr() 174 assert(AM.Scale == 1 || AM.Scale == 2 || AM.Scale == 4 || AM.Scale == 8); in addFullAddress() 183 MIB.addImm(AM.Scale).addReg(AM.IndexReg); in addFullAddress()
|
| H A D | X86ISelDAGToDAG.cpp | 68 unsigned Scale; member 83 : BaseType(RegBase), Base_FrameIndex(0), Scale(1), IndexReg(), Disp(0), in X86ISelAddressMode() 121 dbgs() << " Scale " << Scale << '\n' in dump() 221 SDValue &Scale, SDValue &Index, SDValue &Disp, 224 SDValue ScaleOp, SDValue &Base, SDValue &Scale, 228 SDValue &Scale, SDValue &Index, SDValue &Disp, 231 SDValue &Scale, SDValue &Index, SDValue &Disp, 234 SDValue &Scale, SDValue &Index, SDValue &Disp, 239 SDValue &Base, SDValue &Scale, 245 SDValue &Base, SDValue &Scale, in tryFoldLoad() argument [all …]
|
| H A D | X86FixupLEAs.cpp | 373 const MachineOperand &Scale = MI.getOperand(1 + X86::AddrScaleAmt); in optTwoAddrLEA() local 378 if (Segment.getReg() != 0 || !Disp.isImm() || Scale.getImm() > 1 || in optTwoAddrLEA() 505 const MachineOperand &Scale = MI.getOperand(1 + X86::AddrScaleAmt); in processInstructionForSlowLEA() local 519 if (Scale.getImm() > 1) in processInstructionForSlowLEA() 557 const MachineOperand &Scale = MI.getOperand(1 + X86::AddrScaleAmt); in processInstrForSlow3OpLEA() local 579 bool IsScale1 = Scale.getImm() == 1; in processInstrForSlow3OpLEA() 622 .add(Scale) in processInstrForSlow3OpLEA() 687 .add(Scale) in processInstrForSlow3OpLEA()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Analysis/ |
| H A D | BasicAliasAnalysis.cpp | 284 APInt Scale; member 287 LinearExpression(const ExtendedValue &Val, const APInt &Scale, in LinearExpression() 289 : Val(Val), Scale(Scale), Offset(Offset) {} in LinearExpression() 293 Scale = APInt(BitWidth, 1); in LinearExpression() 354 E.Scale *= RHS; in GetLinearExpression() 369 E.Scale <<= RHS.getLimitedValue(); in GetLinearExpression() 533 APInt Scale(MaxPointerSize, in DecomposeGEPExpression() local 554 .smul_ov(Scale, Overflow); in DecomposeGEPExpression() 559 Scale *= LE.Scale.sextOrTrunc(MaxPointerSize); in DecomposeGEPExpression() 570 Scale += Decomposed.VarIndices[i].Scale; in DecomposeGEPExpression() [all …]
|
| H A D | VectorUtils.cpp | 411 void llvm::narrowShuffleMaskElts(int Scale, ArrayRef<int> Mask, in narrowShuffleMaskElts() argument 413 assert(Scale > 0 && "Unexpected scaling factor"); in narrowShuffleMaskElts() 416 if (Scale == 1) { in narrowShuffleMaskElts() 424 assert(((uint64_t)Scale * MaskElt + (Scale - 1)) <= INT32_MAX && in narrowShuffleMaskElts() 427 for (int SliceElt = 0; SliceElt != Scale; ++SliceElt) in narrowShuffleMaskElts() 428 ScaledMask.push_back(MaskElt < 0 ? MaskElt : Scale * MaskElt + SliceElt); in narrowShuffleMaskElts() 432 bool llvm::widenShuffleMaskElts(int Scale, ArrayRef<int> Mask, in widenShuffleMaskElts() argument 434 assert(Scale > 0 && "Unexpected scaling factor"); in widenShuffleMaskElts() 437 if (Scale == 1) { in widenShuffleMaskElts() 444 if (NumElts % Scale != 0) in widenShuffleMaskElts() [all …]
|
| H A D | BlockFrequencyInfoImpl.cpp | 398 Loop.Scale = in computeLoopScale() 404 << " - scale = " << Loop.Scale << "\n"); in computeLoopScale() 509 << ": mass = " << Loop.Mass << ", scale = " << Loop.Scale in unwrapLoop() 511 Loop.Scale *= Loop.Mass.toScaled(); in unwrapLoop() 513 LLVM_DEBUG(dbgs() << " => combined-scale = " << Loop.Scale << "\n"); in unwrapLoop() 520 Scaled64 &F = Working.isAPackage() ? Working.getPackagedLoop()->Scale in unwrapLoop() 522 Scaled64 New = Loop.Scale * F; in unwrapLoop()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/ADT/ |
| H A D | APFixedPoint.h | 35 FixedPointSemantics(unsigned Width, unsigned Scale, bool IsSigned, in FixedPointSemantics() argument 37 : Width(Width), Scale(Scale), IsSigned(IsSigned), in FixedPointSemantics() 39 assert(Width >= Scale && "Not enough room for the scale"); in FixedPointSemantics() 45 unsigned getScale() const { return Scale; } in getScale() 57 return Width - Scale - 1; in getIntegralBits() 59 return Width - Scale; in getIntegralBits() 88 unsigned Scale : 13; variable
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/Mips/ |
| H A D | MipsConstantIslandPass.cpp | 135 unsigned Bits, Scale; in branchMaxOffsets() local 139 Scale = 2; in branchMaxOffsets() 143 Scale = 2; in branchMaxOffsets() 147 Scale = 2; in branchMaxOffsets() 151 Scale = 2; in branchMaxOffsets() 155 Scale = 2; in branchMaxOffsets() 159 Scale = 2; in branchMaxOffsets() 163 Scale = 2; in branchMaxOffsets() 167 Scale = 2; in branchMaxOffsets() 171 Scale = 2; in branchMaxOffsets() [all …]
|
| /netbsd-src/external/apache2/llvm/dist/clang/utils/TableGen/ |
| H A D | RISCVVEmitter.cpp | 67 VScaleVal Scale = 0; member in __anond814f6670111::RVVType 101 bool isScalar() const { return Scale.hasValue() && Scale.getValue() == 0; } in isScalar() 102 bool isVector() const { return Scale.hasValue() && Scale.getValue() != 0; } in isVector() 339 if (!Scale.hasValue()) in verifyType() 343 unsigned V = Scale.getValue(); in verifyType() 438 BuiltinStr = "q" + utostr(Scale.getValue()) + BuiltinStr; in initBuiltinStr() 448 ClangBuiltinStr += "bool" + utostr(64 / Scale.getValue()) + "_t"; in initClangBuiltinStr() 502 Str += "vbool" + utostr(64 / Scale.getValue()) + "_t"; in initTypeStr() 532 ShortStr = "b" + utostr(64 / Scale.getValue()); in initShortStr() 593 Scale = 0; in applyModifier() [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/MC/MCParser/ |
| H A D | MCTargetAsmParser.h | 69 unsigned Scale; member 73 OffsetName(StringRef()), Scale(1) {} in IntelExpr() 78 OffsetName(offsetName), Scale(1) { in IntelExpr() 80 Scale = scale; in IntelExpr() 90 return (Scale == 1) || in isValid() 91 (hasIndexReg() && (Scale == 2 || Scale == 4 || Scale == 8)); in isValid()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/X86/AsmParser/ |
| H A D | X86Operand.h | 63 unsigned Scale; member 139 if (Mem.Scale) in print() 140 OS << ",Scale=" << Mem.Scale; in print() 196 return Mem.Scale; in getMemScale() 673 Res->Mem.Scale = 1; 686 unsigned BaseReg, unsigned IndexReg, unsigned Scale, SMLoc StartLoc, 697 assert(((Scale == 1 || Scale == 2 || Scale == 4 || Scale == 8)) && 705 Res->Mem.Scale = Scale;
|
| H A D | X86AsmParser.cpp | 49 static bool checkScale(unsigned Scale, StringRef &ErrMsg) { in checkScale() argument 50 if (Scale != 1 && Scale != 2 && Scale != 4 && Scale != 8) { in checkScale() 426 unsigned BaseReg, IndexReg, TmpReg, Scale; member in __anon301d74170111::X86AsmParser::IntelExprStateMachine 451 TmpReg(0), Scale(0), Imm(0), Sym(nullptr), BracCount(0), in IntelExprStateMachine() 461 unsigned getScale() const { return Scale; } in getScale() 663 Scale = 0; in onPlus() 724 Scale = 0; in onMinus() 787 Scale = IC.popOperand(); in onRegister() 788 if (checkScale(Scale, ErrMsg)) in onRegister() 869 Scale = TmpInt; in onInteger() [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Transforms/Scalar/ |
| H A D | LoopStrengthReduce.cpp | 350 int64_t Scale = 0; member 485 if (Scale != 1) in isCanonical() 488 if (Scale == 1 && BaseRegs.empty()) in isCanonical() 521 Scale = 1; in canonicalize() 543 if (Scale != 1) in unscale() 545 Scale = 0; in unscale() 621 if (Scale != 0) { in print() 623 OS << Scale << "*reg("; in print() 1216 bool HasBaseReg, int64_t Scale, 1359 NumBaseParts - (1 + (F.Scale && isAMCompletelyFolded(*TTI, LU, F))); in RateFormula() [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Support/ |
| H A D | BranchProbability.cpp | 56 int Scale = 0; in getBranchProbability() local 59 Scale++; in getBranchProbability() 61 return BranchProbability(Numerator >> Scale, Denominator); in getBranchProbability()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/ARM/ |
| H A D | ARMBaseRegisterInfo.cpp | 516 int Scale = 1; in getFrameIndexInstrOffset() local 523 Scale = 1; in getFrameIndexInstrOffset() 531 Scale = 4; in getFrameIndexInstrOffset() 549 Scale = 4; in getFrameIndexInstrOffset() 555 return InstrOffs * Scale; in getFrameIndexInstrOffset() 714 unsigned Scale = 1; in isFrameOffsetLegal() local 721 Scale = 1; in isFrameOffsetLegal() 732 Scale = 4; in isFrameOffsetLegal() 743 Scale = 4; in isFrameOffsetLegal() 753 if ((Offset & (Scale-1)) != 0) in isFrameOffsetLegal() [all …]
|
| H A D | Thumb2InstrInfo.cpp | 624 unsigned Scale = 1; in rewriteT2FrameIndex() local 646 Scale = 4; in rewriteT2FrameIndex() 648 assert((Offset & (Scale-1)) == 0 && "Can't encode this offset!"); in rewriteT2FrameIndex() 660 Scale = 2; in rewriteT2FrameIndex() 662 assert((Offset & (Scale-1)) == 0 && "Can't encode this offset!"); in rewriteT2FrameIndex() 678 Scale = 1; in rewriteT2FrameIndex() 685 Scale = 1; in rewriteT2FrameIndex() 690 Scale = 4; in rewriteT2FrameIndex() 705 int ImmedOffset = Offset / Scale; in rewriteT2FrameIndex() 707 if ((unsigned)Offset <= Mask * Scale && in rewriteT2FrameIndex() [all …]
|
| H A D | ARMConstantIslandPass.cpp | 746 unsigned Scale = 1; in initializeFunctionInfo() local 761 Scale = 4; in initializeFunctionInfo() 767 Scale = 2; in initializeFunctionInfo() 771 Scale = 2; in initializeFunctionInfo() 777 Scale = 2; in initializeFunctionInfo() 781 Scale = 2; in initializeFunctionInfo() 786 unsigned MaxOffs = ((1 << (Bits-1))-1) * Scale; in initializeFunctionInfo() 806 unsigned Scale = 1; in initializeFunctionInfo() local 830 Scale = 4; in initializeFunctionInfo() 834 Scale = 1; in initializeFunctionInfo() [all …]
|
| H A D | ThumbRegisterInfo.cpp | 387 unsigned Scale = 4; in rewriteFrameIndex() local 389 Offset += InstrOffs * Scale; in rewriteFrameIndex() 390 assert((Offset & (Scale - 1)) == 0 && "Can't encode this offset!"); in rewriteFrameIndex() 394 int ImmedOffset = Offset / Scale; in rewriteFrameIndex() 397 if ((unsigned)Offset <= Mask * Scale) { in rewriteFrameIndex() 422 Offset &= ~(Mask * Scale); in rewriteFrameIndex()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/AArch64/MCTargetDesc/ |
| H A D | AArch64InstPrinter.h | 106 void printUImm12Offset(const MCInst *MI, unsigned OpNum, unsigned Scale, 108 void printAMIndexedWB(const MCInst *MI, unsigned OpNum, unsigned Scale, 111 template <int Scale> 114 printUImm12Offset(MI, OpNum, Scale, O); in printUImm12Offset() 126 template <int Scale>
|
| /netbsd-src/external/apache2/llvm/dist/llvm/lib/Target/AArch64/ |
| H A D | AArch64RegisterInfo.td | 1075 class ZPRExtendAsmOperand<string ShiftExtend, int RegWidth, int Scale, 1077 let Name = "ZPRExtend" # ShiftExtend # RegWidth # Scale 1083 # Scale # ", " 1086 let DiagnosticType = "InvalidZPR" # RegWidth # ShiftExtend # Scale; 1092 int RegWidth, int Scale, string Suffix = ""> 1095 !cast<AsmOperandClass>("ZPR" # RegWidth # "AsmOpndExt" # Repr # Scale # Suffix); 1098 # Scale # ", " 1141 class GPR64ShiftExtendAsmOperand <string AsmOperandName, int Scale, string RegClass> : AsmOperandCl… 1142 let Name = AsmOperandName # Scale; 1143 let PredicateMethod = "isGPR64WithShiftExtend<AArch64::"#RegClass#"RegClassID, " # Scale # ">"; [all …]
|
| H A D | AArch64InstrInfo.cpp | 2568 TypeSize Scale(0U, false); in getMemOperandWithOffsetWidth() local 2572 if (!getMemOpInfo(LdSt.getOpcode(), Scale, Width, Dummy1, Dummy2)) in getMemOperandWithOffsetWidth() 2580 Offset = LdSt.getOperand(2).getImm() * Scale.getKnownMinSize(); in getMemOperandWithOffsetWidth() 2584 Offset = LdSt.getOperand(3).getImm() * Scale.getKnownMinSize(); in getMemOperandWithOffsetWidth() 2586 OffsetIsScalable = Scale.isScalable(); in getMemOperandWithOffsetWidth() 2602 bool AArch64InstrInfo::getMemOpInfo(unsigned Opcode, TypeSize &Scale, in getMemOpInfo() argument 2609 Scale = TypeSize::Fixed(0); in getMemOpInfo() 2616 Scale = TypeSize::Fixed(4); in getMemOpInfo() 2623 Scale = TypeSize::Fixed(1); in getMemOpInfo() 2633 Scale = TypeSize::Fixed(1); in getMemOpInfo() [all …]
|
| /netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/Transforms/ |
| H A D | Instrumentation.h | 190 static inline uint32_t scaleBranchCount(uint64_t Count, uint64_t Scale) { in scaleBranchCount() argument 191 uint64_t Scaled = Count / Scale; in scaleBranchCount()
|
| /netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/Analysis/Utils/ |
| H A D | Local.h | 68 Constant *Scale = ConstantInt::get(IntIdxTy, Size); 72 ConstantExpr::getMul(OC, Scale, false /*NUW*/, isInBounds /*NSW*/);
|
| /netbsd-src/external/apache2/llvm/dist/llvm/include/llvm/Analysis/ |
| H A D | TargetTransformInfoImpl.h | 200 bool HasBaseReg, int64_t Scale, unsigned AddrSpace, 204 return !BaseGV && BaseOffset == 0 && (Scale == 0 || Scale == 1); 275 int64_t Scale, in getScalingFactorCost() argument 278 if (isLegalAddressingMode(Ty, BaseGV, BaseOffset, HasBaseReg, Scale, in getScalingFactorCost() 869 int64_t Scale = 0; variable 904 if (Scale != 0) 907 Scale = ElementSize; 914 BaseOffset.sextOrTrunc(64).getSExtValue(), HasBaseReg, Scale,
|