| /openbsd-src/gnu/llvm/llvm/lib/Transforms/AggressiveInstCombine/ |
| H A D | AggressiveInstCombine.cpp | 674 APInt Offset2(DL.getIndexTypeSizeInBits(Load2Ptr->getType()), 0); in foldLoadsRecursive() local 676 Load2Ptr->stripAndAccumulateConstantOffsets(DL, Offset2, in foldLoadsRecursive() 710 if (Offset2.slt(Offset1)) { in foldLoadsRecursive() 713 std::swap(Offset1, Offset2); in foldLoadsRecursive() 745 if ((Shift2 - Shift1) != ShiftDiff || (Offset2 - Offset1) != PrevSize) in foldLoadsRecursive()
|
| /openbsd-src/gnu/llvm/llvm/lib/CodeGen/SelectionDAG/ |
| H A D | ScheduleDAGSDNodes.cpp | 246 int64_t Offset1, Offset2; in ClusterNeighboringLoads() local 247 if (!TII->areLoadsFromSameBasePtr(Base, User, Offset1, Offset2) || in ClusterNeighboringLoads() 248 Offset1 == Offset2 || in ClusterNeighboringLoads() 256 O2SMap.insert(std::make_pair(Offset2, User)); in ClusterNeighboringLoads() 257 Offsets.push_back(Offset2); in ClusterNeighboringLoads() 258 if (Offset2 < Offset1) in ClusterNeighboringLoads()
|
| /openbsd-src/gnu/llvm/llvm/lib/Target/Mips/ |
| H A D | MicroMipsSizeReduction.cpp | 400 int64_t Offset1, Offset2; in ConsecutiveInstr() local 403 if (!GetImm(MI2, 2, Offset2)) in ConsecutiveInstr() 409 return ((Offset1 == (Offset2 - 4)) && (ConsecutiveRegisters(Reg1, Reg2))); in ConsecutiveInstr()
|
| /openbsd-src/gnu/llvm/llvm/lib/Target/X86/ |
| H A D | X86InstrInfo.h | 437 int64_t &Offset2) const override; 455 int64_t Offset2,
|
| H A D | X86InstrInfo.cpp | 7191 int64_t &Offset1, int64_t &Offset2) const { in areLoadsFromSameBasePtr() 7384 Offset2 = Disp2->getSExtValue(); in areLoadsFromSameBasePtr() 7389 int64_t Offset1, int64_t Offset2, in shouldScheduleLoadsNear() argument 7391 assert(Offset2 > Offset1); in shouldScheduleLoadsNear() 7392 if ((Offset2 - Offset1) / 8 > 64) in shouldScheduleLoadsNear()
|
| /openbsd-src/gnu/llvm/clang/lib/StaticAnalyzer/Checkers/ |
| H A D | ContainerModeling.cpp | 127 SymbolRef Offset2, 961 SymbolRef Offset2, in invalidateIteratorPositions() argument 965 compare(State, Pos.getOffset(), Offset2, Opc2); in invalidateIteratorPositions()
|
| /openbsd-src/gnu/llvm/llvm/lib/Target/ARM/ |
| H A D | ARMBaseInstrInfo.h | 248 int64_t &Offset2) const override; 259 int64_t Offset1, int64_t Offset2,
|
| H A D | ARMBaseInstrInfo.cpp | 1949 int64_t &Offset2) const { in areLoadsFromSameBasePtr() 2010 Offset2 = cast<ConstantSDNode>(Load2->getOperand(1))->getSExtValue(); in areLoadsFromSameBasePtr() 2029 int64_t Offset1, int64_t Offset2, in shouldScheduleLoadsNear() argument 2034 assert(Offset2 > Offset1); in shouldScheduleLoadsNear() 2036 if ((Offset2 - Offset1) / 8 > 64) in shouldScheduleLoadsNear()
|
| /openbsd-src/gnu/llvm/llvm/include/llvm/CodeGen/ |
| H A D | TargetInstrInfo.h | 1362 int64_t &Offset2) const { in areLoadsFromSameBasePtr() argument 1375 int64_t Offset1, int64_t Offset2, in shouldScheduleLoadsNear() argument
|
| /openbsd-src/gnu/llvm/llvm/lib/Transforms/Scalar/ |
| H A D | SeparateConstOffsetFromGEP.cpp | 1359 Value *Offset2 = Second->getOperand(1); in swapGEPOperand() local 1360 First->setOperand(1, Offset2); in swapGEPOperand()
|
| H A D | ConstraintElimination.cpp | 433 int64_t Offset2 = BDec.Offset; in getConstraint() local 483 if (AddOverflow(Offset1, Offset2, OffsetSum)) in getConstraint()
|
| /openbsd-src/gnu/llvm/llvm/lib/Target/SystemZ/ |
| H A D | SystemZInstrInfo.cpp | 1638 int64_t Offset2 = (MCID.TSFlags & SystemZII::Is128Bit ? Offset + 8 : Offset); in getOpcodeForOffset() local 1639 if (isUInt<12>(Offset) && isUInt<12>(Offset2)) { in getOpcodeForOffset() 1649 if (isInt<20>(Offset) && isInt<20>(Offset2)) { in getOpcodeForOffset()
|
| /openbsd-src/gnu/llvm/llvm/lib/Target/AArch64/ |
| H A D | AArch64InstrInfo.cpp | 3312 int64_t Offset2, unsigned Opcode2) { in shouldClusterFI() argument 3329 ObjectOffset2 += Offset2; in shouldClusterFI() 3382 int64_t Offset2 = SecondLdSt.getOperand(2).getImm(); in shouldClusterMemOps() local 3383 if (hasUnscaledLdStOffset(SecondOpc) && !scaleOffset(SecondOpc, Offset2)) in shouldClusterMemOps() 3393 assert((!BaseOp1.isIdenticalTo(BaseOp2) || Offset1 <= Offset2) && in shouldClusterMemOps() 3399 BaseOp2.getIndex(), Offset2, SecondOpc); in shouldClusterMemOps() 3402 assert(Offset1 <= Offset2 && "Caller should have ordered offsets."); in shouldClusterMemOps() 3404 return Offset1 + 1 == Offset2; in shouldClusterMemOps()
|
| H A D | AArch64InstrFormats.td | 10347 int Offset1, int Offset2, int Offset4, int Offset8> { 10373 !cast<DAGOperand>("GPR64pi" # Offset2)>; 10376 !cast<DAGOperand>("GPR64pi" # Offset2)>; 10392 defm : SIMDLdrAliases<NAME, asm, "4h", Count, Offset2, 64>; 10393 defm : SIMDLdrAliases<NAME, asm, "8h", Count, Offset2, 128>;
|
| /openbsd-src/gnu/llvm/llvm/lib/CodeGen/ |
| H A D | MachinePipeliner.cpp | 795 int64_t Offset1, Offset2; in addLoopCarriedDependences() local 799 TII->getMemOperandWithOffset(MI, BaseOp2, Offset2, in addLoopCarriedDependences() 803 (int)Offset1 < (int)Offset2) { in addLoopCarriedDependences()
|
| H A D | CodeGenPrepare.cpp | 2239 uint64_t Offset2 = Offset.getLimitedValue(); in optimizeCallInst() local 2240 if (!isAligned(PrefAlign, Offset2)) in optimizeCallInst() 2244 DL->getTypeAllocSize(AI->getAllocatedType()) >= MinSize + Offset2) in optimizeCallInst() 2253 DL->getTypeAllocSize(GV->getValueType()) >= MinSize + Offset2) in optimizeCallInst()
|
| /openbsd-src/gnu/llvm/llvm/lib/Transforms/InstCombine/ |
| H A D | InstCombineAndOrXor.cpp | 1153 const APInt *Offset1 = nullptr, *Offset2 = nullptr; in foldAndOrOfICmpsUsingRanges() local 1158 if (match(V2, m_Add(m_Value(X), m_APInt(Offset2)))) in foldAndOrOfICmpsUsingRanges() 1172 if (Offset2) in foldAndOrOfICmpsUsingRanges() 1173 CR2 = CR2.subtract(*Offset2); in foldAndOrOfICmpsUsingRanges()
|
| /openbsd-src/gnu/llvm/clang/tools/c-index-test/ |
| H A D | c-index-test.c | 1778 long long Offset2 = clang_Cursor_getOffsetOfField(cursor); in PrintTypeSize() local 1779 if (Offset == Offset2){ in PrintTypeSize() 1783 printf(" [offsetof=%lld/%lld]", Offset, Offset2); in PrintTypeSize()
|
| /openbsd-src/gnu/llvm/llvm/lib/Target/PowerPC/ |
| H A D | PPCInstrInfo.cpp | 2941 int64_t Offset1 = 0, Offset2 = 0; in shouldClusterMemOps() local 2945 !getMemOperandWithOffsetWidth(SecondLdSt, Base2, Offset2, Width2, TRI) || in shouldClusterMemOps() 2952 assert(Offset1 <= Offset2 && "Caller should have ordered offsets."); in shouldClusterMemOps() 2953 return Offset1 + Width1 == Offset2; in shouldClusterMemOps()
|
| H A D | PPCISelLowering.cpp | 13317 int64_t Offset1 = 0, Offset2 = 0; in isConsecutiveLSLoc() local 13319 getBaseWithConstantOffset(BaseLoc, Base2, Offset2, DAG); in isConsecutiveLSLoc() 13320 if (Base1 == Base2 && Offset1 == (Offset2 + Dist * Bytes)) in isConsecutiveLSLoc() 13327 Offset2 = 0; in isConsecutiveLSLoc() 13329 bool isGA2 = TLI.isGAPlusOffset(BaseLoc.getNode(), GV2, Offset2); in isConsecutiveLSLoc() 13331 return Offset1 == (Offset2 + Dist*Bytes); in isConsecutiveLSLoc()
|
| /openbsd-src/gnu/llvm/llvm/lib/Analysis/ |
| H A D | ValueTracking.cpp | 7503 APInt Offset2(DL.getIndexTypeSizeInBits(Ptr2->getType()), 0); in isPointerOffset() local 7505 Ptr2 = Ptr2->stripAndAccumulateConstantOffsets(DL, Offset2, true); in isPointerOffset() 7509 return Offset2.getSExtValue() - Offset1.getSExtValue(); in isPointerOffset() 7533 return *IOffset2 - *IOffset1 + Offset2.getSExtValue() - in isPointerOffset()
|
| /openbsd-src/gnu/llvm/llvm/lib/Transforms/Vectorize/ |
| H A D | SLPVectorizer.cpp | 4776 int Offset2 = Pair2.second; in canFormVector() local 4777 return Offset1 < Offset2; in canFormVector()
|
| /openbsd-src/gnu/llvm/llvm/docs/ |
| H A D | LangRef.rst | 6183 Offset2)`` if either ``(BaseTy1, Offset1)`` is reachable from ``(Base2, 6184 Offset2)`` via the ``Parent`` relation or vice versa.
|