Lines Matching defs:StackProbeSize
659 const uint64_t StackProbeSize = TLI.getStackProbeSize(MF);
660 uint64_t ProbeChunk = StackProbeSize * 8;
666 // BuildStackAlignAND ensures that only MaxAlign % StackProbeSize bits left
670 MaxAlign % StackProbeSize);
673 MaxAlign % StackProbeSize);
687 const uint64_t StackProbeSize = TLI.getStackProbeSize(MF);
691 assert(AlignOffset < StackProbeSize);
694 if (StackProbeSize < Offset + AlignOffset) {
696 uint64_t StackAdjustment = StackProbeSize - AlignOffset;
711 CurrentOffset = StackProbeSize - AlignOffset;
717 while (CurrentOffset + StackProbeSize < Offset) {
718 BuildStackAdjustment(MBB, MBBI, DL, -StackProbeSize, /*InEpilogue=*/false)
724 MCCFIInstruction::createAdjustCfaOffset(nullptr, StackProbeSize));
732 CurrentOffset += StackProbeSize;
768 const uint64_t StackProbeSize = TLI.getStackProbeSize(MF);
771 if (AlignOffset < StackProbeSize) {
807 const unsigned BoundOffset = alignDown(Offset, StackProbeSize);
833 BuildStackAdjustment(*testMBB, testMBB->end(), DL, -StackProbeSize,
864 const uint64_t TailOffset = Offset % StackProbeSize;
1254 const uint64_t StackProbeSize = TLI.getStackProbeSize(MF);
1257 // We want to make sure that (in worst case) less than StackProbeSize bytes
1260 if (Reg == StackPtr && EmitInlineStackProbe && MaxAlign >= StackProbeSize) {
1317 .addImm(StackProbeSize)
1347 .addImm(StackProbeSize)
1606 unsigned StackProbeSize = STI.getTargetLowering()->getStackProbeSize(MF);
1958 if (AlignedNumBytes >= StackProbeSize && EmitStackProbeCall) {