Lines Matching defs:StackProbeSize
669 const uint64_t StackProbeSize = TLI.getStackProbeSize(MF);
670 uint64_t ProbeChunk = StackProbeSize * 8;
676 // BuildStackAlignAND ensures that only MaxAlign % StackProbeSize bits left
680 MaxAlign % StackProbeSize);
683 MaxAlign % StackProbeSize);
697 const uint64_t StackProbeSize = TLI.getStackProbeSize(MF);
701 assert(AlignOffset < StackProbeSize);
704 if (StackProbeSize < Offset + AlignOffset) {
706 uint64_t StackAdjustment = StackProbeSize - AlignOffset;
721 CurrentOffset = StackProbeSize - AlignOffset;
727 while (CurrentOffset + StackProbeSize < Offset) {
728 BuildStackAdjustment(MBB, MBBI, DL, -StackProbeSize, /*InEpilogue=*/false)
734 MCCFIInstruction::createAdjustCfaOffset(nullptr, StackProbeSize));
742 CurrentOffset += StackProbeSize;
778 const uint64_t StackProbeSize = TLI.getStackProbeSize(MF);
781 if (AlignOffset < StackProbeSize) {
813 const uint64_t BoundOffset = alignDown(Offset, StackProbeSize);
862 BuildStackAdjustment(*testMBB, testMBB->end(), DL, -StackProbeSize,
893 const uint64_t TailOffset = Offset % StackProbeSize;
1290 const uint64_t StackProbeSize = TLI.getStackProbeSize(MF);
1293 // We want to make sure that (in worst case) less than StackProbeSize bytes
1296 if (Reg == StackPtr && EmitInlineStackProbe && MaxAlign >= StackProbeSize) {
1353 .addImm(StackProbeSize)
1383 .addImm(StackProbeSize)
1642 unsigned StackProbeSize = STI.getTargetLowering()->getStackProbeSize(MF);
1994 if (AlignedNumBytes >= StackProbeSize && EmitStackProbeCall) {