Lines Matching defs:H

67 static inline void unmap(LargeBlock::Header *H) {
68 // Note that the `H->MapMap` is stored on the pages managed by itself. Take
71 MemMapT MemMap = H->MemMap;
93 UNUSED uptr HeadersSize, UNUSED LargeBlock::Header **H,
97 void store(UNUSED Options Options, LargeBlock::Header *H) { unmap(H); }
218 void store(const Options &Options, LargeBlock::Header *H) EXCLUDES(Mutex) {
219 if (!canCache(H->CommitSize))
220 return unmap(H);
228 Entry.CommitBase = H->CommitBase;
229 Entry.CommitSize = H->CommitSize;
230 Entry.BlockBegin = reinterpret_cast<uptr>(H + 1);
231 Entry.MemMap = H->MemMap;
299 LargeBlock::Header **H, bool *Zeroed) EXCLUDES(Mutex) {
358 *H = reinterpret_cast<LargeBlock::Header *>(
363 uptr NewBlockBegin = reinterpret_cast<uptr>(*H + 1);
374 (*H)->CommitBase = Entry.CommitBase;
375 (*H)->CommitSize = Entry.CommitSize;
376 (*H)->MemMap = Entry.MemMap;
541 for (const auto &H : InUseBlocks) {
542 uptr Ptr = reinterpret_cast<uptr>(&H) + LargeBlock::getHeaderSize();
600 LargeBlock::Header *H;
602 if (Cache.retrieve(Options, Size, Alignment, getHeadersSize(), &H,
604 const uptr BlockEnd = H->CommitBase + H->CommitSize;
607 uptr HInt = reinterpret_cast<uptr>(H);
617 InUseBlocks.push_back(H);
618 AllocatedBytes += H->CommitSize;
619 FragmentedBytes += H->MemMap.getCapacity() - H->CommitSize;
621 Stats.add(StatAllocated, H->CommitSize);
622 Stats.add(StatMapped, H->MemMap.getCapacity());
679 LargeBlock::Header *H = reinterpret_cast<LargeBlock::Header *>(
683 reinterpret_cast<uptr>(H + 1));
684 H->CommitBase = CommitBase;
685 H->CommitSize = CommitSize;
686 H->MemMap = MemMap;
691 InUseBlocks.push_back(H);
693 FragmentedBytes += H->MemMap.getCapacity() - CommitSize;
698 Stats.add(StatMapped, H->MemMap.getCapacity());
706 LargeBlock::Header *H = LargeBlock::getHeader<Config>(Ptr);
707 const uptr CommitSize = H->CommitSize;
710 InUseBlocks.remove(H);
712 FragmentedBytes -= H->MemMap.getCapacity() - CommitSize;
715 Stats.sub(StatMapped, H->MemMap.getCapacity());
717 Cache.store(Options, H);