1349cc55cSDimitry Andric //===- LoadStoreOpt.cpp ----------- Generic memory optimizations -*- C++ -*-==// 2349cc55cSDimitry Andric // 3349cc55cSDimitry Andric // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. 4349cc55cSDimitry Andric // See https://llvm.org/LICENSE.txt for license information. 5349cc55cSDimitry Andric // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception 6349cc55cSDimitry Andric // 7349cc55cSDimitry Andric //===----------------------------------------------------------------------===// 8349cc55cSDimitry Andric /// \file 9349cc55cSDimitry Andric /// This file implements the LoadStoreOpt optimization pass. 10349cc55cSDimitry Andric //===----------------------------------------------------------------------===// 11349cc55cSDimitry Andric 12349cc55cSDimitry Andric #include "llvm/CodeGen/GlobalISel/LoadStoreOpt.h" 1306c3fb27SDimitry Andric #include "llvm/ADT/STLExtras.h" 1406c3fb27SDimitry Andric #include "llvm/ADT/SmallPtrSet.h" 15349cc55cSDimitry Andric #include "llvm/ADT/Statistic.h" 16349cc55cSDimitry Andric #include "llvm/Analysis/AliasAnalysis.h" 17349cc55cSDimitry Andric #include "llvm/Analysis/MemoryLocation.h" 18349cc55cSDimitry Andric #include "llvm/Analysis/OptimizationRemarkEmitter.h" 19349cc55cSDimitry Andric #include "llvm/CodeGen/GlobalISel/GenericMachineInstrs.h" 20349cc55cSDimitry Andric #include "llvm/CodeGen/GlobalISel/LegalizerInfo.h" 21349cc55cSDimitry Andric #include "llvm/CodeGen/GlobalISel/MIPatternMatch.h" 22349cc55cSDimitry Andric #include "llvm/CodeGen/GlobalISel/Utils.h" 2306c3fb27SDimitry Andric #include "llvm/CodeGen/LowLevelTypeUtils.h" 24349cc55cSDimitry Andric #include "llvm/CodeGen/MachineBasicBlock.h" 25349cc55cSDimitry Andric #include "llvm/CodeGen/MachineFrameInfo.h" 26349cc55cSDimitry Andric #include "llvm/CodeGen/MachineFunction.h" 27349cc55cSDimitry Andric #include "llvm/CodeGen/MachineInstr.h" 28349cc55cSDimitry Andric #include "llvm/CodeGen/MachineOptimizationRemarkEmitter.h" 29349cc55cSDimitry Andric #include "llvm/CodeGen/MachineRegisterInfo.h" 30349cc55cSDimitry Andric #include "llvm/CodeGen/Register.h" 31349cc55cSDimitry Andric #include "llvm/CodeGen/TargetLowering.h" 32349cc55cSDimitry Andric #include "llvm/CodeGen/TargetOpcodes.h" 33349cc55cSDimitry Andric #include "llvm/IR/DebugInfoMetadata.h" 34349cc55cSDimitry Andric #include "llvm/InitializePasses.h" 35349cc55cSDimitry Andric #include "llvm/Support/AtomicOrdering.h" 36349cc55cSDimitry Andric #include "llvm/Support/Casting.h" 37349cc55cSDimitry Andric #include "llvm/Support/Debug.h" 38349cc55cSDimitry Andric #include "llvm/Support/ErrorHandling.h" 39349cc55cSDimitry Andric #include "llvm/Support/MathExtras.h" 40349cc55cSDimitry Andric #include <algorithm> 41349cc55cSDimitry Andric 42349cc55cSDimitry Andric #define DEBUG_TYPE "loadstore-opt" 43349cc55cSDimitry Andric 44349cc55cSDimitry Andric using namespace llvm; 45349cc55cSDimitry Andric using namespace ore; 46349cc55cSDimitry Andric using namespace MIPatternMatch; 47349cc55cSDimitry Andric 48349cc55cSDimitry Andric STATISTIC(NumStoresMerged, "Number of stores merged"); 49349cc55cSDimitry Andric 50349cc55cSDimitry Andric const unsigned MaxStoreSizeToForm = 128; 51349cc55cSDimitry Andric 52349cc55cSDimitry Andric char LoadStoreOpt::ID = 0; 53349cc55cSDimitry Andric INITIALIZE_PASS_BEGIN(LoadStoreOpt, DEBUG_TYPE, "Generic memory optimizations", 54349cc55cSDimitry Andric false, false) 55349cc55cSDimitry Andric INITIALIZE_PASS_END(LoadStoreOpt, DEBUG_TYPE, "Generic memory optimizations", 56349cc55cSDimitry Andric false, false) 57349cc55cSDimitry Andric 58349cc55cSDimitry Andric LoadStoreOpt::LoadStoreOpt(std::function<bool(const MachineFunction &)> F) 59349cc55cSDimitry Andric : MachineFunctionPass(ID), DoNotRunPass(F) {} 60349cc55cSDimitry Andric 61349cc55cSDimitry Andric LoadStoreOpt::LoadStoreOpt() 62349cc55cSDimitry Andric : LoadStoreOpt([](const MachineFunction &) { return false; }) {} 63349cc55cSDimitry Andric 64349cc55cSDimitry Andric void LoadStoreOpt::init(MachineFunction &MF) { 65349cc55cSDimitry Andric this->MF = &MF; 66349cc55cSDimitry Andric MRI = &MF.getRegInfo(); 67349cc55cSDimitry Andric AA = &getAnalysis<AAResultsWrapperPass>().getAAResults(); 68349cc55cSDimitry Andric TLI = MF.getSubtarget().getTargetLowering(); 69349cc55cSDimitry Andric LI = MF.getSubtarget().getLegalizerInfo(); 70349cc55cSDimitry Andric Builder.setMF(MF); 71349cc55cSDimitry Andric IsPreLegalizer = !MF.getProperties().hasProperty( 72349cc55cSDimitry Andric MachineFunctionProperties::Property::Legalized); 73349cc55cSDimitry Andric InstsToErase.clear(); 74349cc55cSDimitry Andric } 75349cc55cSDimitry Andric 76349cc55cSDimitry Andric void LoadStoreOpt::getAnalysisUsage(AnalysisUsage &AU) const { 77349cc55cSDimitry Andric AU.addRequired<AAResultsWrapperPass>(); 7881ad6265SDimitry Andric AU.setPreservesAll(); 79349cc55cSDimitry Andric getSelectionDAGFallbackAnalysisUsage(AU); 80349cc55cSDimitry Andric MachineFunctionPass::getAnalysisUsage(AU); 81349cc55cSDimitry Andric } 82349cc55cSDimitry Andric 83349cc55cSDimitry Andric BaseIndexOffset GISelAddressing::getPointerInfo(Register Ptr, 84349cc55cSDimitry Andric MachineRegisterInfo &MRI) { 85349cc55cSDimitry Andric BaseIndexOffset Info; 86349cc55cSDimitry Andric Register PtrAddRHS; 875678d1d9SDimitry Andric Register BaseReg; 885678d1d9SDimitry Andric if (!mi_match(Ptr, MRI, m_GPtrAdd(m_Reg(BaseReg), m_Reg(PtrAddRHS)))) { 895678d1d9SDimitry Andric Info.setBase(Ptr); 905678d1d9SDimitry Andric Info.setOffset(0); 91349cc55cSDimitry Andric return Info; 92349cc55cSDimitry Andric } 935678d1d9SDimitry Andric Info.setBase(BaseReg); 94349cc55cSDimitry Andric auto RHSCst = getIConstantVRegValWithLookThrough(PtrAddRHS, MRI); 95349cc55cSDimitry Andric if (RHSCst) 965678d1d9SDimitry Andric Info.setOffset(RHSCst->Value.getSExtValue()); 97349cc55cSDimitry Andric 98349cc55cSDimitry Andric // Just recognize a simple case for now. In future we'll need to match 99349cc55cSDimitry Andric // indexing patterns for base + index + constant. 1005678d1d9SDimitry Andric Info.setIndex(PtrAddRHS); 101349cc55cSDimitry Andric return Info; 102349cc55cSDimitry Andric } 103349cc55cSDimitry Andric 104349cc55cSDimitry Andric bool GISelAddressing::aliasIsKnownForLoadStore(const MachineInstr &MI1, 105349cc55cSDimitry Andric const MachineInstr &MI2, 106349cc55cSDimitry Andric bool &IsAlias, 107349cc55cSDimitry Andric MachineRegisterInfo &MRI) { 108349cc55cSDimitry Andric auto *LdSt1 = dyn_cast<GLoadStore>(&MI1); 109349cc55cSDimitry Andric auto *LdSt2 = dyn_cast<GLoadStore>(&MI2); 110349cc55cSDimitry Andric if (!LdSt1 || !LdSt2) 111349cc55cSDimitry Andric return false; 112349cc55cSDimitry Andric 113349cc55cSDimitry Andric BaseIndexOffset BasePtr0 = getPointerInfo(LdSt1->getPointerReg(), MRI); 114349cc55cSDimitry Andric BaseIndexOffset BasePtr1 = getPointerInfo(LdSt2->getPointerReg(), MRI); 115349cc55cSDimitry Andric 1165678d1d9SDimitry Andric if (!BasePtr0.getBase().isValid() || !BasePtr1.getBase().isValid()) 117349cc55cSDimitry Andric return false; 118349cc55cSDimitry Andric 119*0fca6ea1SDimitry Andric LocationSize Size1 = LdSt1->getMemSize(); 120*0fca6ea1SDimitry Andric LocationSize Size2 = LdSt2->getMemSize(); 121349cc55cSDimitry Andric 122349cc55cSDimitry Andric int64_t PtrDiff; 1235678d1d9SDimitry Andric if (BasePtr0.getBase() == BasePtr1.getBase() && BasePtr0.hasValidOffset() && 1245678d1d9SDimitry Andric BasePtr1.hasValidOffset()) { 1255678d1d9SDimitry Andric PtrDiff = BasePtr1.getOffset() - BasePtr0.getOffset(); 126349cc55cSDimitry Andric // If the size of memory access is unknown, do not use it to do analysis. 127349cc55cSDimitry Andric // One example of unknown size memory access is to load/store scalable 128349cc55cSDimitry Andric // vector objects on the stack. 129349cc55cSDimitry Andric // BasePtr1 is PtrDiff away from BasePtr0. They alias if none of the 130349cc55cSDimitry Andric // following situations arise: 131*0fca6ea1SDimitry Andric if (PtrDiff >= 0 && Size1.hasValue() && !Size1.isScalable()) { 132349cc55cSDimitry Andric // [----BasePtr0----] 133349cc55cSDimitry Andric // [---BasePtr1--] 134349cc55cSDimitry Andric // ========PtrDiff========> 135*0fca6ea1SDimitry Andric IsAlias = !((int64_t)Size1.getValue() <= PtrDiff); 136349cc55cSDimitry Andric return true; 137349cc55cSDimitry Andric } 138*0fca6ea1SDimitry Andric if (PtrDiff < 0 && Size2.hasValue() && !Size2.isScalable()) { 139349cc55cSDimitry Andric // [----BasePtr0----] 140349cc55cSDimitry Andric // [---BasePtr1--] 141349cc55cSDimitry Andric // =====(-PtrDiff)====> 142*0fca6ea1SDimitry Andric IsAlias = !((PtrDiff + (int64_t)Size2.getValue()) <= 0); 143349cc55cSDimitry Andric return true; 144349cc55cSDimitry Andric } 145349cc55cSDimitry Andric return false; 146349cc55cSDimitry Andric } 147349cc55cSDimitry Andric 148349cc55cSDimitry Andric // If both BasePtr0 and BasePtr1 are FrameIndexes, we will not be 149349cc55cSDimitry Andric // able to calculate their relative offset if at least one arises 150349cc55cSDimitry Andric // from an alloca. However, these allocas cannot overlap and we 151349cc55cSDimitry Andric // can infer there is no alias. 1525678d1d9SDimitry Andric auto *Base0Def = getDefIgnoringCopies(BasePtr0.getBase(), MRI); 1535678d1d9SDimitry Andric auto *Base1Def = getDefIgnoringCopies(BasePtr1.getBase(), MRI); 154349cc55cSDimitry Andric if (!Base0Def || !Base1Def) 155349cc55cSDimitry Andric return false; // Couldn't tell anything. 156349cc55cSDimitry Andric 157349cc55cSDimitry Andric 158349cc55cSDimitry Andric if (Base0Def->getOpcode() != Base1Def->getOpcode()) 159349cc55cSDimitry Andric return false; 160349cc55cSDimitry Andric 161349cc55cSDimitry Andric if (Base0Def->getOpcode() == TargetOpcode::G_FRAME_INDEX) { 162349cc55cSDimitry Andric MachineFrameInfo &MFI = Base0Def->getMF()->getFrameInfo(); 163349cc55cSDimitry Andric // If the bases have the same frame index but we couldn't find a 164349cc55cSDimitry Andric // constant offset, (indices are different) be conservative. 165349cc55cSDimitry Andric if (Base0Def != Base1Def && 166349cc55cSDimitry Andric (!MFI.isFixedObjectIndex(Base0Def->getOperand(1).getIndex()) || 167349cc55cSDimitry Andric !MFI.isFixedObjectIndex(Base1Def->getOperand(1).getIndex()))) { 168349cc55cSDimitry Andric IsAlias = false; 169349cc55cSDimitry Andric return true; 170349cc55cSDimitry Andric } 171349cc55cSDimitry Andric } 172349cc55cSDimitry Andric 173349cc55cSDimitry Andric // This implementation is a lot more primitive than the SDAG one for now. 174349cc55cSDimitry Andric // FIXME: what about constant pools? 175349cc55cSDimitry Andric if (Base0Def->getOpcode() == TargetOpcode::G_GLOBAL_VALUE) { 176349cc55cSDimitry Andric auto GV0 = Base0Def->getOperand(1).getGlobal(); 177349cc55cSDimitry Andric auto GV1 = Base1Def->getOperand(1).getGlobal(); 178349cc55cSDimitry Andric if (GV0 != GV1) { 179349cc55cSDimitry Andric IsAlias = false; 180349cc55cSDimitry Andric return true; 181349cc55cSDimitry Andric } 182349cc55cSDimitry Andric } 183349cc55cSDimitry Andric 184349cc55cSDimitry Andric // Can't tell anything about aliasing. 185349cc55cSDimitry Andric return false; 186349cc55cSDimitry Andric } 187349cc55cSDimitry Andric 188349cc55cSDimitry Andric bool GISelAddressing::instMayAlias(const MachineInstr &MI, 189349cc55cSDimitry Andric const MachineInstr &Other, 190349cc55cSDimitry Andric MachineRegisterInfo &MRI, 191349cc55cSDimitry Andric AliasAnalysis *AA) { 192349cc55cSDimitry Andric struct MemUseCharacteristics { 193349cc55cSDimitry Andric bool IsVolatile; 194349cc55cSDimitry Andric bool IsAtomic; 195349cc55cSDimitry Andric Register BasePtr; 196349cc55cSDimitry Andric int64_t Offset; 197*0fca6ea1SDimitry Andric LocationSize NumBytes; 198349cc55cSDimitry Andric MachineMemOperand *MMO; 199349cc55cSDimitry Andric }; 200349cc55cSDimitry Andric 201349cc55cSDimitry Andric auto getCharacteristics = 202349cc55cSDimitry Andric [&](const MachineInstr *MI) -> MemUseCharacteristics { 203349cc55cSDimitry Andric if (const auto *LS = dyn_cast<GLoadStore>(MI)) { 204349cc55cSDimitry Andric Register BaseReg; 205349cc55cSDimitry Andric int64_t Offset = 0; 206349cc55cSDimitry Andric // No pre/post-inc addressing modes are considered here, unlike in SDAG. 207349cc55cSDimitry Andric if (!mi_match(LS->getPointerReg(), MRI, 208349cc55cSDimitry Andric m_GPtrAdd(m_Reg(BaseReg), m_ICst(Offset)))) { 209349cc55cSDimitry Andric BaseReg = LS->getPointerReg(); 210349cc55cSDimitry Andric Offset = 0; 211349cc55cSDimitry Andric } 212349cc55cSDimitry Andric 213*0fca6ea1SDimitry Andric LocationSize Size = LS->getMMO().getSize(); 214349cc55cSDimitry Andric return {LS->isVolatile(), LS->isAtomic(), BaseReg, 215349cc55cSDimitry Andric Offset /*base offset*/, Size, &LS->getMMO()}; 216349cc55cSDimitry Andric } 217349cc55cSDimitry Andric // FIXME: support recognizing lifetime instructions. 218349cc55cSDimitry Andric // Default. 219349cc55cSDimitry Andric return {false /*isvolatile*/, 220*0fca6ea1SDimitry Andric /*isAtomic*/ false, 221*0fca6ea1SDimitry Andric Register(), 222*0fca6ea1SDimitry Andric (int64_t)0 /*offset*/, 223*0fca6ea1SDimitry Andric LocationSize::beforeOrAfterPointer() /*size*/, 224349cc55cSDimitry Andric (MachineMemOperand *)nullptr}; 225349cc55cSDimitry Andric }; 226349cc55cSDimitry Andric MemUseCharacteristics MUC0 = getCharacteristics(&MI), 227349cc55cSDimitry Andric MUC1 = getCharacteristics(&Other); 228349cc55cSDimitry Andric 229349cc55cSDimitry Andric // If they are to the same address, then they must be aliases. 230349cc55cSDimitry Andric if (MUC0.BasePtr.isValid() && MUC0.BasePtr == MUC1.BasePtr && 231349cc55cSDimitry Andric MUC0.Offset == MUC1.Offset) 232349cc55cSDimitry Andric return true; 233349cc55cSDimitry Andric 234349cc55cSDimitry Andric // If they are both volatile then they cannot be reordered. 235349cc55cSDimitry Andric if (MUC0.IsVolatile && MUC1.IsVolatile) 236349cc55cSDimitry Andric return true; 237349cc55cSDimitry Andric 238349cc55cSDimitry Andric // Be conservative about atomics for the moment 239349cc55cSDimitry Andric // TODO: This is way overconservative for unordered atomics (see D66309) 240349cc55cSDimitry Andric if (MUC0.IsAtomic && MUC1.IsAtomic) 241349cc55cSDimitry Andric return true; 242349cc55cSDimitry Andric 243349cc55cSDimitry Andric // If one operation reads from invariant memory, and the other may store, they 244349cc55cSDimitry Andric // cannot alias. 245349cc55cSDimitry Andric if (MUC0.MMO && MUC1.MMO) { 246349cc55cSDimitry Andric if ((MUC0.MMO->isInvariant() && MUC1.MMO->isStore()) || 247349cc55cSDimitry Andric (MUC1.MMO->isInvariant() && MUC0.MMO->isStore())) 248349cc55cSDimitry Andric return false; 249349cc55cSDimitry Andric } 250349cc55cSDimitry Andric 251*0fca6ea1SDimitry Andric // If NumBytes is scalable and offset is not 0, conservatively return may 252*0fca6ea1SDimitry Andric // alias 253*0fca6ea1SDimitry Andric if ((MUC0.NumBytes.isScalable() && MUC0.Offset != 0) || 254*0fca6ea1SDimitry Andric (MUC1.NumBytes.isScalable() && MUC1.Offset != 0)) 255*0fca6ea1SDimitry Andric return true; 256*0fca6ea1SDimitry Andric 257*0fca6ea1SDimitry Andric const bool BothNotScalable = 258*0fca6ea1SDimitry Andric !MUC0.NumBytes.isScalable() && !MUC1.NumBytes.isScalable(); 259*0fca6ea1SDimitry Andric 260349cc55cSDimitry Andric // Try to prove that there is aliasing, or that there is no aliasing. Either 261349cc55cSDimitry Andric // way, we can return now. If nothing can be proved, proceed with more tests. 262349cc55cSDimitry Andric bool IsAlias; 263*0fca6ea1SDimitry Andric if (BothNotScalable && 264*0fca6ea1SDimitry Andric GISelAddressing::aliasIsKnownForLoadStore(MI, Other, IsAlias, MRI)) 265349cc55cSDimitry Andric return IsAlias; 266349cc55cSDimitry Andric 267349cc55cSDimitry Andric // The following all rely on MMO0 and MMO1 being valid. 268349cc55cSDimitry Andric if (!MUC0.MMO || !MUC1.MMO) 269349cc55cSDimitry Andric return true; 270349cc55cSDimitry Andric 271349cc55cSDimitry Andric // FIXME: port the alignment based alias analysis from SDAG's isAlias(). 272349cc55cSDimitry Andric int64_t SrcValOffset0 = MUC0.MMO->getOffset(); 273349cc55cSDimitry Andric int64_t SrcValOffset1 = MUC1.MMO->getOffset(); 274*0fca6ea1SDimitry Andric LocationSize Size0 = MUC0.NumBytes; 275*0fca6ea1SDimitry Andric LocationSize Size1 = MUC1.NumBytes; 276*0fca6ea1SDimitry Andric if (AA && MUC0.MMO->getValue() && MUC1.MMO->getValue() && Size0.hasValue() && 277*0fca6ea1SDimitry Andric Size1.hasValue()) { 278349cc55cSDimitry Andric // Use alias analysis information. 279349cc55cSDimitry Andric int64_t MinOffset = std::min(SrcValOffset0, SrcValOffset1); 280*0fca6ea1SDimitry Andric int64_t Overlap0 = 281*0fca6ea1SDimitry Andric Size0.getValue().getKnownMinValue() + SrcValOffset0 - MinOffset; 282*0fca6ea1SDimitry Andric int64_t Overlap1 = 283*0fca6ea1SDimitry Andric Size1.getValue().getKnownMinValue() + SrcValOffset1 - MinOffset; 284*0fca6ea1SDimitry Andric LocationSize Loc0 = 285*0fca6ea1SDimitry Andric Size0.isScalable() ? Size0 : LocationSize::precise(Overlap0); 286*0fca6ea1SDimitry Andric LocationSize Loc1 = 287*0fca6ea1SDimitry Andric Size1.isScalable() ? Size1 : LocationSize::precise(Overlap1); 288*0fca6ea1SDimitry Andric 289*0fca6ea1SDimitry Andric if (AA->isNoAlias( 290*0fca6ea1SDimitry Andric MemoryLocation(MUC0.MMO->getValue(), Loc0, MUC0.MMO->getAAInfo()), 291*0fca6ea1SDimitry Andric MemoryLocation(MUC1.MMO->getValue(), Loc1, MUC1.MMO->getAAInfo()))) 292349cc55cSDimitry Andric return false; 293349cc55cSDimitry Andric } 294349cc55cSDimitry Andric 295349cc55cSDimitry Andric // Otherwise we have to assume they alias. 296349cc55cSDimitry Andric return true; 297349cc55cSDimitry Andric } 298349cc55cSDimitry Andric 299349cc55cSDimitry Andric /// Returns true if the instruction creates an unavoidable hazard that 300349cc55cSDimitry Andric /// forces a boundary between store merge candidates. 301349cc55cSDimitry Andric static bool isInstHardMergeHazard(MachineInstr &MI) { 302349cc55cSDimitry Andric return MI.hasUnmodeledSideEffects() || MI.hasOrderedMemoryRef(); 303349cc55cSDimitry Andric } 304349cc55cSDimitry Andric 305349cc55cSDimitry Andric bool LoadStoreOpt::mergeStores(SmallVectorImpl<GStore *> &StoresToMerge) { 306349cc55cSDimitry Andric // Try to merge all the stores in the vector, splitting into separate segments 307349cc55cSDimitry Andric // as necessary. 308349cc55cSDimitry Andric assert(StoresToMerge.size() > 1 && "Expected multiple stores to merge"); 309349cc55cSDimitry Andric LLT OrigTy = MRI->getType(StoresToMerge[0]->getValueReg()); 310349cc55cSDimitry Andric LLT PtrTy = MRI->getType(StoresToMerge[0]->getPointerReg()); 311349cc55cSDimitry Andric unsigned AS = PtrTy.getAddressSpace(); 312349cc55cSDimitry Andric // Ensure the legal store info is computed for this address space. 313349cc55cSDimitry Andric initializeStoreMergeTargetInfo(AS); 314349cc55cSDimitry Andric const auto &LegalSizes = LegalStoreSizes[AS]; 315349cc55cSDimitry Andric 316349cc55cSDimitry Andric #ifndef NDEBUG 317fcaf7f86SDimitry Andric for (auto *StoreMI : StoresToMerge) 318349cc55cSDimitry Andric assert(MRI->getType(StoreMI->getValueReg()) == OrigTy); 319349cc55cSDimitry Andric #endif 320349cc55cSDimitry Andric 321*0fca6ea1SDimitry Andric const auto &DL = MF->getFunction().getDataLayout(); 322349cc55cSDimitry Andric bool AnyMerged = false; 323349cc55cSDimitry Andric do { 32406c3fb27SDimitry Andric unsigned NumPow2 = llvm::bit_floor(StoresToMerge.size()); 325bdd1243dSDimitry Andric unsigned MaxSizeBits = NumPow2 * OrigTy.getSizeInBits().getFixedValue(); 326349cc55cSDimitry Andric // Compute the biggest store we can generate to handle the number of stores. 327349cc55cSDimitry Andric unsigned MergeSizeBits; 328349cc55cSDimitry Andric for (MergeSizeBits = MaxSizeBits; MergeSizeBits > 1; MergeSizeBits /= 2) { 329349cc55cSDimitry Andric LLT StoreTy = LLT::scalar(MergeSizeBits); 330349cc55cSDimitry Andric EVT StoreEVT = 331349cc55cSDimitry Andric getApproximateEVTForLLT(StoreTy, DL, MF->getFunction().getContext()); 332349cc55cSDimitry Andric if (LegalSizes.size() > MergeSizeBits && LegalSizes[MergeSizeBits] && 333349cc55cSDimitry Andric TLI->canMergeStoresTo(AS, StoreEVT, *MF) && 334349cc55cSDimitry Andric (TLI->isTypeLegal(StoreEVT))) 335349cc55cSDimitry Andric break; // We can generate a MergeSize bits store. 336349cc55cSDimitry Andric } 337349cc55cSDimitry Andric if (MergeSizeBits <= OrigTy.getSizeInBits()) 338349cc55cSDimitry Andric return AnyMerged; // No greater merge. 339349cc55cSDimitry Andric 340349cc55cSDimitry Andric unsigned NumStoresToMerge = MergeSizeBits / OrigTy.getSizeInBits(); 341349cc55cSDimitry Andric // Perform the actual merging. 342349cc55cSDimitry Andric SmallVector<GStore *, 8> SingleMergeStores( 343349cc55cSDimitry Andric StoresToMerge.begin(), StoresToMerge.begin() + NumStoresToMerge); 344349cc55cSDimitry Andric AnyMerged |= doSingleStoreMerge(SingleMergeStores); 345349cc55cSDimitry Andric StoresToMerge.erase(StoresToMerge.begin(), 346349cc55cSDimitry Andric StoresToMerge.begin() + NumStoresToMerge); 347349cc55cSDimitry Andric } while (StoresToMerge.size() > 1); 348349cc55cSDimitry Andric return AnyMerged; 349349cc55cSDimitry Andric } 350349cc55cSDimitry Andric 351349cc55cSDimitry Andric bool LoadStoreOpt::isLegalOrBeforeLegalizer(const LegalityQuery &Query, 352349cc55cSDimitry Andric MachineFunction &MF) const { 353349cc55cSDimitry Andric auto Action = LI->getAction(Query).Action; 354349cc55cSDimitry Andric // If the instruction is unsupported, it can't be legalized at all. 355349cc55cSDimitry Andric if (Action == LegalizeActions::Unsupported) 356349cc55cSDimitry Andric return false; 357349cc55cSDimitry Andric return IsPreLegalizer || Action == LegalizeAction::Legal; 358349cc55cSDimitry Andric } 359349cc55cSDimitry Andric 360349cc55cSDimitry Andric bool LoadStoreOpt::doSingleStoreMerge(SmallVectorImpl<GStore *> &Stores) { 361349cc55cSDimitry Andric assert(Stores.size() > 1); 362349cc55cSDimitry Andric // We know that all the stores are consecutive and there are no aliasing 363349cc55cSDimitry Andric // operations in the range. However, the values that are being stored may be 364349cc55cSDimitry Andric // generated anywhere before each store. To ensure we have the values 365349cc55cSDimitry Andric // available, we materialize the wide value and new store at the place of the 366349cc55cSDimitry Andric // final store in the merge sequence. 367349cc55cSDimitry Andric GStore *FirstStore = Stores[0]; 368349cc55cSDimitry Andric const unsigned NumStores = Stores.size(); 369349cc55cSDimitry Andric LLT SmallTy = MRI->getType(FirstStore->getValueReg()); 370349cc55cSDimitry Andric LLT WideValueTy = 371bdd1243dSDimitry Andric LLT::scalar(NumStores * SmallTy.getSizeInBits().getFixedValue()); 372349cc55cSDimitry Andric 373349cc55cSDimitry Andric // For each store, compute pairwise merged debug locs. 374bdd1243dSDimitry Andric DebugLoc MergedLoc = Stores.front()->getDebugLoc(); 375bdd1243dSDimitry Andric for (auto *Store : drop_begin(Stores)) 376bdd1243dSDimitry Andric MergedLoc = DILocation::getMergedLocation(MergedLoc, Store->getDebugLoc()); 377bdd1243dSDimitry Andric 378349cc55cSDimitry Andric Builder.setInstr(*Stores.back()); 379349cc55cSDimitry Andric Builder.setDebugLoc(MergedLoc); 380349cc55cSDimitry Andric 381349cc55cSDimitry Andric // If all of the store values are constants, then create a wide constant 382349cc55cSDimitry Andric // directly. Otherwise, we need to generate some instructions to merge the 383349cc55cSDimitry Andric // existing values together into a wider type. 384349cc55cSDimitry Andric SmallVector<APInt, 8> ConstantVals; 385fcaf7f86SDimitry Andric for (auto *Store : Stores) { 386349cc55cSDimitry Andric auto MaybeCst = 387349cc55cSDimitry Andric getIConstantVRegValWithLookThrough(Store->getValueReg(), *MRI); 388349cc55cSDimitry Andric if (!MaybeCst) { 389349cc55cSDimitry Andric ConstantVals.clear(); 390349cc55cSDimitry Andric break; 391349cc55cSDimitry Andric } 392349cc55cSDimitry Andric ConstantVals.emplace_back(MaybeCst->Value); 393349cc55cSDimitry Andric } 394349cc55cSDimitry Andric 395349cc55cSDimitry Andric Register WideReg; 396349cc55cSDimitry Andric auto *WideMMO = 397349cc55cSDimitry Andric MF->getMachineMemOperand(&FirstStore->getMMO(), 0, WideValueTy); 398349cc55cSDimitry Andric if (ConstantVals.empty()) { 399349cc55cSDimitry Andric // Mimic the SDAG behaviour here and don't try to do anything for unknown 400349cc55cSDimitry Andric // values. In future, we should also support the cases of loads and 401349cc55cSDimitry Andric // extracted vector elements. 402349cc55cSDimitry Andric return false; 403349cc55cSDimitry Andric } 404349cc55cSDimitry Andric 405349cc55cSDimitry Andric assert(ConstantVals.size() == NumStores); 406349cc55cSDimitry Andric // Check if our wide constant is legal. 407349cc55cSDimitry Andric if (!isLegalOrBeforeLegalizer({TargetOpcode::G_CONSTANT, {WideValueTy}}, *MF)) 408349cc55cSDimitry Andric return false; 409349cc55cSDimitry Andric APInt WideConst(WideValueTy.getSizeInBits(), 0); 410349cc55cSDimitry Andric for (unsigned Idx = 0; Idx < ConstantVals.size(); ++Idx) { 411349cc55cSDimitry Andric // Insert the smaller constant into the corresponding position in the 412349cc55cSDimitry Andric // wider one. 413349cc55cSDimitry Andric WideConst.insertBits(ConstantVals[Idx], Idx * SmallTy.getSizeInBits()); 414349cc55cSDimitry Andric } 415349cc55cSDimitry Andric WideReg = Builder.buildConstant(WideValueTy, WideConst).getReg(0); 416349cc55cSDimitry Andric auto NewStore = 417349cc55cSDimitry Andric Builder.buildStore(WideReg, FirstStore->getPointerReg(), *WideMMO); 418349cc55cSDimitry Andric (void) NewStore; 41906c3fb27SDimitry Andric LLVM_DEBUG(dbgs() << "Merged " << Stores.size() 42006c3fb27SDimitry Andric << " stores into merged store: " << *NewStore); 42106c3fb27SDimitry Andric LLVM_DEBUG(for (auto *MI : Stores) dbgs() << " " << *MI;); 422349cc55cSDimitry Andric NumStoresMerged += Stores.size(); 423349cc55cSDimitry Andric 424349cc55cSDimitry Andric MachineOptimizationRemarkEmitter MORE(*MF, nullptr); 425349cc55cSDimitry Andric MORE.emit([&]() { 426349cc55cSDimitry Andric MachineOptimizationRemark R(DEBUG_TYPE, "MergedStore", 427349cc55cSDimitry Andric FirstStore->getDebugLoc(), 428349cc55cSDimitry Andric FirstStore->getParent()); 429349cc55cSDimitry Andric R << "Merged " << NV("NumMerged", Stores.size()) << " stores of " 430349cc55cSDimitry Andric << NV("OrigWidth", SmallTy.getSizeInBytes()) 431349cc55cSDimitry Andric << " bytes into a single store of " 432349cc55cSDimitry Andric << NV("NewWidth", WideValueTy.getSizeInBytes()) << " bytes"; 433349cc55cSDimitry Andric return R; 434349cc55cSDimitry Andric }); 435349cc55cSDimitry Andric 436fcaf7f86SDimitry Andric for (auto *MI : Stores) 437349cc55cSDimitry Andric InstsToErase.insert(MI); 438349cc55cSDimitry Andric return true; 439349cc55cSDimitry Andric } 440349cc55cSDimitry Andric 441349cc55cSDimitry Andric bool LoadStoreOpt::processMergeCandidate(StoreMergeCandidate &C) { 442349cc55cSDimitry Andric if (C.Stores.size() < 2) { 443349cc55cSDimitry Andric C.reset(); 444349cc55cSDimitry Andric return false; 445349cc55cSDimitry Andric } 446349cc55cSDimitry Andric 447349cc55cSDimitry Andric LLVM_DEBUG(dbgs() << "Checking store merge candidate with " << C.Stores.size() 448349cc55cSDimitry Andric << " stores, starting with " << *C.Stores[0]); 449349cc55cSDimitry Andric // We know that the stores in the candidate are adjacent. 450349cc55cSDimitry Andric // Now we need to check if any potential aliasing instructions recorded 451349cc55cSDimitry Andric // during the search alias with load/stores added to the candidate after. 452349cc55cSDimitry Andric // For example, if we have the candidate: 453349cc55cSDimitry Andric // C.Stores = [ST1, ST2, ST3, ST4] 454349cc55cSDimitry Andric // and after seeing ST2 we saw a load LD1, which did not alias with ST1 or 455349cc55cSDimitry Andric // ST2, then we would have recorded it into the PotentialAliases structure 456349cc55cSDimitry Andric // with the associated index value of "1". Then we see ST3 and ST4 and add 457349cc55cSDimitry Andric // them to the candidate group. We know that LD1 does not alias with ST1 or 458349cc55cSDimitry Andric // ST2, since we already did that check. However we don't yet know if it 459349cc55cSDimitry Andric // may alias ST3 and ST4, so we perform those checks now. 460349cc55cSDimitry Andric SmallVector<GStore *> StoresToMerge; 461349cc55cSDimitry Andric 462349cc55cSDimitry Andric auto DoesStoreAliasWithPotential = [&](unsigned Idx, GStore &CheckStore) { 463349cc55cSDimitry Andric for (auto AliasInfo : reverse(C.PotentialAliases)) { 464349cc55cSDimitry Andric MachineInstr *PotentialAliasOp = AliasInfo.first; 465349cc55cSDimitry Andric unsigned PreCheckedIdx = AliasInfo.second; 46606c3fb27SDimitry Andric if (static_cast<unsigned>(Idx) < PreCheckedIdx) { 46706c3fb27SDimitry Andric // Once our store index is lower than the index associated with the 46806c3fb27SDimitry Andric // potential alias, we know that we've already checked for this alias 46906c3fb27SDimitry Andric // and all of the earlier potential aliases too. 47006c3fb27SDimitry Andric return false; 47106c3fb27SDimitry Andric } 472349cc55cSDimitry Andric // Need to check this alias. 473349cc55cSDimitry Andric if (GISelAddressing::instMayAlias(CheckStore, *PotentialAliasOp, *MRI, 474349cc55cSDimitry Andric AA)) { 475349cc55cSDimitry Andric LLVM_DEBUG(dbgs() << "Potential alias " << *PotentialAliasOp 476349cc55cSDimitry Andric << " detected\n"); 477349cc55cSDimitry Andric return true; 478349cc55cSDimitry Andric } 479349cc55cSDimitry Andric } 480349cc55cSDimitry Andric return false; 481349cc55cSDimitry Andric }; 482349cc55cSDimitry Andric // Start from the last store in the group, and check if it aliases with any 483349cc55cSDimitry Andric // of the potential aliasing operations in the list. 484349cc55cSDimitry Andric for (int StoreIdx = C.Stores.size() - 1; StoreIdx >= 0; --StoreIdx) { 485349cc55cSDimitry Andric auto *CheckStore = C.Stores[StoreIdx]; 486349cc55cSDimitry Andric if (DoesStoreAliasWithPotential(StoreIdx, *CheckStore)) 487349cc55cSDimitry Andric continue; 488349cc55cSDimitry Andric StoresToMerge.emplace_back(CheckStore); 489349cc55cSDimitry Andric } 490349cc55cSDimitry Andric 491349cc55cSDimitry Andric LLVM_DEBUG(dbgs() << StoresToMerge.size() 492349cc55cSDimitry Andric << " stores remaining after alias checks. Merging...\n"); 493349cc55cSDimitry Andric 494349cc55cSDimitry Andric // Now we've checked for aliasing hazards, merge any stores left. 495349cc55cSDimitry Andric C.reset(); 496349cc55cSDimitry Andric if (StoresToMerge.size() < 2) 497349cc55cSDimitry Andric return false; 498349cc55cSDimitry Andric return mergeStores(StoresToMerge); 499349cc55cSDimitry Andric } 500349cc55cSDimitry Andric 501349cc55cSDimitry Andric bool LoadStoreOpt::operationAliasesWithCandidate(MachineInstr &MI, 502349cc55cSDimitry Andric StoreMergeCandidate &C) { 503349cc55cSDimitry Andric if (C.Stores.empty()) 504349cc55cSDimitry Andric return false; 505349cc55cSDimitry Andric return llvm::any_of(C.Stores, [&](MachineInstr *OtherMI) { 506349cc55cSDimitry Andric return instMayAlias(MI, *OtherMI, *MRI, AA); 507349cc55cSDimitry Andric }); 508349cc55cSDimitry Andric } 509349cc55cSDimitry Andric 510349cc55cSDimitry Andric void LoadStoreOpt::StoreMergeCandidate::addPotentialAlias(MachineInstr &MI) { 511349cc55cSDimitry Andric PotentialAliases.emplace_back(std::make_pair(&MI, Stores.size() - 1)); 512349cc55cSDimitry Andric } 513349cc55cSDimitry Andric 514349cc55cSDimitry Andric bool LoadStoreOpt::addStoreToCandidate(GStore &StoreMI, 515349cc55cSDimitry Andric StoreMergeCandidate &C) { 516349cc55cSDimitry Andric // Check if the given store writes to an adjacent address, and other 517349cc55cSDimitry Andric // requirements. 518349cc55cSDimitry Andric LLT ValueTy = MRI->getType(StoreMI.getValueReg()); 519349cc55cSDimitry Andric LLT PtrTy = MRI->getType(StoreMI.getPointerReg()); 520349cc55cSDimitry Andric 521349cc55cSDimitry Andric // Only handle scalars. 522349cc55cSDimitry Andric if (!ValueTy.isScalar()) 523349cc55cSDimitry Andric return false; 524349cc55cSDimitry Andric 525349cc55cSDimitry Andric // Don't allow truncating stores for now. 526349cc55cSDimitry Andric if (StoreMI.getMemSizeInBits() != ValueTy.getSizeInBits()) 527349cc55cSDimitry Andric return false; 528349cc55cSDimitry Andric 52981ad6265SDimitry Andric // Avoid adding volatile or ordered stores to the candidate. We already have a 53081ad6265SDimitry Andric // check for this in instMayAlias() but that only get's called later between 53181ad6265SDimitry Andric // potential aliasing hazards. 53281ad6265SDimitry Andric if (!StoreMI.isSimple()) 53381ad6265SDimitry Andric return false; 53481ad6265SDimitry Andric 535349cc55cSDimitry Andric Register StoreAddr = StoreMI.getPointerReg(); 536349cc55cSDimitry Andric auto BIO = getPointerInfo(StoreAddr, *MRI); 5375678d1d9SDimitry Andric Register StoreBase = BIO.getBase(); 538349cc55cSDimitry Andric if (C.Stores.empty()) { 5395678d1d9SDimitry Andric C.BasePtr = StoreBase; 5405678d1d9SDimitry Andric if (!BIO.hasValidOffset()) { 5415678d1d9SDimitry Andric C.CurrentLowestOffset = 0; 5425678d1d9SDimitry Andric } else { 5435678d1d9SDimitry Andric C.CurrentLowestOffset = BIO.getOffset(); 5445678d1d9SDimitry Andric } 545349cc55cSDimitry Andric // This is the first store of the candidate. 546349cc55cSDimitry Andric // If the offset can't possibly allow for a lower addressed store with the 547349cc55cSDimitry Andric // same base, don't bother adding it. 5485678d1d9SDimitry Andric if (BIO.hasValidOffset() && 5495678d1d9SDimitry Andric BIO.getOffset() < static_cast<int64_t>(ValueTy.getSizeInBytes())) 550349cc55cSDimitry Andric return false; 551349cc55cSDimitry Andric C.Stores.emplace_back(&StoreMI); 552349cc55cSDimitry Andric LLVM_DEBUG(dbgs() << "Starting a new merge candidate group with: " 553349cc55cSDimitry Andric << StoreMI); 554349cc55cSDimitry Andric return true; 555349cc55cSDimitry Andric } 556349cc55cSDimitry Andric 557349cc55cSDimitry Andric // Check the store is the same size as the existing ones in the candidate. 558349cc55cSDimitry Andric if (MRI->getType(C.Stores[0]->getValueReg()).getSizeInBits() != 559349cc55cSDimitry Andric ValueTy.getSizeInBits()) 560349cc55cSDimitry Andric return false; 561349cc55cSDimitry Andric 562349cc55cSDimitry Andric if (MRI->getType(C.Stores[0]->getPointerReg()).getAddressSpace() != 563349cc55cSDimitry Andric PtrTy.getAddressSpace()) 564349cc55cSDimitry Andric return false; 565349cc55cSDimitry Andric 566349cc55cSDimitry Andric // There are other stores in the candidate. Check that the store address 567349cc55cSDimitry Andric // writes to the next lowest adjacent address. 568349cc55cSDimitry Andric if (C.BasePtr != StoreBase) 569349cc55cSDimitry Andric return false; 5705678d1d9SDimitry Andric // If we don't have a valid offset, we can't guarantee to be an adjacent 5715678d1d9SDimitry Andric // offset. 5725678d1d9SDimitry Andric if (!BIO.hasValidOffset()) 5735678d1d9SDimitry Andric return false; 5745678d1d9SDimitry Andric if ((C.CurrentLowestOffset - 5755678d1d9SDimitry Andric static_cast<int64_t>(ValueTy.getSizeInBytes())) != BIO.getOffset()) 576349cc55cSDimitry Andric return false; 577349cc55cSDimitry Andric 578349cc55cSDimitry Andric // This writes to an adjacent address. Allow it. 579349cc55cSDimitry Andric C.Stores.emplace_back(&StoreMI); 580349cc55cSDimitry Andric C.CurrentLowestOffset = C.CurrentLowestOffset - ValueTy.getSizeInBytes(); 581349cc55cSDimitry Andric LLVM_DEBUG(dbgs() << "Candidate added store: " << StoreMI); 582349cc55cSDimitry Andric return true; 583349cc55cSDimitry Andric } 584349cc55cSDimitry Andric 585349cc55cSDimitry Andric bool LoadStoreOpt::mergeBlockStores(MachineBasicBlock &MBB) { 586349cc55cSDimitry Andric bool Changed = false; 587349cc55cSDimitry Andric // Walk through the block bottom-up, looking for merging candidates. 588349cc55cSDimitry Andric StoreMergeCandidate Candidate; 5890eae32dcSDimitry Andric for (MachineInstr &MI : llvm::reverse(MBB)) { 590349cc55cSDimitry Andric if (InstsToErase.contains(&MI)) 591349cc55cSDimitry Andric continue; 592349cc55cSDimitry Andric 5930eae32dcSDimitry Andric if (auto *StoreMI = dyn_cast<GStore>(&MI)) { 594349cc55cSDimitry Andric // We have a G_STORE. Add it to the candidate if it writes to an adjacent 595349cc55cSDimitry Andric // address. 596349cc55cSDimitry Andric if (!addStoreToCandidate(*StoreMI, Candidate)) { 597349cc55cSDimitry Andric // Store wasn't eligible to be added. May need to record it as a 598349cc55cSDimitry Andric // potential alias. 599349cc55cSDimitry Andric if (operationAliasesWithCandidate(*StoreMI, Candidate)) { 600349cc55cSDimitry Andric Changed |= processMergeCandidate(Candidate); 601349cc55cSDimitry Andric continue; 602349cc55cSDimitry Andric } 603349cc55cSDimitry Andric Candidate.addPotentialAlias(*StoreMI); 604349cc55cSDimitry Andric } 605349cc55cSDimitry Andric continue; 606349cc55cSDimitry Andric } 607349cc55cSDimitry Andric 608349cc55cSDimitry Andric // If we don't have any stores yet, this instruction can't pose a problem. 609349cc55cSDimitry Andric if (Candidate.Stores.empty()) 610349cc55cSDimitry Andric continue; 611349cc55cSDimitry Andric 612349cc55cSDimitry Andric // We're dealing with some other kind of instruction. 613349cc55cSDimitry Andric if (isInstHardMergeHazard(MI)) { 614349cc55cSDimitry Andric Changed |= processMergeCandidate(Candidate); 615349cc55cSDimitry Andric Candidate.Stores.clear(); 616349cc55cSDimitry Andric continue; 617349cc55cSDimitry Andric } 618349cc55cSDimitry Andric 619349cc55cSDimitry Andric if (!MI.mayLoadOrStore()) 620349cc55cSDimitry Andric continue; 621349cc55cSDimitry Andric 622349cc55cSDimitry Andric if (operationAliasesWithCandidate(MI, Candidate)) { 623349cc55cSDimitry Andric // We have a potential alias, so process the current candidate if we can 624349cc55cSDimitry Andric // and then continue looking for a new candidate. 625349cc55cSDimitry Andric Changed |= processMergeCandidate(Candidate); 626349cc55cSDimitry Andric continue; 627349cc55cSDimitry Andric } 628349cc55cSDimitry Andric 629349cc55cSDimitry Andric // Record this instruction as a potential alias for future stores that are 630349cc55cSDimitry Andric // added to the candidate. 631349cc55cSDimitry Andric Candidate.addPotentialAlias(MI); 632349cc55cSDimitry Andric } 633349cc55cSDimitry Andric 634349cc55cSDimitry Andric // Process any candidate left after finishing searching the entire block. 635349cc55cSDimitry Andric Changed |= processMergeCandidate(Candidate); 636349cc55cSDimitry Andric 637349cc55cSDimitry Andric // Erase instructions now that we're no longer iterating over the block. 638349cc55cSDimitry Andric for (auto *MI : InstsToErase) 639349cc55cSDimitry Andric MI->eraseFromParent(); 640349cc55cSDimitry Andric InstsToErase.clear(); 641349cc55cSDimitry Andric return Changed; 642349cc55cSDimitry Andric } 643349cc55cSDimitry Andric 64406c3fb27SDimitry Andric /// Check if the store \p Store is a truncstore that can be merged. That is, 64506c3fb27SDimitry Andric /// it's a store of a shifted value of \p SrcVal. If \p SrcVal is an empty 64606c3fb27SDimitry Andric /// Register then it does not need to match and SrcVal is set to the source 64706c3fb27SDimitry Andric /// value found. 64806c3fb27SDimitry Andric /// On match, returns the start byte offset of the \p SrcVal that is being 64906c3fb27SDimitry Andric /// stored. 65006c3fb27SDimitry Andric static std::optional<int64_t> 65106c3fb27SDimitry Andric getTruncStoreByteOffset(GStore &Store, Register &SrcVal, 65206c3fb27SDimitry Andric MachineRegisterInfo &MRI) { 65306c3fb27SDimitry Andric Register TruncVal; 65406c3fb27SDimitry Andric if (!mi_match(Store.getValueReg(), MRI, m_GTrunc(m_Reg(TruncVal)))) 65506c3fb27SDimitry Andric return std::nullopt; 65606c3fb27SDimitry Andric 65706c3fb27SDimitry Andric // The shift amount must be a constant multiple of the narrow type. 65806c3fb27SDimitry Andric // It is translated to the offset address in the wide source value "y". 65906c3fb27SDimitry Andric // 66006c3fb27SDimitry Andric // x = G_LSHR y, ShiftAmtC 66106c3fb27SDimitry Andric // s8 z = G_TRUNC x 66206c3fb27SDimitry Andric // store z, ... 66306c3fb27SDimitry Andric Register FoundSrcVal; 66406c3fb27SDimitry Andric int64_t ShiftAmt; 66506c3fb27SDimitry Andric if (!mi_match(TruncVal, MRI, 66606c3fb27SDimitry Andric m_any_of(m_GLShr(m_Reg(FoundSrcVal), m_ICst(ShiftAmt)), 66706c3fb27SDimitry Andric m_GAShr(m_Reg(FoundSrcVal), m_ICst(ShiftAmt))))) { 66806c3fb27SDimitry Andric if (!SrcVal.isValid() || TruncVal == SrcVal) { 66906c3fb27SDimitry Andric if (!SrcVal.isValid()) 67006c3fb27SDimitry Andric SrcVal = TruncVal; 67106c3fb27SDimitry Andric return 0; // If it's the lowest index store. 67206c3fb27SDimitry Andric } 67306c3fb27SDimitry Andric return std::nullopt; 67406c3fb27SDimitry Andric } 67506c3fb27SDimitry Andric 67606c3fb27SDimitry Andric unsigned NarrowBits = Store.getMMO().getMemoryType().getScalarSizeInBits(); 67706c3fb27SDimitry Andric if (ShiftAmt % NarrowBits != 0) 67806c3fb27SDimitry Andric return std::nullopt; 67906c3fb27SDimitry Andric const unsigned Offset = ShiftAmt / NarrowBits; 68006c3fb27SDimitry Andric 68106c3fb27SDimitry Andric if (SrcVal.isValid() && FoundSrcVal != SrcVal) 68206c3fb27SDimitry Andric return std::nullopt; 68306c3fb27SDimitry Andric 68406c3fb27SDimitry Andric if (!SrcVal.isValid()) 68506c3fb27SDimitry Andric SrcVal = FoundSrcVal; 68606c3fb27SDimitry Andric else if (MRI.getType(SrcVal) != MRI.getType(FoundSrcVal)) 68706c3fb27SDimitry Andric return std::nullopt; 68806c3fb27SDimitry Andric return Offset; 68906c3fb27SDimitry Andric } 69006c3fb27SDimitry Andric 69106c3fb27SDimitry Andric /// Match a pattern where a wide type scalar value is stored by several narrow 69206c3fb27SDimitry Andric /// stores. Fold it into a single store or a BSWAP and a store if the targets 69306c3fb27SDimitry Andric /// supports it. 69406c3fb27SDimitry Andric /// 69506c3fb27SDimitry Andric /// Assuming little endian target: 69606c3fb27SDimitry Andric /// i8 *p = ... 69706c3fb27SDimitry Andric /// i32 val = ... 69806c3fb27SDimitry Andric /// p[0] = (val >> 0) & 0xFF; 69906c3fb27SDimitry Andric /// p[1] = (val >> 8) & 0xFF; 70006c3fb27SDimitry Andric /// p[2] = (val >> 16) & 0xFF; 70106c3fb27SDimitry Andric /// p[3] = (val >> 24) & 0xFF; 70206c3fb27SDimitry Andric /// => 70306c3fb27SDimitry Andric /// *((i32)p) = val; 70406c3fb27SDimitry Andric /// 70506c3fb27SDimitry Andric /// i8 *p = ... 70606c3fb27SDimitry Andric /// i32 val = ... 70706c3fb27SDimitry Andric /// p[0] = (val >> 24) & 0xFF; 70806c3fb27SDimitry Andric /// p[1] = (val >> 16) & 0xFF; 70906c3fb27SDimitry Andric /// p[2] = (val >> 8) & 0xFF; 71006c3fb27SDimitry Andric /// p[3] = (val >> 0) & 0xFF; 71106c3fb27SDimitry Andric /// => 71206c3fb27SDimitry Andric /// *((i32)p) = BSWAP(val); 71306c3fb27SDimitry Andric bool LoadStoreOpt::mergeTruncStore(GStore &StoreMI, 71406c3fb27SDimitry Andric SmallPtrSetImpl<GStore *> &DeletedStores) { 71506c3fb27SDimitry Andric LLT MemTy = StoreMI.getMMO().getMemoryType(); 71606c3fb27SDimitry Andric 71706c3fb27SDimitry Andric // We only handle merging simple stores of 1-4 bytes. 71806c3fb27SDimitry Andric if (!MemTy.isScalar()) 71906c3fb27SDimitry Andric return false; 72006c3fb27SDimitry Andric switch (MemTy.getSizeInBits()) { 72106c3fb27SDimitry Andric case 8: 72206c3fb27SDimitry Andric case 16: 72306c3fb27SDimitry Andric case 32: 72406c3fb27SDimitry Andric break; 72506c3fb27SDimitry Andric default: 72606c3fb27SDimitry Andric return false; 72706c3fb27SDimitry Andric } 72806c3fb27SDimitry Andric if (!StoreMI.isSimple()) 72906c3fb27SDimitry Andric return false; 73006c3fb27SDimitry Andric 73106c3fb27SDimitry Andric // We do a simple search for mergeable stores prior to this one. 73206c3fb27SDimitry Andric // Any potential alias hazard along the way terminates the search. 73306c3fb27SDimitry Andric SmallVector<GStore *> FoundStores; 73406c3fb27SDimitry Andric 73506c3fb27SDimitry Andric // We're looking for: 73606c3fb27SDimitry Andric // 1) a (store(trunc(...))) 73706c3fb27SDimitry Andric // 2) of an LSHR/ASHR of a single wide value, by the appropriate shift to get 73806c3fb27SDimitry Andric // the partial value stored. 73906c3fb27SDimitry Andric // 3) where the offsets form either a little or big-endian sequence. 74006c3fb27SDimitry Andric 74106c3fb27SDimitry Andric auto &LastStore = StoreMI; 74206c3fb27SDimitry Andric 74306c3fb27SDimitry Andric // The single base pointer that all stores must use. 74406c3fb27SDimitry Andric Register BaseReg; 74506c3fb27SDimitry Andric int64_t LastOffset; 74606c3fb27SDimitry Andric if (!mi_match(LastStore.getPointerReg(), *MRI, 74706c3fb27SDimitry Andric m_GPtrAdd(m_Reg(BaseReg), m_ICst(LastOffset)))) { 74806c3fb27SDimitry Andric BaseReg = LastStore.getPointerReg(); 74906c3fb27SDimitry Andric LastOffset = 0; 75006c3fb27SDimitry Andric } 75106c3fb27SDimitry Andric 75206c3fb27SDimitry Andric GStore *LowestIdxStore = &LastStore; 75306c3fb27SDimitry Andric int64_t LowestIdxOffset = LastOffset; 75406c3fb27SDimitry Andric 75506c3fb27SDimitry Andric Register WideSrcVal; 75606c3fb27SDimitry Andric auto LowestShiftAmt = getTruncStoreByteOffset(LastStore, WideSrcVal, *MRI); 75706c3fb27SDimitry Andric if (!LowestShiftAmt) 75806c3fb27SDimitry Andric return false; // Didn't match a trunc. 75906c3fb27SDimitry Andric assert(WideSrcVal.isValid()); 76006c3fb27SDimitry Andric 76106c3fb27SDimitry Andric LLT WideStoreTy = MRI->getType(WideSrcVal); 76206c3fb27SDimitry Andric // The wide type might not be a multiple of the memory type, e.g. s48 and s32. 76306c3fb27SDimitry Andric if (WideStoreTy.getSizeInBits() % MemTy.getSizeInBits() != 0) 76406c3fb27SDimitry Andric return false; 76506c3fb27SDimitry Andric const unsigned NumStoresRequired = 76606c3fb27SDimitry Andric WideStoreTy.getSizeInBits() / MemTy.getSizeInBits(); 76706c3fb27SDimitry Andric 76806c3fb27SDimitry Andric SmallVector<int64_t, 8> OffsetMap(NumStoresRequired, INT64_MAX); 76906c3fb27SDimitry Andric OffsetMap[*LowestShiftAmt] = LastOffset; 77006c3fb27SDimitry Andric FoundStores.emplace_back(&LastStore); 77106c3fb27SDimitry Andric 77206c3fb27SDimitry Andric const int MaxInstsToCheck = 10; 77306c3fb27SDimitry Andric int NumInstsChecked = 0; 77406c3fb27SDimitry Andric for (auto II = ++LastStore.getReverseIterator(); 77506c3fb27SDimitry Andric II != LastStore.getParent()->rend() && NumInstsChecked < MaxInstsToCheck; 77606c3fb27SDimitry Andric ++II) { 77706c3fb27SDimitry Andric NumInstsChecked++; 77806c3fb27SDimitry Andric GStore *NewStore; 77906c3fb27SDimitry Andric if ((NewStore = dyn_cast<GStore>(&*II))) { 78006c3fb27SDimitry Andric if (NewStore->getMMO().getMemoryType() != MemTy || !NewStore->isSimple()) 78106c3fb27SDimitry Andric break; 78206c3fb27SDimitry Andric } else if (II->isLoadFoldBarrier() || II->mayLoad()) { 78306c3fb27SDimitry Andric break; 78406c3fb27SDimitry Andric } else { 78506c3fb27SDimitry Andric continue; // This is a safe instruction we can look past. 78606c3fb27SDimitry Andric } 78706c3fb27SDimitry Andric 78806c3fb27SDimitry Andric Register NewBaseReg; 78906c3fb27SDimitry Andric int64_t MemOffset; 79006c3fb27SDimitry Andric // Check we're storing to the same base + some offset. 79106c3fb27SDimitry Andric if (!mi_match(NewStore->getPointerReg(), *MRI, 79206c3fb27SDimitry Andric m_GPtrAdd(m_Reg(NewBaseReg), m_ICst(MemOffset)))) { 79306c3fb27SDimitry Andric NewBaseReg = NewStore->getPointerReg(); 79406c3fb27SDimitry Andric MemOffset = 0; 79506c3fb27SDimitry Andric } 79606c3fb27SDimitry Andric if (BaseReg != NewBaseReg) 79706c3fb27SDimitry Andric break; 79806c3fb27SDimitry Andric 79906c3fb27SDimitry Andric auto ShiftByteOffset = getTruncStoreByteOffset(*NewStore, WideSrcVal, *MRI); 80006c3fb27SDimitry Andric if (!ShiftByteOffset) 80106c3fb27SDimitry Andric break; 80206c3fb27SDimitry Andric if (MemOffset < LowestIdxOffset) { 80306c3fb27SDimitry Andric LowestIdxOffset = MemOffset; 80406c3fb27SDimitry Andric LowestIdxStore = NewStore; 80506c3fb27SDimitry Andric } 80606c3fb27SDimitry Andric 80706c3fb27SDimitry Andric // Map the offset in the store and the offset in the combined value, and 80806c3fb27SDimitry Andric // early return if it has been set before. 80906c3fb27SDimitry Andric if (*ShiftByteOffset < 0 || *ShiftByteOffset >= NumStoresRequired || 81006c3fb27SDimitry Andric OffsetMap[*ShiftByteOffset] != INT64_MAX) 81106c3fb27SDimitry Andric break; 81206c3fb27SDimitry Andric OffsetMap[*ShiftByteOffset] = MemOffset; 81306c3fb27SDimitry Andric 81406c3fb27SDimitry Andric FoundStores.emplace_back(NewStore); 81506c3fb27SDimitry Andric // Reset counter since we've found a matching inst. 81606c3fb27SDimitry Andric NumInstsChecked = 0; 81706c3fb27SDimitry Andric if (FoundStores.size() == NumStoresRequired) 81806c3fb27SDimitry Andric break; 81906c3fb27SDimitry Andric } 82006c3fb27SDimitry Andric 82106c3fb27SDimitry Andric if (FoundStores.size() != NumStoresRequired) { 82206c3fb27SDimitry Andric if (FoundStores.size() == 1) 82306c3fb27SDimitry Andric return false; 82406c3fb27SDimitry Andric // We didn't find enough stores to merge into the size of the original 82506c3fb27SDimitry Andric // source value, but we may be able to generate a smaller store if we 82606c3fb27SDimitry Andric // truncate the source value. 82706c3fb27SDimitry Andric WideStoreTy = LLT::scalar(FoundStores.size() * MemTy.getScalarSizeInBits()); 82806c3fb27SDimitry Andric } 82906c3fb27SDimitry Andric 83006c3fb27SDimitry Andric unsigned NumStoresFound = FoundStores.size(); 83106c3fb27SDimitry Andric 83206c3fb27SDimitry Andric const auto &DL = LastStore.getMF()->getDataLayout(); 83306c3fb27SDimitry Andric auto &C = LastStore.getMF()->getFunction().getContext(); 83406c3fb27SDimitry Andric // Check that a store of the wide type is both allowed and fast on the target 83506c3fb27SDimitry Andric unsigned Fast = 0; 83606c3fb27SDimitry Andric bool Allowed = TLI->allowsMemoryAccess( 83706c3fb27SDimitry Andric C, DL, WideStoreTy, LowestIdxStore->getMMO(), &Fast); 83806c3fb27SDimitry Andric if (!Allowed || !Fast) 83906c3fb27SDimitry Andric return false; 84006c3fb27SDimitry Andric 84106c3fb27SDimitry Andric // Check if the pieces of the value are going to the expected places in memory 84206c3fb27SDimitry Andric // to merge the stores. 84306c3fb27SDimitry Andric unsigned NarrowBits = MemTy.getScalarSizeInBits(); 84406c3fb27SDimitry Andric auto checkOffsets = [&](bool MatchLittleEndian) { 84506c3fb27SDimitry Andric if (MatchLittleEndian) { 84606c3fb27SDimitry Andric for (unsigned i = 0; i != NumStoresFound; ++i) 84706c3fb27SDimitry Andric if (OffsetMap[i] != i * (NarrowBits / 8) + LowestIdxOffset) 84806c3fb27SDimitry Andric return false; 84906c3fb27SDimitry Andric } else { // MatchBigEndian by reversing loop counter. 85006c3fb27SDimitry Andric for (unsigned i = 0, j = NumStoresFound - 1; i != NumStoresFound; 85106c3fb27SDimitry Andric ++i, --j) 85206c3fb27SDimitry Andric if (OffsetMap[j] != i * (NarrowBits / 8) + LowestIdxOffset) 85306c3fb27SDimitry Andric return false; 85406c3fb27SDimitry Andric } 85506c3fb27SDimitry Andric return true; 85606c3fb27SDimitry Andric }; 85706c3fb27SDimitry Andric 85806c3fb27SDimitry Andric // Check if the offsets line up for the native data layout of this target. 85906c3fb27SDimitry Andric bool NeedBswap = false; 86006c3fb27SDimitry Andric bool NeedRotate = false; 86106c3fb27SDimitry Andric if (!checkOffsets(DL.isLittleEndian())) { 86206c3fb27SDimitry Andric // Special-case: check if byte offsets line up for the opposite endian. 86306c3fb27SDimitry Andric if (NarrowBits == 8 && checkOffsets(DL.isBigEndian())) 86406c3fb27SDimitry Andric NeedBswap = true; 86506c3fb27SDimitry Andric else if (NumStoresFound == 2 && checkOffsets(DL.isBigEndian())) 86606c3fb27SDimitry Andric NeedRotate = true; 86706c3fb27SDimitry Andric else 86806c3fb27SDimitry Andric return false; 86906c3fb27SDimitry Andric } 87006c3fb27SDimitry Andric 87106c3fb27SDimitry Andric if (NeedBswap && 87206c3fb27SDimitry Andric !isLegalOrBeforeLegalizer({TargetOpcode::G_BSWAP, {WideStoreTy}}, *MF)) 87306c3fb27SDimitry Andric return false; 87406c3fb27SDimitry Andric if (NeedRotate && 87506c3fb27SDimitry Andric !isLegalOrBeforeLegalizer( 87606c3fb27SDimitry Andric {TargetOpcode::G_ROTR, {WideStoreTy, WideStoreTy}}, *MF)) 87706c3fb27SDimitry Andric return false; 87806c3fb27SDimitry Andric 87906c3fb27SDimitry Andric Builder.setInstrAndDebugLoc(StoreMI); 88006c3fb27SDimitry Andric 88106c3fb27SDimitry Andric if (WideStoreTy != MRI->getType(WideSrcVal)) 88206c3fb27SDimitry Andric WideSrcVal = Builder.buildTrunc(WideStoreTy, WideSrcVal).getReg(0); 88306c3fb27SDimitry Andric 88406c3fb27SDimitry Andric if (NeedBswap) { 88506c3fb27SDimitry Andric WideSrcVal = Builder.buildBSwap(WideStoreTy, WideSrcVal).getReg(0); 88606c3fb27SDimitry Andric } else if (NeedRotate) { 88706c3fb27SDimitry Andric assert(WideStoreTy.getSizeInBits() % 2 == 0 && 88806c3fb27SDimitry Andric "Unexpected type for rotate"); 88906c3fb27SDimitry Andric auto RotAmt = 89006c3fb27SDimitry Andric Builder.buildConstant(WideStoreTy, WideStoreTy.getSizeInBits() / 2); 89106c3fb27SDimitry Andric WideSrcVal = 89206c3fb27SDimitry Andric Builder.buildRotateRight(WideStoreTy, WideSrcVal, RotAmt).getReg(0); 89306c3fb27SDimitry Andric } 89406c3fb27SDimitry Andric 89506c3fb27SDimitry Andric Builder.buildStore(WideSrcVal, LowestIdxStore->getPointerReg(), 89606c3fb27SDimitry Andric LowestIdxStore->getMMO().getPointerInfo(), 89706c3fb27SDimitry Andric LowestIdxStore->getMMO().getAlign()); 89806c3fb27SDimitry Andric 89906c3fb27SDimitry Andric // Erase the old stores. 90006c3fb27SDimitry Andric for (auto *ST : FoundStores) { 90106c3fb27SDimitry Andric ST->eraseFromParent(); 90206c3fb27SDimitry Andric DeletedStores.insert(ST); 90306c3fb27SDimitry Andric } 90406c3fb27SDimitry Andric return true; 90506c3fb27SDimitry Andric } 90606c3fb27SDimitry Andric 90706c3fb27SDimitry Andric bool LoadStoreOpt::mergeTruncStoresBlock(MachineBasicBlock &BB) { 90806c3fb27SDimitry Andric bool Changed = false; 90906c3fb27SDimitry Andric SmallVector<GStore *, 16> Stores; 91006c3fb27SDimitry Andric SmallPtrSet<GStore *, 8> DeletedStores; 91106c3fb27SDimitry Andric // Walk up the block so we can see the most eligible stores. 91206c3fb27SDimitry Andric for (MachineInstr &MI : llvm::reverse(BB)) 91306c3fb27SDimitry Andric if (auto *StoreMI = dyn_cast<GStore>(&MI)) 91406c3fb27SDimitry Andric Stores.emplace_back(StoreMI); 91506c3fb27SDimitry Andric 91606c3fb27SDimitry Andric for (auto *StoreMI : Stores) { 91706c3fb27SDimitry Andric if (DeletedStores.count(StoreMI)) 91806c3fb27SDimitry Andric continue; 91906c3fb27SDimitry Andric if (mergeTruncStore(*StoreMI, DeletedStores)) 92006c3fb27SDimitry Andric Changed = true; 92106c3fb27SDimitry Andric } 92206c3fb27SDimitry Andric return Changed; 92306c3fb27SDimitry Andric } 92406c3fb27SDimitry Andric 925349cc55cSDimitry Andric bool LoadStoreOpt::mergeFunctionStores(MachineFunction &MF) { 926349cc55cSDimitry Andric bool Changed = false; 927349cc55cSDimitry Andric for (auto &BB : MF){ 928349cc55cSDimitry Andric Changed |= mergeBlockStores(BB); 92906c3fb27SDimitry Andric Changed |= mergeTruncStoresBlock(BB); 930349cc55cSDimitry Andric } 93106c3fb27SDimitry Andric 93206c3fb27SDimitry Andric // Erase all dead instructions left over by the merging. 93306c3fb27SDimitry Andric if (Changed) { 93406c3fb27SDimitry Andric for (auto &BB : MF) { 93506c3fb27SDimitry Andric for (auto &I : make_early_inc_range(make_range(BB.rbegin(), BB.rend()))) { 93606c3fb27SDimitry Andric if (isTriviallyDead(I, *MRI)) 93706c3fb27SDimitry Andric I.eraseFromParent(); 93806c3fb27SDimitry Andric } 93906c3fb27SDimitry Andric } 94006c3fb27SDimitry Andric } 94106c3fb27SDimitry Andric 942349cc55cSDimitry Andric return Changed; 943349cc55cSDimitry Andric } 944349cc55cSDimitry Andric 945349cc55cSDimitry Andric void LoadStoreOpt::initializeStoreMergeTargetInfo(unsigned AddrSpace) { 946349cc55cSDimitry Andric // Query the legalizer info to record what store types are legal. 947349cc55cSDimitry Andric // We record this because we don't want to bother trying to merge stores into 948349cc55cSDimitry Andric // illegal ones, which would just result in being split again. 949349cc55cSDimitry Andric 950349cc55cSDimitry Andric if (LegalStoreSizes.count(AddrSpace)) { 951349cc55cSDimitry Andric assert(LegalStoreSizes[AddrSpace].any()); 952349cc55cSDimitry Andric return; // Already cached sizes for this address space. 953349cc55cSDimitry Andric } 954349cc55cSDimitry Andric 955349cc55cSDimitry Andric // Need to reserve at least MaxStoreSizeToForm + 1 bits. 956349cc55cSDimitry Andric BitVector LegalSizes(MaxStoreSizeToForm * 2); 957349cc55cSDimitry Andric const auto &LI = *MF->getSubtarget().getLegalizerInfo(); 958*0fca6ea1SDimitry Andric const auto &DL = MF->getFunction().getDataLayout(); 9595f757f3fSDimitry Andric Type *IRPtrTy = PointerType::get(MF->getFunction().getContext(), AddrSpace); 9605f757f3fSDimitry Andric LLT PtrTy = getLLTForType(*IRPtrTy, DL); 961349cc55cSDimitry Andric // We assume that we're not going to be generating any stores wider than 962349cc55cSDimitry Andric // MaxStoreSizeToForm bits for now. 963349cc55cSDimitry Andric for (unsigned Size = 2; Size <= MaxStoreSizeToForm; Size *= 2) { 964349cc55cSDimitry Andric LLT Ty = LLT::scalar(Size); 965349cc55cSDimitry Andric SmallVector<LegalityQuery::MemDesc, 2> MemDescrs( 966349cc55cSDimitry Andric {{Ty, Ty.getSizeInBits(), AtomicOrdering::NotAtomic}}); 967349cc55cSDimitry Andric SmallVector<LLT> StoreTys({Ty, PtrTy}); 968349cc55cSDimitry Andric LegalityQuery Q(TargetOpcode::G_STORE, StoreTys, MemDescrs); 969349cc55cSDimitry Andric LegalizeActionStep ActionStep = LI.getAction(Q); 970349cc55cSDimitry Andric if (ActionStep.Action == LegalizeActions::Legal) 971349cc55cSDimitry Andric LegalSizes.set(Size); 972349cc55cSDimitry Andric } 973349cc55cSDimitry Andric assert(LegalSizes.any() && "Expected some store sizes to be legal!"); 974349cc55cSDimitry Andric LegalStoreSizes[AddrSpace] = LegalSizes; 975349cc55cSDimitry Andric } 976349cc55cSDimitry Andric 977349cc55cSDimitry Andric bool LoadStoreOpt::runOnMachineFunction(MachineFunction &MF) { 978349cc55cSDimitry Andric // If the ISel pipeline failed, do not bother running that pass. 979349cc55cSDimitry Andric if (MF.getProperties().hasProperty( 980349cc55cSDimitry Andric MachineFunctionProperties::Property::FailedISel)) 981349cc55cSDimitry Andric return false; 982349cc55cSDimitry Andric 983349cc55cSDimitry Andric LLVM_DEBUG(dbgs() << "Begin memory optimizations for: " << MF.getName() 984349cc55cSDimitry Andric << '\n'); 985349cc55cSDimitry Andric 986349cc55cSDimitry Andric init(MF); 987349cc55cSDimitry Andric bool Changed = false; 988349cc55cSDimitry Andric Changed |= mergeFunctionStores(MF); 989349cc55cSDimitry Andric 990349cc55cSDimitry Andric LegalStoreSizes.clear(); 991349cc55cSDimitry Andric return Changed; 992349cc55cSDimitry Andric } 993