1 //===- IRBuilder.cpp - Builder for LLVM Instrs ----------------------------===// 2 // 3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. 4 // See https://llvm.org/LICENSE.txt for license information. 5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception 6 // 7 //===----------------------------------------------------------------------===// 8 // 9 // This file implements the IRBuilder class, which is used as a convenient way 10 // to create LLVM instructions with a consistent and simplified interface. 11 // 12 //===----------------------------------------------------------------------===// 13 14 #include "llvm/IR/IRBuilder.h" 15 #include "llvm/ADT/ArrayRef.h" 16 #include "llvm/IR/Constant.h" 17 #include "llvm/IR/Constants.h" 18 #include "llvm/IR/DebugInfoMetadata.h" 19 #include "llvm/IR/DerivedTypes.h" 20 #include "llvm/IR/Function.h" 21 #include "llvm/IR/GlobalValue.h" 22 #include "llvm/IR/GlobalVariable.h" 23 #include "llvm/IR/IntrinsicInst.h" 24 #include "llvm/IR/Intrinsics.h" 25 #include "llvm/IR/LLVMContext.h" 26 #include "llvm/IR/Module.h" 27 #include "llvm/IR/NoFolder.h" 28 #include "llvm/IR/Operator.h" 29 #include "llvm/IR/Statepoint.h" 30 #include "llvm/IR/Type.h" 31 #include "llvm/IR/Value.h" 32 #include "llvm/Support/Casting.h" 33 #include <cassert> 34 #include <cstdint> 35 #include <optional> 36 #include <vector> 37 38 using namespace llvm; 39 40 /// CreateGlobalString - Make a new global variable with an initializer that 41 /// has array of i8 type filled in with the nul terminated string value 42 /// specified. If Name is specified, it is the name of the global variable 43 /// created. 44 GlobalVariable *IRBuilderBase::CreateGlobalString(StringRef Str, 45 const Twine &Name, 46 unsigned AddressSpace, 47 Module *M, bool AddNull) { 48 Constant *StrConstant = ConstantDataArray::getString(Context, Str, AddNull); 49 if (!M) 50 M = BB->getParent()->getParent(); 51 auto *GV = new GlobalVariable( 52 *M, StrConstant->getType(), true, GlobalValue::PrivateLinkage, 53 StrConstant, Name, nullptr, GlobalVariable::NotThreadLocal, AddressSpace); 54 GV->setUnnamedAddr(GlobalValue::UnnamedAddr::Global); 55 GV->setAlignment(Align(1)); 56 return GV; 57 } 58 59 Type *IRBuilderBase::getCurrentFunctionReturnType() const { 60 assert(BB && BB->getParent() && "No current function!"); 61 return BB->getParent()->getReturnType(); 62 } 63 64 DebugLoc IRBuilderBase::getCurrentDebugLocation() const { 65 for (auto &KV : MetadataToCopy) 66 if (KV.first == LLVMContext::MD_dbg) 67 return {cast<DILocation>(KV.second)}; 68 69 return {}; 70 } 71 void IRBuilderBase::SetInstDebugLocation(Instruction *I) const { 72 for (const auto &KV : MetadataToCopy) 73 if (KV.first == LLVMContext::MD_dbg) { 74 I->setDebugLoc(DebugLoc(KV.second)); 75 return; 76 } 77 } 78 79 CallInst * 80 IRBuilderBase::createCallHelper(Function *Callee, ArrayRef<Value *> Ops, 81 const Twine &Name, FMFSource FMFSource, 82 ArrayRef<OperandBundleDef> OpBundles) { 83 CallInst *CI = CreateCall(Callee, Ops, OpBundles, Name); 84 if (isa<FPMathOperator>(CI)) 85 CI->setFastMathFlags(FMFSource.get(FMF)); 86 return CI; 87 } 88 89 Value *IRBuilderBase::CreateVScale(Constant *Scaling, const Twine &Name) { 90 assert(isa<ConstantInt>(Scaling) && "Expected constant integer"); 91 if (cast<ConstantInt>(Scaling)->isZero()) 92 return Scaling; 93 CallInst *CI = 94 CreateIntrinsic(Intrinsic::vscale, {Scaling->getType()}, {}, {}, Name); 95 return cast<ConstantInt>(Scaling)->isOne() ? CI : CreateMul(CI, Scaling); 96 } 97 98 Value *IRBuilderBase::CreateElementCount(Type *DstType, ElementCount EC) { 99 Constant *MinEC = ConstantInt::get(DstType, EC.getKnownMinValue()); 100 return EC.isScalable() ? CreateVScale(MinEC) : MinEC; 101 } 102 103 Value *IRBuilderBase::CreateTypeSize(Type *DstType, TypeSize Size) { 104 Constant *MinSize = ConstantInt::get(DstType, Size.getKnownMinValue()); 105 return Size.isScalable() ? CreateVScale(MinSize) : MinSize; 106 } 107 108 Value *IRBuilderBase::CreateStepVector(Type *DstType, const Twine &Name) { 109 Type *STy = DstType->getScalarType(); 110 if (isa<ScalableVectorType>(DstType)) { 111 Type *StepVecType = DstType; 112 // TODO: We expect this special case (element type < 8 bits) to be 113 // temporary - once the intrinsic properly supports < 8 bits this code 114 // can be removed. 115 if (STy->getScalarSizeInBits() < 8) 116 StepVecType = 117 VectorType::get(getInt8Ty(), cast<ScalableVectorType>(DstType)); 118 Value *Res = CreateIntrinsic(Intrinsic::stepvector, {StepVecType}, {}, 119 nullptr, Name); 120 if (StepVecType != DstType) 121 Res = CreateTrunc(Res, DstType); 122 return Res; 123 } 124 125 unsigned NumEls = cast<FixedVectorType>(DstType)->getNumElements(); 126 127 // Create a vector of consecutive numbers from zero to VF. 128 SmallVector<Constant *, 8> Indices; 129 for (unsigned i = 0; i < NumEls; ++i) 130 Indices.push_back(ConstantInt::get(STy, i)); 131 132 // Add the consecutive indices to the vector value. 133 return ConstantVector::get(Indices); 134 } 135 136 CallInst *IRBuilderBase::CreateMemSet(Value *Ptr, Value *Val, Value *Size, 137 MaybeAlign Align, bool isVolatile, 138 MDNode *TBAATag, MDNode *ScopeTag, 139 MDNode *NoAliasTag) { 140 Value *Ops[] = {Ptr, Val, Size, getInt1(isVolatile)}; 141 Type *Tys[] = {Ptr->getType(), Size->getType()}; 142 143 CallInst *CI = CreateIntrinsic(Intrinsic::memset, Tys, Ops); 144 145 if (Align) 146 cast<MemSetInst>(CI)->setDestAlignment(*Align); 147 148 // Set the TBAA info if present. 149 if (TBAATag) 150 CI->setMetadata(LLVMContext::MD_tbaa, TBAATag); 151 152 if (ScopeTag) 153 CI->setMetadata(LLVMContext::MD_alias_scope, ScopeTag); 154 155 if (NoAliasTag) 156 CI->setMetadata(LLVMContext::MD_noalias, NoAliasTag); 157 158 return CI; 159 } 160 161 CallInst *IRBuilderBase::CreateMemSetInline(Value *Dst, MaybeAlign DstAlign, 162 Value *Val, Value *Size, 163 bool IsVolatile, MDNode *TBAATag, 164 MDNode *ScopeTag, 165 MDNode *NoAliasTag) { 166 Value *Ops[] = {Dst, Val, Size, getInt1(IsVolatile)}; 167 Type *Tys[] = {Dst->getType(), Size->getType()}; 168 169 CallInst *CI = CreateIntrinsic(Intrinsic::memset_inline, Tys, Ops); 170 171 if (DstAlign) 172 cast<MemSetInlineInst>(CI)->setDestAlignment(*DstAlign); 173 174 // Set the TBAA info if present. 175 if (TBAATag) 176 CI->setMetadata(LLVMContext::MD_tbaa, TBAATag); 177 178 if (ScopeTag) 179 CI->setMetadata(LLVMContext::MD_alias_scope, ScopeTag); 180 181 if (NoAliasTag) 182 CI->setMetadata(LLVMContext::MD_noalias, NoAliasTag); 183 184 return CI; 185 } 186 187 CallInst *IRBuilderBase::CreateElementUnorderedAtomicMemSet( 188 Value *Ptr, Value *Val, Value *Size, Align Alignment, uint32_t ElementSize, 189 MDNode *TBAATag, MDNode *ScopeTag, MDNode *NoAliasTag) { 190 191 Value *Ops[] = {Ptr, Val, Size, getInt32(ElementSize)}; 192 Type *Tys[] = {Ptr->getType(), Size->getType()}; 193 194 CallInst *CI = 195 CreateIntrinsic(Intrinsic::memset_element_unordered_atomic, Tys, Ops); 196 197 cast<AtomicMemSetInst>(CI)->setDestAlignment(Alignment); 198 199 // Set the TBAA info if present. 200 if (TBAATag) 201 CI->setMetadata(LLVMContext::MD_tbaa, TBAATag); 202 203 if (ScopeTag) 204 CI->setMetadata(LLVMContext::MD_alias_scope, ScopeTag); 205 206 if (NoAliasTag) 207 CI->setMetadata(LLVMContext::MD_noalias, NoAliasTag); 208 209 return CI; 210 } 211 212 CallInst *IRBuilderBase::CreateMemTransferInst( 213 Intrinsic::ID IntrID, Value *Dst, MaybeAlign DstAlign, Value *Src, 214 MaybeAlign SrcAlign, Value *Size, bool isVolatile, MDNode *TBAATag, 215 MDNode *TBAAStructTag, MDNode *ScopeTag, MDNode *NoAliasTag) { 216 assert((IntrID == Intrinsic::memcpy || IntrID == Intrinsic::memcpy_inline || 217 IntrID == Intrinsic::memmove) && 218 "Unexpected intrinsic ID"); 219 Value *Ops[] = {Dst, Src, Size, getInt1(isVolatile)}; 220 Type *Tys[] = {Dst->getType(), Src->getType(), Size->getType()}; 221 222 CallInst *CI = CreateIntrinsic(IntrID, Tys, Ops); 223 224 auto* MCI = cast<MemTransferInst>(CI); 225 if (DstAlign) 226 MCI->setDestAlignment(*DstAlign); 227 if (SrcAlign) 228 MCI->setSourceAlignment(*SrcAlign); 229 230 // Set the TBAA info if present. 231 if (TBAATag) 232 CI->setMetadata(LLVMContext::MD_tbaa, TBAATag); 233 234 // Set the TBAA Struct info if present. 235 if (TBAAStructTag) 236 CI->setMetadata(LLVMContext::MD_tbaa_struct, TBAAStructTag); 237 238 if (ScopeTag) 239 CI->setMetadata(LLVMContext::MD_alias_scope, ScopeTag); 240 241 if (NoAliasTag) 242 CI->setMetadata(LLVMContext::MD_noalias, NoAliasTag); 243 244 return CI; 245 } 246 247 CallInst *IRBuilderBase::CreateElementUnorderedAtomicMemCpy( 248 Value *Dst, Align DstAlign, Value *Src, Align SrcAlign, Value *Size, 249 uint32_t ElementSize, MDNode *TBAATag, MDNode *TBAAStructTag, 250 MDNode *ScopeTag, MDNode *NoAliasTag) { 251 assert(DstAlign >= ElementSize && 252 "Pointer alignment must be at least element size"); 253 assert(SrcAlign >= ElementSize && 254 "Pointer alignment must be at least element size"); 255 Value *Ops[] = {Dst, Src, Size, getInt32(ElementSize)}; 256 Type *Tys[] = {Dst->getType(), Src->getType(), Size->getType()}; 257 258 CallInst *CI = 259 CreateIntrinsic(Intrinsic::memcpy_element_unordered_atomic, Tys, Ops); 260 261 // Set the alignment of the pointer args. 262 auto *AMCI = cast<AtomicMemCpyInst>(CI); 263 AMCI->setDestAlignment(DstAlign); 264 AMCI->setSourceAlignment(SrcAlign); 265 266 // Set the TBAA info if present. 267 if (TBAATag) 268 CI->setMetadata(LLVMContext::MD_tbaa, TBAATag); 269 270 // Set the TBAA Struct info if present. 271 if (TBAAStructTag) 272 CI->setMetadata(LLVMContext::MD_tbaa_struct, TBAAStructTag); 273 274 if (ScopeTag) 275 CI->setMetadata(LLVMContext::MD_alias_scope, ScopeTag); 276 277 if (NoAliasTag) 278 CI->setMetadata(LLVMContext::MD_noalias, NoAliasTag); 279 280 return CI; 281 } 282 283 /// isConstantOne - Return true only if val is constant int 1 284 static bool isConstantOne(const Value *Val) { 285 assert(Val && "isConstantOne does not work with nullptr Val"); 286 const ConstantInt *CVal = dyn_cast<ConstantInt>(Val); 287 return CVal && CVal->isOne(); 288 } 289 290 CallInst *IRBuilderBase::CreateMalloc(Type *IntPtrTy, Type *AllocTy, 291 Value *AllocSize, Value *ArraySize, 292 ArrayRef<OperandBundleDef> OpB, 293 Function *MallocF, const Twine &Name) { 294 // malloc(type) becomes: 295 // i8* malloc(typeSize) 296 // malloc(type, arraySize) becomes: 297 // i8* malloc(typeSize*arraySize) 298 if (!ArraySize) 299 ArraySize = ConstantInt::get(IntPtrTy, 1); 300 else if (ArraySize->getType() != IntPtrTy) 301 ArraySize = CreateIntCast(ArraySize, IntPtrTy, false); 302 303 if (!isConstantOne(ArraySize)) { 304 if (isConstantOne(AllocSize)) { 305 AllocSize = ArraySize; // Operand * 1 = Operand 306 } else { 307 // Multiply type size by the array size... 308 AllocSize = CreateMul(ArraySize, AllocSize, "mallocsize"); 309 } 310 } 311 312 assert(AllocSize->getType() == IntPtrTy && "malloc arg is wrong size"); 313 // Create the call to Malloc. 314 Module *M = BB->getParent()->getParent(); 315 Type *BPTy = PointerType::getUnqual(Context); 316 FunctionCallee MallocFunc = MallocF; 317 if (!MallocFunc) 318 // prototype malloc as "void *malloc(size_t)" 319 MallocFunc = M->getOrInsertFunction("malloc", BPTy, IntPtrTy); 320 CallInst *MCall = CreateCall(MallocFunc, AllocSize, OpB, Name); 321 322 MCall->setTailCall(); 323 if (Function *F = dyn_cast<Function>(MallocFunc.getCallee())) { 324 MCall->setCallingConv(F->getCallingConv()); 325 F->setReturnDoesNotAlias(); 326 } 327 328 assert(!MCall->getType()->isVoidTy() && "Malloc has void return type"); 329 330 return MCall; 331 } 332 333 CallInst *IRBuilderBase::CreateMalloc(Type *IntPtrTy, Type *AllocTy, 334 Value *AllocSize, Value *ArraySize, 335 Function *MallocF, const Twine &Name) { 336 337 return CreateMalloc(IntPtrTy, AllocTy, AllocSize, ArraySize, {}, MallocF, 338 Name); 339 } 340 341 /// CreateFree - Generate the IR for a call to the builtin free function. 342 CallInst *IRBuilderBase::CreateFree(Value *Source, 343 ArrayRef<OperandBundleDef> Bundles) { 344 assert(Source->getType()->isPointerTy() && 345 "Can not free something of nonpointer type!"); 346 347 Module *M = BB->getParent()->getParent(); 348 349 Type *VoidTy = Type::getVoidTy(M->getContext()); 350 Type *VoidPtrTy = PointerType::getUnqual(M->getContext()); 351 // prototype free as "void free(void*)" 352 FunctionCallee FreeFunc = M->getOrInsertFunction("free", VoidTy, VoidPtrTy); 353 CallInst *Result = CreateCall(FreeFunc, Source, Bundles, ""); 354 Result->setTailCall(); 355 if (Function *F = dyn_cast<Function>(FreeFunc.getCallee())) 356 Result->setCallingConv(F->getCallingConv()); 357 358 return Result; 359 } 360 361 CallInst *IRBuilderBase::CreateElementUnorderedAtomicMemMove( 362 Value *Dst, Align DstAlign, Value *Src, Align SrcAlign, Value *Size, 363 uint32_t ElementSize, MDNode *TBAATag, MDNode *TBAAStructTag, 364 MDNode *ScopeTag, MDNode *NoAliasTag) { 365 assert(DstAlign >= ElementSize && 366 "Pointer alignment must be at least element size"); 367 assert(SrcAlign >= ElementSize && 368 "Pointer alignment must be at least element size"); 369 Value *Ops[] = {Dst, Src, Size, getInt32(ElementSize)}; 370 Type *Tys[] = {Dst->getType(), Src->getType(), Size->getType()}; 371 372 CallInst *CI = 373 CreateIntrinsic(Intrinsic::memmove_element_unordered_atomic, Tys, Ops); 374 375 // Set the alignment of the pointer args. 376 CI->addParamAttr(0, Attribute::getWithAlignment(CI->getContext(), DstAlign)); 377 CI->addParamAttr(1, Attribute::getWithAlignment(CI->getContext(), SrcAlign)); 378 379 // Set the TBAA info if present. 380 if (TBAATag) 381 CI->setMetadata(LLVMContext::MD_tbaa, TBAATag); 382 383 // Set the TBAA Struct info if present. 384 if (TBAAStructTag) 385 CI->setMetadata(LLVMContext::MD_tbaa_struct, TBAAStructTag); 386 387 if (ScopeTag) 388 CI->setMetadata(LLVMContext::MD_alias_scope, ScopeTag); 389 390 if (NoAliasTag) 391 CI->setMetadata(LLVMContext::MD_noalias, NoAliasTag); 392 393 return CI; 394 } 395 396 CallInst *IRBuilderBase::getReductionIntrinsic(Intrinsic::ID ID, Value *Src) { 397 Value *Ops[] = {Src}; 398 Type *Tys[] = { Src->getType() }; 399 return CreateIntrinsic(ID, Tys, Ops); 400 } 401 402 CallInst *IRBuilderBase::CreateFAddReduce(Value *Acc, Value *Src) { 403 Value *Ops[] = {Acc, Src}; 404 return CreateIntrinsic(Intrinsic::vector_reduce_fadd, {Src->getType()}, Ops); 405 } 406 407 CallInst *IRBuilderBase::CreateFMulReduce(Value *Acc, Value *Src) { 408 Value *Ops[] = {Acc, Src}; 409 return CreateIntrinsic(Intrinsic::vector_reduce_fmul, {Src->getType()}, Ops); 410 } 411 412 CallInst *IRBuilderBase::CreateAddReduce(Value *Src) { 413 return getReductionIntrinsic(Intrinsic::vector_reduce_add, Src); 414 } 415 416 CallInst *IRBuilderBase::CreateMulReduce(Value *Src) { 417 return getReductionIntrinsic(Intrinsic::vector_reduce_mul, Src); 418 } 419 420 CallInst *IRBuilderBase::CreateAndReduce(Value *Src) { 421 return getReductionIntrinsic(Intrinsic::vector_reduce_and, Src); 422 } 423 424 CallInst *IRBuilderBase::CreateOrReduce(Value *Src) { 425 return getReductionIntrinsic(Intrinsic::vector_reduce_or, Src); 426 } 427 428 CallInst *IRBuilderBase::CreateXorReduce(Value *Src) { 429 return getReductionIntrinsic(Intrinsic::vector_reduce_xor, Src); 430 } 431 432 CallInst *IRBuilderBase::CreateIntMaxReduce(Value *Src, bool IsSigned) { 433 auto ID = 434 IsSigned ? Intrinsic::vector_reduce_smax : Intrinsic::vector_reduce_umax; 435 return getReductionIntrinsic(ID, Src); 436 } 437 438 CallInst *IRBuilderBase::CreateIntMinReduce(Value *Src, bool IsSigned) { 439 auto ID = 440 IsSigned ? Intrinsic::vector_reduce_smin : Intrinsic::vector_reduce_umin; 441 return getReductionIntrinsic(ID, Src); 442 } 443 444 CallInst *IRBuilderBase::CreateFPMaxReduce(Value *Src) { 445 return getReductionIntrinsic(Intrinsic::vector_reduce_fmax, Src); 446 } 447 448 CallInst *IRBuilderBase::CreateFPMinReduce(Value *Src) { 449 return getReductionIntrinsic(Intrinsic::vector_reduce_fmin, Src); 450 } 451 452 CallInst *IRBuilderBase::CreateFPMaximumReduce(Value *Src) { 453 return getReductionIntrinsic(Intrinsic::vector_reduce_fmaximum, Src); 454 } 455 456 CallInst *IRBuilderBase::CreateFPMinimumReduce(Value *Src) { 457 return getReductionIntrinsic(Intrinsic::vector_reduce_fminimum, Src); 458 } 459 460 CallInst *IRBuilderBase::CreateLifetimeStart(Value *Ptr, ConstantInt *Size) { 461 assert(isa<PointerType>(Ptr->getType()) && 462 "lifetime.start only applies to pointers."); 463 if (!Size) 464 Size = getInt64(-1); 465 else 466 assert(Size->getType() == getInt64Ty() && 467 "lifetime.start requires the size to be an i64"); 468 Value *Ops[] = { Size, Ptr }; 469 return CreateIntrinsic(Intrinsic::lifetime_start, {Ptr->getType()}, Ops); 470 } 471 472 CallInst *IRBuilderBase::CreateLifetimeEnd(Value *Ptr, ConstantInt *Size) { 473 assert(isa<PointerType>(Ptr->getType()) && 474 "lifetime.end only applies to pointers."); 475 if (!Size) 476 Size = getInt64(-1); 477 else 478 assert(Size->getType() == getInt64Ty() && 479 "lifetime.end requires the size to be an i64"); 480 Value *Ops[] = { Size, Ptr }; 481 return CreateIntrinsic(Intrinsic::lifetime_end, {Ptr->getType()}, Ops); 482 } 483 484 CallInst *IRBuilderBase::CreateInvariantStart(Value *Ptr, ConstantInt *Size) { 485 486 assert(isa<PointerType>(Ptr->getType()) && 487 "invariant.start only applies to pointers."); 488 if (!Size) 489 Size = getInt64(-1); 490 else 491 assert(Size->getType() == getInt64Ty() && 492 "invariant.start requires the size to be an i64"); 493 494 Value *Ops[] = {Size, Ptr}; 495 // Fill in the single overloaded type: memory object type. 496 Type *ObjectPtr[1] = {Ptr->getType()}; 497 return CreateIntrinsic(Intrinsic::invariant_start, ObjectPtr, Ops); 498 } 499 500 static MaybeAlign getAlign(Value *Ptr) { 501 if (auto *O = dyn_cast<GlobalObject>(Ptr)) 502 return O->getAlign(); 503 if (auto *A = dyn_cast<GlobalAlias>(Ptr)) 504 return A->getAliaseeObject()->getAlign(); 505 return {}; 506 } 507 508 CallInst *IRBuilderBase::CreateThreadLocalAddress(Value *Ptr) { 509 assert(isa<GlobalValue>(Ptr) && cast<GlobalValue>(Ptr)->isThreadLocal() && 510 "threadlocal_address only applies to thread local variables."); 511 CallInst *CI = CreateIntrinsic(llvm::Intrinsic::threadlocal_address, 512 {Ptr->getType()}, {Ptr}); 513 if (MaybeAlign A = getAlign(Ptr)) { 514 CI->addParamAttr(0, Attribute::getWithAlignment(CI->getContext(), *A)); 515 CI->addRetAttr(Attribute::getWithAlignment(CI->getContext(), *A)); 516 } 517 return CI; 518 } 519 520 CallInst * 521 IRBuilderBase::CreateAssumption(Value *Cond, 522 ArrayRef<OperandBundleDef> OpBundles) { 523 assert(Cond->getType() == getInt1Ty() && 524 "an assumption condition must be of type i1"); 525 526 Value *Ops[] = { Cond }; 527 Module *M = BB->getParent()->getParent(); 528 Function *FnAssume = Intrinsic::getOrInsertDeclaration(M, Intrinsic::assume); 529 return CreateCall(FnAssume, Ops, OpBundles); 530 } 531 532 Instruction *IRBuilderBase::CreateNoAliasScopeDeclaration(Value *Scope) { 533 return CreateIntrinsic(Intrinsic::experimental_noalias_scope_decl, {}, 534 {Scope}); 535 } 536 537 /// Create a call to a Masked Load intrinsic. 538 /// \p Ty - vector type to load 539 /// \p Ptr - base pointer for the load 540 /// \p Alignment - alignment of the source location 541 /// \p Mask - vector of booleans which indicates what vector lanes should 542 /// be accessed in memory 543 /// \p PassThru - pass-through value that is used to fill the masked-off lanes 544 /// of the result 545 /// \p Name - name of the result variable 546 CallInst *IRBuilderBase::CreateMaskedLoad(Type *Ty, Value *Ptr, Align Alignment, 547 Value *Mask, Value *PassThru, 548 const Twine &Name) { 549 auto *PtrTy = cast<PointerType>(Ptr->getType()); 550 assert(Ty->isVectorTy() && "Type should be vector"); 551 assert(Mask && "Mask should not be all-ones (null)"); 552 if (!PassThru) 553 PassThru = PoisonValue::get(Ty); 554 Type *OverloadedTypes[] = { Ty, PtrTy }; 555 Value *Ops[] = {Ptr, getInt32(Alignment.value()), Mask, PassThru}; 556 return CreateMaskedIntrinsic(Intrinsic::masked_load, Ops, 557 OverloadedTypes, Name); 558 } 559 560 /// Create a call to a Masked Store intrinsic. 561 /// \p Val - data to be stored, 562 /// \p Ptr - base pointer for the store 563 /// \p Alignment - alignment of the destination location 564 /// \p Mask - vector of booleans which indicates what vector lanes should 565 /// be accessed in memory 566 CallInst *IRBuilderBase::CreateMaskedStore(Value *Val, Value *Ptr, 567 Align Alignment, Value *Mask) { 568 auto *PtrTy = cast<PointerType>(Ptr->getType()); 569 Type *DataTy = Val->getType(); 570 assert(DataTy->isVectorTy() && "Val should be a vector"); 571 assert(Mask && "Mask should not be all-ones (null)"); 572 Type *OverloadedTypes[] = { DataTy, PtrTy }; 573 Value *Ops[] = {Val, Ptr, getInt32(Alignment.value()), Mask}; 574 return CreateMaskedIntrinsic(Intrinsic::masked_store, Ops, OverloadedTypes); 575 } 576 577 /// Create a call to a Masked intrinsic, with given intrinsic Id, 578 /// an array of operands - Ops, and an array of overloaded types - 579 /// OverloadedTypes. 580 CallInst *IRBuilderBase::CreateMaskedIntrinsic(Intrinsic::ID Id, 581 ArrayRef<Value *> Ops, 582 ArrayRef<Type *> OverloadedTypes, 583 const Twine &Name) { 584 return CreateIntrinsic(Id, OverloadedTypes, Ops, {}, Name); 585 } 586 587 /// Create a call to a Masked Gather intrinsic. 588 /// \p Ty - vector type to gather 589 /// \p Ptrs - vector of pointers for loading 590 /// \p Align - alignment for one element 591 /// \p Mask - vector of booleans which indicates what vector lanes should 592 /// be accessed in memory 593 /// \p PassThru - pass-through value that is used to fill the masked-off lanes 594 /// of the result 595 /// \p Name - name of the result variable 596 CallInst *IRBuilderBase::CreateMaskedGather(Type *Ty, Value *Ptrs, 597 Align Alignment, Value *Mask, 598 Value *PassThru, 599 const Twine &Name) { 600 auto *VecTy = cast<VectorType>(Ty); 601 ElementCount NumElts = VecTy->getElementCount(); 602 auto *PtrsTy = cast<VectorType>(Ptrs->getType()); 603 assert(NumElts == PtrsTy->getElementCount() && "Element count mismatch"); 604 605 if (!Mask) 606 Mask = getAllOnesMask(NumElts); 607 608 if (!PassThru) 609 PassThru = PoisonValue::get(Ty); 610 611 Type *OverloadedTypes[] = {Ty, PtrsTy}; 612 Value *Ops[] = {Ptrs, getInt32(Alignment.value()), Mask, PassThru}; 613 614 // We specify only one type when we create this intrinsic. Types of other 615 // arguments are derived from this type. 616 return CreateMaskedIntrinsic(Intrinsic::masked_gather, Ops, OverloadedTypes, 617 Name); 618 } 619 620 /// Create a call to a Masked Scatter intrinsic. 621 /// \p Data - data to be stored, 622 /// \p Ptrs - the vector of pointers, where the \p Data elements should be 623 /// stored 624 /// \p Align - alignment for one element 625 /// \p Mask - vector of booleans which indicates what vector lanes should 626 /// be accessed in memory 627 CallInst *IRBuilderBase::CreateMaskedScatter(Value *Data, Value *Ptrs, 628 Align Alignment, Value *Mask) { 629 auto *PtrsTy = cast<VectorType>(Ptrs->getType()); 630 auto *DataTy = cast<VectorType>(Data->getType()); 631 ElementCount NumElts = PtrsTy->getElementCount(); 632 633 if (!Mask) 634 Mask = getAllOnesMask(NumElts); 635 636 Type *OverloadedTypes[] = {DataTy, PtrsTy}; 637 Value *Ops[] = {Data, Ptrs, getInt32(Alignment.value()), Mask}; 638 639 // We specify only one type when we create this intrinsic. Types of other 640 // arguments are derived from this type. 641 return CreateMaskedIntrinsic(Intrinsic::masked_scatter, Ops, OverloadedTypes); 642 } 643 644 /// Create a call to Masked Expand Load intrinsic 645 /// \p Ty - vector type to load 646 /// \p Ptr - base pointer for the load 647 /// \p Align - alignment of \p Ptr 648 /// \p Mask - vector of booleans which indicates what vector lanes should 649 /// be accessed in memory 650 /// \p PassThru - pass-through value that is used to fill the masked-off lanes 651 /// of the result 652 /// \p Name - name of the result variable 653 CallInst *IRBuilderBase::CreateMaskedExpandLoad(Type *Ty, Value *Ptr, 654 MaybeAlign Align, Value *Mask, 655 Value *PassThru, 656 const Twine &Name) { 657 assert(Ty->isVectorTy() && "Type should be vector"); 658 assert(Mask && "Mask should not be all-ones (null)"); 659 if (!PassThru) 660 PassThru = PoisonValue::get(Ty); 661 Type *OverloadedTypes[] = {Ty}; 662 Value *Ops[] = {Ptr, Mask, PassThru}; 663 CallInst *CI = CreateMaskedIntrinsic(Intrinsic::masked_expandload, Ops, 664 OverloadedTypes, Name); 665 if (Align) 666 CI->addParamAttr(0, Attribute::getWithAlignment(CI->getContext(), *Align)); 667 return CI; 668 } 669 670 /// Create a call to Masked Compress Store intrinsic 671 /// \p Val - data to be stored, 672 /// \p Ptr - base pointer for the store 673 /// \p Align - alignment of \p Ptr 674 /// \p Mask - vector of booleans which indicates what vector lanes should 675 /// be accessed in memory 676 CallInst *IRBuilderBase::CreateMaskedCompressStore(Value *Val, Value *Ptr, 677 MaybeAlign Align, 678 Value *Mask) { 679 Type *DataTy = Val->getType(); 680 assert(DataTy->isVectorTy() && "Val should be a vector"); 681 assert(Mask && "Mask should not be all-ones (null)"); 682 Type *OverloadedTypes[] = {DataTy}; 683 Value *Ops[] = {Val, Ptr, Mask}; 684 CallInst *CI = CreateMaskedIntrinsic(Intrinsic::masked_compressstore, Ops, 685 OverloadedTypes); 686 if (Align) 687 CI->addParamAttr(1, Attribute::getWithAlignment(CI->getContext(), *Align)); 688 return CI; 689 } 690 691 template <typename T0> 692 static std::vector<Value *> 693 getStatepointArgs(IRBuilderBase &B, uint64_t ID, uint32_t NumPatchBytes, 694 Value *ActualCallee, uint32_t Flags, ArrayRef<T0> CallArgs) { 695 std::vector<Value *> Args; 696 Args.push_back(B.getInt64(ID)); 697 Args.push_back(B.getInt32(NumPatchBytes)); 698 Args.push_back(ActualCallee); 699 Args.push_back(B.getInt32(CallArgs.size())); 700 Args.push_back(B.getInt32(Flags)); 701 llvm::append_range(Args, CallArgs); 702 // GC Transition and Deopt args are now always handled via operand bundle. 703 // They will be removed from the signature of gc.statepoint shortly. 704 Args.push_back(B.getInt32(0)); 705 Args.push_back(B.getInt32(0)); 706 // GC args are now encoded in the gc-live operand bundle 707 return Args; 708 } 709 710 template<typename T1, typename T2, typename T3> 711 static std::vector<OperandBundleDef> 712 getStatepointBundles(std::optional<ArrayRef<T1>> TransitionArgs, 713 std::optional<ArrayRef<T2>> DeoptArgs, 714 ArrayRef<T3> GCArgs) { 715 std::vector<OperandBundleDef> Rval; 716 if (DeoptArgs) { 717 SmallVector<Value*, 16> DeoptValues; 718 llvm::append_range(DeoptValues, *DeoptArgs); 719 Rval.emplace_back("deopt", DeoptValues); 720 } 721 if (TransitionArgs) { 722 SmallVector<Value*, 16> TransitionValues; 723 llvm::append_range(TransitionValues, *TransitionArgs); 724 Rval.emplace_back("gc-transition", TransitionValues); 725 } 726 if (GCArgs.size()) { 727 SmallVector<Value*, 16> LiveValues; 728 llvm::append_range(LiveValues, GCArgs); 729 Rval.emplace_back("gc-live", LiveValues); 730 } 731 return Rval; 732 } 733 734 template <typename T0, typename T1, typename T2, typename T3> 735 static CallInst *CreateGCStatepointCallCommon( 736 IRBuilderBase *Builder, uint64_t ID, uint32_t NumPatchBytes, 737 FunctionCallee ActualCallee, uint32_t Flags, ArrayRef<T0> CallArgs, 738 std::optional<ArrayRef<T1>> TransitionArgs, 739 std::optional<ArrayRef<T2>> DeoptArgs, ArrayRef<T3> GCArgs, 740 const Twine &Name) { 741 Module *M = Builder->GetInsertBlock()->getParent()->getParent(); 742 // Fill in the one generic type'd argument (the function is also vararg) 743 Function *FnStatepoint = Intrinsic::getOrInsertDeclaration( 744 M, Intrinsic::experimental_gc_statepoint, 745 {ActualCallee.getCallee()->getType()}); 746 747 std::vector<Value *> Args = getStatepointArgs( 748 *Builder, ID, NumPatchBytes, ActualCallee.getCallee(), Flags, CallArgs); 749 750 CallInst *CI = Builder->CreateCall( 751 FnStatepoint, Args, 752 getStatepointBundles(TransitionArgs, DeoptArgs, GCArgs), Name); 753 CI->addParamAttr(2, 754 Attribute::get(Builder->getContext(), Attribute::ElementType, 755 ActualCallee.getFunctionType())); 756 return CI; 757 } 758 759 CallInst *IRBuilderBase::CreateGCStatepointCall( 760 uint64_t ID, uint32_t NumPatchBytes, FunctionCallee ActualCallee, 761 ArrayRef<Value *> CallArgs, std::optional<ArrayRef<Value *>> DeoptArgs, 762 ArrayRef<Value *> GCArgs, const Twine &Name) { 763 return CreateGCStatepointCallCommon<Value *, Value *, Value *, Value *>( 764 this, ID, NumPatchBytes, ActualCallee, uint32_t(StatepointFlags::None), 765 CallArgs, std::nullopt /* No Transition Args */, DeoptArgs, GCArgs, Name); 766 } 767 768 CallInst *IRBuilderBase::CreateGCStatepointCall( 769 uint64_t ID, uint32_t NumPatchBytes, FunctionCallee ActualCallee, 770 uint32_t Flags, ArrayRef<Value *> CallArgs, 771 std::optional<ArrayRef<Use>> TransitionArgs, 772 std::optional<ArrayRef<Use>> DeoptArgs, ArrayRef<Value *> GCArgs, 773 const Twine &Name) { 774 return CreateGCStatepointCallCommon<Value *, Use, Use, Value *>( 775 this, ID, NumPatchBytes, ActualCallee, Flags, CallArgs, TransitionArgs, 776 DeoptArgs, GCArgs, Name); 777 } 778 779 CallInst *IRBuilderBase::CreateGCStatepointCall( 780 uint64_t ID, uint32_t NumPatchBytes, FunctionCallee ActualCallee, 781 ArrayRef<Use> CallArgs, std::optional<ArrayRef<Value *>> DeoptArgs, 782 ArrayRef<Value *> GCArgs, const Twine &Name) { 783 return CreateGCStatepointCallCommon<Use, Value *, Value *, Value *>( 784 this, ID, NumPatchBytes, ActualCallee, uint32_t(StatepointFlags::None), 785 CallArgs, std::nullopt, DeoptArgs, GCArgs, Name); 786 } 787 788 template <typename T0, typename T1, typename T2, typename T3> 789 static InvokeInst *CreateGCStatepointInvokeCommon( 790 IRBuilderBase *Builder, uint64_t ID, uint32_t NumPatchBytes, 791 FunctionCallee ActualInvokee, BasicBlock *NormalDest, 792 BasicBlock *UnwindDest, uint32_t Flags, ArrayRef<T0> InvokeArgs, 793 std::optional<ArrayRef<T1>> TransitionArgs, 794 std::optional<ArrayRef<T2>> DeoptArgs, ArrayRef<T3> GCArgs, 795 const Twine &Name) { 796 Module *M = Builder->GetInsertBlock()->getParent()->getParent(); 797 // Fill in the one generic type'd argument (the function is also vararg) 798 Function *FnStatepoint = Intrinsic::getOrInsertDeclaration( 799 M, Intrinsic::experimental_gc_statepoint, 800 {ActualInvokee.getCallee()->getType()}); 801 802 std::vector<Value *> Args = 803 getStatepointArgs(*Builder, ID, NumPatchBytes, ActualInvokee.getCallee(), 804 Flags, InvokeArgs); 805 806 InvokeInst *II = Builder->CreateInvoke( 807 FnStatepoint, NormalDest, UnwindDest, Args, 808 getStatepointBundles(TransitionArgs, DeoptArgs, GCArgs), Name); 809 II->addParamAttr(2, 810 Attribute::get(Builder->getContext(), Attribute::ElementType, 811 ActualInvokee.getFunctionType())); 812 return II; 813 } 814 815 InvokeInst *IRBuilderBase::CreateGCStatepointInvoke( 816 uint64_t ID, uint32_t NumPatchBytes, FunctionCallee ActualInvokee, 817 BasicBlock *NormalDest, BasicBlock *UnwindDest, 818 ArrayRef<Value *> InvokeArgs, std::optional<ArrayRef<Value *>> DeoptArgs, 819 ArrayRef<Value *> GCArgs, const Twine &Name) { 820 return CreateGCStatepointInvokeCommon<Value *, Value *, Value *, Value *>( 821 this, ID, NumPatchBytes, ActualInvokee, NormalDest, UnwindDest, 822 uint32_t(StatepointFlags::None), InvokeArgs, 823 std::nullopt /* No Transition Args*/, DeoptArgs, GCArgs, Name); 824 } 825 826 InvokeInst *IRBuilderBase::CreateGCStatepointInvoke( 827 uint64_t ID, uint32_t NumPatchBytes, FunctionCallee ActualInvokee, 828 BasicBlock *NormalDest, BasicBlock *UnwindDest, uint32_t Flags, 829 ArrayRef<Value *> InvokeArgs, std::optional<ArrayRef<Use>> TransitionArgs, 830 std::optional<ArrayRef<Use>> DeoptArgs, ArrayRef<Value *> GCArgs, 831 const Twine &Name) { 832 return CreateGCStatepointInvokeCommon<Value *, Use, Use, Value *>( 833 this, ID, NumPatchBytes, ActualInvokee, NormalDest, UnwindDest, Flags, 834 InvokeArgs, TransitionArgs, DeoptArgs, GCArgs, Name); 835 } 836 837 InvokeInst *IRBuilderBase::CreateGCStatepointInvoke( 838 uint64_t ID, uint32_t NumPatchBytes, FunctionCallee ActualInvokee, 839 BasicBlock *NormalDest, BasicBlock *UnwindDest, ArrayRef<Use> InvokeArgs, 840 std::optional<ArrayRef<Value *>> DeoptArgs, ArrayRef<Value *> GCArgs, 841 const Twine &Name) { 842 return CreateGCStatepointInvokeCommon<Use, Value *, Value *, Value *>( 843 this, ID, NumPatchBytes, ActualInvokee, NormalDest, UnwindDest, 844 uint32_t(StatepointFlags::None), InvokeArgs, std::nullopt, DeoptArgs, 845 GCArgs, Name); 846 } 847 848 CallInst *IRBuilderBase::CreateGCResult(Instruction *Statepoint, 849 Type *ResultType, const Twine &Name) { 850 Intrinsic::ID ID = Intrinsic::experimental_gc_result; 851 Type *Types[] = {ResultType}; 852 853 Value *Args[] = {Statepoint}; 854 return CreateIntrinsic(ID, Types, Args, {}, Name); 855 } 856 857 CallInst *IRBuilderBase::CreateGCRelocate(Instruction *Statepoint, 858 int BaseOffset, int DerivedOffset, 859 Type *ResultType, const Twine &Name) { 860 Type *Types[] = {ResultType}; 861 862 Value *Args[] = {Statepoint, getInt32(BaseOffset), getInt32(DerivedOffset)}; 863 return CreateIntrinsic(Intrinsic::experimental_gc_relocate, Types, Args, {}, 864 Name); 865 } 866 867 CallInst *IRBuilderBase::CreateGCGetPointerBase(Value *DerivedPtr, 868 const Twine &Name) { 869 Type *PtrTy = DerivedPtr->getType(); 870 return CreateIntrinsic(Intrinsic::experimental_gc_get_pointer_base, 871 {PtrTy, PtrTy}, {DerivedPtr}, {}, Name); 872 } 873 874 CallInst *IRBuilderBase::CreateGCGetPointerOffset(Value *DerivedPtr, 875 const Twine &Name) { 876 Type *PtrTy = DerivedPtr->getType(); 877 return CreateIntrinsic(Intrinsic::experimental_gc_get_pointer_offset, {PtrTy}, 878 {DerivedPtr}, {}, Name); 879 } 880 881 CallInst *IRBuilderBase::CreateUnaryIntrinsic(Intrinsic::ID ID, Value *V, 882 FMFSource FMFSource, 883 const Twine &Name) { 884 Module *M = BB->getModule(); 885 Function *Fn = Intrinsic::getOrInsertDeclaration(M, ID, {V->getType()}); 886 return createCallHelper(Fn, {V}, Name, FMFSource); 887 } 888 889 Value *IRBuilderBase::CreateBinaryIntrinsic(Intrinsic::ID ID, Value *LHS, 890 Value *RHS, FMFSource FMFSource, 891 const Twine &Name) { 892 Module *M = BB->getModule(); 893 Function *Fn = Intrinsic::getOrInsertDeclaration(M, ID, {LHS->getType()}); 894 if (Value *V = Folder.FoldBinaryIntrinsic(ID, LHS, RHS, Fn->getReturnType(), 895 /*FMFSource=*/nullptr)) 896 return V; 897 return createCallHelper(Fn, {LHS, RHS}, Name, FMFSource); 898 } 899 900 CallInst *IRBuilderBase::CreateIntrinsic(Intrinsic::ID ID, 901 ArrayRef<Type *> Types, 902 ArrayRef<Value *> Args, 903 FMFSource FMFSource, 904 const Twine &Name) { 905 Module *M = BB->getModule(); 906 Function *Fn = Intrinsic::getOrInsertDeclaration(M, ID, Types); 907 return createCallHelper(Fn, Args, Name, FMFSource); 908 } 909 910 CallInst *IRBuilderBase::CreateIntrinsic(Type *RetTy, Intrinsic::ID ID, 911 ArrayRef<Value *> Args, 912 FMFSource FMFSource, 913 const Twine &Name) { 914 Module *M = BB->getModule(); 915 916 SmallVector<Intrinsic::IITDescriptor> Table; 917 Intrinsic::getIntrinsicInfoTableEntries(ID, Table); 918 ArrayRef<Intrinsic::IITDescriptor> TableRef(Table); 919 920 SmallVector<Type *> ArgTys; 921 ArgTys.reserve(Args.size()); 922 for (auto &I : Args) 923 ArgTys.push_back(I->getType()); 924 FunctionType *FTy = FunctionType::get(RetTy, ArgTys, false); 925 SmallVector<Type *> OverloadTys; 926 Intrinsic::MatchIntrinsicTypesResult Res = 927 matchIntrinsicSignature(FTy, TableRef, OverloadTys); 928 (void)Res; 929 assert(Res == Intrinsic::MatchIntrinsicTypes_Match && TableRef.empty() && 930 "Wrong types for intrinsic!"); 931 // TODO: Handle varargs intrinsics. 932 933 Function *Fn = Intrinsic::getOrInsertDeclaration(M, ID, OverloadTys); 934 return createCallHelper(Fn, Args, Name, FMFSource); 935 } 936 937 CallInst *IRBuilderBase::CreateConstrainedFPBinOp( 938 Intrinsic::ID ID, Value *L, Value *R, FMFSource FMFSource, 939 const Twine &Name, MDNode *FPMathTag, std::optional<RoundingMode> Rounding, 940 std::optional<fp::ExceptionBehavior> Except) { 941 Value *RoundingV = getConstrainedFPRounding(Rounding); 942 Value *ExceptV = getConstrainedFPExcept(Except); 943 944 FastMathFlags UseFMF = FMFSource.get(FMF); 945 946 CallInst *C = CreateIntrinsic(ID, {L->getType()}, 947 {L, R, RoundingV, ExceptV}, nullptr, Name); 948 setConstrainedFPCallAttr(C); 949 setFPAttrs(C, FPMathTag, UseFMF); 950 return C; 951 } 952 953 CallInst *IRBuilderBase::CreateConstrainedFPUnroundedBinOp( 954 Intrinsic::ID ID, Value *L, Value *R, FMFSource FMFSource, 955 const Twine &Name, MDNode *FPMathTag, 956 std::optional<fp::ExceptionBehavior> Except) { 957 Value *ExceptV = getConstrainedFPExcept(Except); 958 959 FastMathFlags UseFMF = FMFSource.get(FMF); 960 961 CallInst *C = 962 CreateIntrinsic(ID, {L->getType()}, {L, R, ExceptV}, nullptr, Name); 963 setConstrainedFPCallAttr(C); 964 setFPAttrs(C, FPMathTag, UseFMF); 965 return C; 966 } 967 968 Value *IRBuilderBase::CreateNAryOp(unsigned Opc, ArrayRef<Value *> Ops, 969 const Twine &Name, MDNode *FPMathTag) { 970 if (Instruction::isBinaryOp(Opc)) { 971 assert(Ops.size() == 2 && "Invalid number of operands!"); 972 return CreateBinOp(static_cast<Instruction::BinaryOps>(Opc), 973 Ops[0], Ops[1], Name, FPMathTag); 974 } 975 if (Instruction::isUnaryOp(Opc)) { 976 assert(Ops.size() == 1 && "Invalid number of operands!"); 977 return CreateUnOp(static_cast<Instruction::UnaryOps>(Opc), 978 Ops[0], Name, FPMathTag); 979 } 980 llvm_unreachable("Unexpected opcode!"); 981 } 982 983 CallInst *IRBuilderBase::CreateConstrainedFPCast( 984 Intrinsic::ID ID, Value *V, Type *DestTy, FMFSource FMFSource, 985 const Twine &Name, MDNode *FPMathTag, std::optional<RoundingMode> Rounding, 986 std::optional<fp::ExceptionBehavior> Except) { 987 Value *ExceptV = getConstrainedFPExcept(Except); 988 989 FastMathFlags UseFMF = FMFSource.get(FMF); 990 991 CallInst *C; 992 if (Intrinsic::hasConstrainedFPRoundingModeOperand(ID)) { 993 Value *RoundingV = getConstrainedFPRounding(Rounding); 994 C = CreateIntrinsic(ID, {DestTy, V->getType()}, {V, RoundingV, ExceptV}, 995 nullptr, Name); 996 } else 997 C = CreateIntrinsic(ID, {DestTy, V->getType()}, {V, ExceptV}, nullptr, 998 Name); 999 1000 setConstrainedFPCallAttr(C); 1001 1002 if (isa<FPMathOperator>(C)) 1003 setFPAttrs(C, FPMathTag, UseFMF); 1004 return C; 1005 } 1006 1007 Value *IRBuilderBase::CreateFCmpHelper(CmpInst::Predicate P, Value *LHS, 1008 Value *RHS, const Twine &Name, 1009 MDNode *FPMathTag, FMFSource FMFSource, 1010 bool IsSignaling) { 1011 if (IsFPConstrained) { 1012 auto ID = IsSignaling ? Intrinsic::experimental_constrained_fcmps 1013 : Intrinsic::experimental_constrained_fcmp; 1014 return CreateConstrainedFPCmp(ID, P, LHS, RHS, Name); 1015 } 1016 1017 if (auto *V = Folder.FoldCmp(P, LHS, RHS)) 1018 return V; 1019 return Insert( 1020 setFPAttrs(new FCmpInst(P, LHS, RHS), FPMathTag, FMFSource.get(FMF)), 1021 Name); 1022 } 1023 1024 CallInst *IRBuilderBase::CreateConstrainedFPCmp( 1025 Intrinsic::ID ID, CmpInst::Predicate P, Value *L, Value *R, 1026 const Twine &Name, std::optional<fp::ExceptionBehavior> Except) { 1027 Value *PredicateV = getConstrainedFPPredicate(P); 1028 Value *ExceptV = getConstrainedFPExcept(Except); 1029 1030 CallInst *C = CreateIntrinsic(ID, {L->getType()}, 1031 {L, R, PredicateV, ExceptV}, nullptr, Name); 1032 setConstrainedFPCallAttr(C); 1033 return C; 1034 } 1035 1036 CallInst *IRBuilderBase::CreateConstrainedFPCall( 1037 Function *Callee, ArrayRef<Value *> Args, const Twine &Name, 1038 std::optional<RoundingMode> Rounding, 1039 std::optional<fp::ExceptionBehavior> Except) { 1040 llvm::SmallVector<Value *, 6> UseArgs; 1041 1042 append_range(UseArgs, Args); 1043 1044 if (Intrinsic::hasConstrainedFPRoundingModeOperand(Callee->getIntrinsicID())) 1045 UseArgs.push_back(getConstrainedFPRounding(Rounding)); 1046 UseArgs.push_back(getConstrainedFPExcept(Except)); 1047 1048 CallInst *C = CreateCall(Callee, UseArgs, Name); 1049 setConstrainedFPCallAttr(C); 1050 return C; 1051 } 1052 1053 Value *IRBuilderBase::CreateSelect(Value *C, Value *True, Value *False, 1054 const Twine &Name, Instruction *MDFrom) { 1055 return CreateSelectFMF(C, True, False, {}, Name, MDFrom); 1056 } 1057 1058 Value *IRBuilderBase::CreateSelectFMF(Value *C, Value *True, Value *False, 1059 FMFSource FMFSource, const Twine &Name, 1060 Instruction *MDFrom) { 1061 if (auto *V = Folder.FoldSelect(C, True, False)) 1062 return V; 1063 1064 SelectInst *Sel = SelectInst::Create(C, True, False); 1065 if (MDFrom) { 1066 MDNode *Prof = MDFrom->getMetadata(LLVMContext::MD_prof); 1067 MDNode *Unpred = MDFrom->getMetadata(LLVMContext::MD_unpredictable); 1068 Sel = addBranchMetadata(Sel, Prof, Unpred); 1069 } 1070 if (isa<FPMathOperator>(Sel)) 1071 setFPAttrs(Sel, /*MDNode=*/nullptr, FMFSource.get(FMF)); 1072 return Insert(Sel, Name); 1073 } 1074 1075 Value *IRBuilderBase::CreatePtrDiff(Type *ElemTy, Value *LHS, Value *RHS, 1076 const Twine &Name) { 1077 assert(LHS->getType() == RHS->getType() && 1078 "Pointer subtraction operand types must match!"); 1079 Value *LHS_int = CreatePtrToInt(LHS, Type::getInt64Ty(Context)); 1080 Value *RHS_int = CreatePtrToInt(RHS, Type::getInt64Ty(Context)); 1081 Value *Difference = CreateSub(LHS_int, RHS_int); 1082 return CreateExactSDiv(Difference, ConstantExpr::getSizeOf(ElemTy), 1083 Name); 1084 } 1085 1086 Value *IRBuilderBase::CreateLaunderInvariantGroup(Value *Ptr) { 1087 assert(isa<PointerType>(Ptr->getType()) && 1088 "launder.invariant.group only applies to pointers."); 1089 auto *PtrType = Ptr->getType(); 1090 Module *M = BB->getParent()->getParent(); 1091 Function *FnLaunderInvariantGroup = Intrinsic::getOrInsertDeclaration( 1092 M, Intrinsic::launder_invariant_group, {PtrType}); 1093 1094 assert(FnLaunderInvariantGroup->getReturnType() == PtrType && 1095 FnLaunderInvariantGroup->getFunctionType()->getParamType(0) == 1096 PtrType && 1097 "LaunderInvariantGroup should take and return the same type"); 1098 1099 return CreateCall(FnLaunderInvariantGroup, {Ptr}); 1100 } 1101 1102 Value *IRBuilderBase::CreateStripInvariantGroup(Value *Ptr) { 1103 assert(isa<PointerType>(Ptr->getType()) && 1104 "strip.invariant.group only applies to pointers."); 1105 1106 auto *PtrType = Ptr->getType(); 1107 Module *M = BB->getParent()->getParent(); 1108 Function *FnStripInvariantGroup = Intrinsic::getOrInsertDeclaration( 1109 M, Intrinsic::strip_invariant_group, {PtrType}); 1110 1111 assert(FnStripInvariantGroup->getReturnType() == PtrType && 1112 FnStripInvariantGroup->getFunctionType()->getParamType(0) == 1113 PtrType && 1114 "StripInvariantGroup should take and return the same type"); 1115 1116 return CreateCall(FnStripInvariantGroup, {Ptr}); 1117 } 1118 1119 Value *IRBuilderBase::CreateVectorReverse(Value *V, const Twine &Name) { 1120 auto *Ty = cast<VectorType>(V->getType()); 1121 if (isa<ScalableVectorType>(Ty)) { 1122 Module *M = BB->getParent()->getParent(); 1123 Function *F = 1124 Intrinsic::getOrInsertDeclaration(M, Intrinsic::vector_reverse, Ty); 1125 return Insert(CallInst::Create(F, V), Name); 1126 } 1127 // Keep the original behaviour for fixed vector 1128 SmallVector<int, 8> ShuffleMask; 1129 int NumElts = Ty->getElementCount().getKnownMinValue(); 1130 for (int i = 0; i < NumElts; ++i) 1131 ShuffleMask.push_back(NumElts - i - 1); 1132 return CreateShuffleVector(V, ShuffleMask, Name); 1133 } 1134 1135 Value *IRBuilderBase::CreateVectorSplice(Value *V1, Value *V2, int64_t Imm, 1136 const Twine &Name) { 1137 assert(isa<VectorType>(V1->getType()) && "Unexpected type"); 1138 assert(V1->getType() == V2->getType() && 1139 "Splice expects matching operand types!"); 1140 1141 if (auto *VTy = dyn_cast<ScalableVectorType>(V1->getType())) { 1142 Module *M = BB->getParent()->getParent(); 1143 Function *F = 1144 Intrinsic::getOrInsertDeclaration(M, Intrinsic::vector_splice, VTy); 1145 1146 Value *Ops[] = {V1, V2, getInt32(Imm)}; 1147 return Insert(CallInst::Create(F, Ops), Name); 1148 } 1149 1150 unsigned NumElts = cast<FixedVectorType>(V1->getType())->getNumElements(); 1151 assert(((-Imm <= NumElts) || (Imm < NumElts)) && 1152 "Invalid immediate for vector splice!"); 1153 1154 // Keep the original behaviour for fixed vector 1155 unsigned Idx = (NumElts + Imm) % NumElts; 1156 SmallVector<int, 8> Mask; 1157 for (unsigned I = 0; I < NumElts; ++I) 1158 Mask.push_back(Idx + I); 1159 1160 return CreateShuffleVector(V1, V2, Mask); 1161 } 1162 1163 Value *IRBuilderBase::CreateVectorSplat(unsigned NumElts, Value *V, 1164 const Twine &Name) { 1165 auto EC = ElementCount::getFixed(NumElts); 1166 return CreateVectorSplat(EC, V, Name); 1167 } 1168 1169 Value *IRBuilderBase::CreateVectorSplat(ElementCount EC, Value *V, 1170 const Twine &Name) { 1171 assert(EC.isNonZero() && "Cannot splat to an empty vector!"); 1172 1173 // First insert it into a poison vector so we can shuffle it. 1174 Value *Poison = PoisonValue::get(VectorType::get(V->getType(), EC)); 1175 V = CreateInsertElement(Poison, V, getInt64(0), Name + ".splatinsert"); 1176 1177 // Shuffle the value across the desired number of elements. 1178 SmallVector<int, 16> Zeros; 1179 Zeros.resize(EC.getKnownMinValue()); 1180 return CreateShuffleVector(V, Zeros, Name + ".splat"); 1181 } 1182 1183 Value *IRBuilderBase::CreatePreserveArrayAccessIndex( 1184 Type *ElTy, Value *Base, unsigned Dimension, unsigned LastIndex, 1185 MDNode *DbgInfo) { 1186 auto *BaseType = Base->getType(); 1187 assert(isa<PointerType>(BaseType) && 1188 "Invalid Base ptr type for preserve.array.access.index."); 1189 1190 Value *LastIndexV = getInt32(LastIndex); 1191 Constant *Zero = ConstantInt::get(Type::getInt32Ty(Context), 0); 1192 SmallVector<Value *, 4> IdxList(Dimension, Zero); 1193 IdxList.push_back(LastIndexV); 1194 1195 Type *ResultType = GetElementPtrInst::getGEPReturnType(Base, IdxList); 1196 1197 Value *DimV = getInt32(Dimension); 1198 CallInst *Fn = 1199 CreateIntrinsic(Intrinsic::preserve_array_access_index, 1200 {ResultType, BaseType}, {Base, DimV, LastIndexV}); 1201 Fn->addParamAttr( 1202 0, Attribute::get(Fn->getContext(), Attribute::ElementType, ElTy)); 1203 if (DbgInfo) 1204 Fn->setMetadata(LLVMContext::MD_preserve_access_index, DbgInfo); 1205 1206 return Fn; 1207 } 1208 1209 Value *IRBuilderBase::CreatePreserveUnionAccessIndex( 1210 Value *Base, unsigned FieldIndex, MDNode *DbgInfo) { 1211 assert(isa<PointerType>(Base->getType()) && 1212 "Invalid Base ptr type for preserve.union.access.index."); 1213 auto *BaseType = Base->getType(); 1214 1215 Value *DIIndex = getInt32(FieldIndex); 1216 CallInst *Fn = CreateIntrinsic(Intrinsic::preserve_union_access_index, 1217 {BaseType, BaseType}, {Base, DIIndex}); 1218 if (DbgInfo) 1219 Fn->setMetadata(LLVMContext::MD_preserve_access_index, DbgInfo); 1220 1221 return Fn; 1222 } 1223 1224 Value *IRBuilderBase::CreatePreserveStructAccessIndex( 1225 Type *ElTy, Value *Base, unsigned Index, unsigned FieldIndex, 1226 MDNode *DbgInfo) { 1227 auto *BaseType = Base->getType(); 1228 assert(isa<PointerType>(BaseType) && 1229 "Invalid Base ptr type for preserve.struct.access.index."); 1230 1231 Value *GEPIndex = getInt32(Index); 1232 Constant *Zero = ConstantInt::get(Type::getInt32Ty(Context), 0); 1233 Type *ResultType = 1234 GetElementPtrInst::getGEPReturnType(Base, {Zero, GEPIndex}); 1235 1236 Value *DIIndex = getInt32(FieldIndex); 1237 CallInst *Fn = 1238 CreateIntrinsic(Intrinsic::preserve_struct_access_index, 1239 {ResultType, BaseType}, {Base, GEPIndex, DIIndex}); 1240 Fn->addParamAttr( 1241 0, Attribute::get(Fn->getContext(), Attribute::ElementType, ElTy)); 1242 if (DbgInfo) 1243 Fn->setMetadata(LLVMContext::MD_preserve_access_index, DbgInfo); 1244 1245 return Fn; 1246 } 1247 1248 Value *IRBuilderBase::createIsFPClass(Value *FPNum, unsigned Test) { 1249 ConstantInt *TestV = getInt32(Test); 1250 return CreateIntrinsic(Intrinsic::is_fpclass, {FPNum->getType()}, 1251 {FPNum, TestV}); 1252 } 1253 1254 CallInst *IRBuilderBase::CreateAlignmentAssumptionHelper(const DataLayout &DL, 1255 Value *PtrValue, 1256 Value *AlignValue, 1257 Value *OffsetValue) { 1258 SmallVector<Value *, 4> Vals({PtrValue, AlignValue}); 1259 if (OffsetValue) 1260 Vals.push_back(OffsetValue); 1261 OperandBundleDefT<Value *> AlignOpB("align", Vals); 1262 return CreateAssumption(ConstantInt::getTrue(getContext()), {AlignOpB}); 1263 } 1264 1265 CallInst *IRBuilderBase::CreateAlignmentAssumption(const DataLayout &DL, 1266 Value *PtrValue, 1267 unsigned Alignment, 1268 Value *OffsetValue) { 1269 assert(isa<PointerType>(PtrValue->getType()) && 1270 "trying to create an alignment assumption on a non-pointer?"); 1271 assert(Alignment != 0 && "Invalid Alignment"); 1272 auto *PtrTy = cast<PointerType>(PtrValue->getType()); 1273 Type *IntPtrTy = getIntPtrTy(DL, PtrTy->getAddressSpace()); 1274 Value *AlignValue = ConstantInt::get(IntPtrTy, Alignment); 1275 return CreateAlignmentAssumptionHelper(DL, PtrValue, AlignValue, OffsetValue); 1276 } 1277 1278 CallInst *IRBuilderBase::CreateAlignmentAssumption(const DataLayout &DL, 1279 Value *PtrValue, 1280 Value *Alignment, 1281 Value *OffsetValue) { 1282 assert(isa<PointerType>(PtrValue->getType()) && 1283 "trying to create an alignment assumption on a non-pointer?"); 1284 return CreateAlignmentAssumptionHelper(DL, PtrValue, Alignment, OffsetValue); 1285 } 1286 1287 IRBuilderDefaultInserter::~IRBuilderDefaultInserter() = default; 1288 IRBuilderCallbackInserter::~IRBuilderCallbackInserter() = default; 1289 IRBuilderFolder::~IRBuilderFolder() = default; 1290 void ConstantFolder::anchor() {} 1291 void NoFolder::anchor() {} 1292