/llvm-project/llvm/lib/Target/AMDGPU/Utils/ |
H A D | SIDefinesUtils.h | 28 unsigned Shift = 0; in getShiftMask() local 45 uint32_t Shift, MCContext &Ctx) { in maskShiftSet() 64 uint32_t Shift, MCContext &Ctx) { in maskShiftGet()
|
H A D | AMDKernelCodeTInfo.h | 52 #define PRINTCOMP(GetMacro, Shift) \ argument 60 #define PARSECOMP(SetMacro, Shift) \ argument 72 #define COMPPGM(name, aname, GetMacro, SetMacro, Shift) \ argument
|
H A D | AMDGPUAsmUtils.h | 38 unsigned Shift; member
|
/llvm-project/llvm/lib/Target/AMDGPU/MCTargetDesc/ |
H A D | AMDGPUMCKernelDescriptor.cpp | 83 uint32_t Shift, uint32_t Mask, in bits_set() argument 92 const MCExpr *MCKernelDescriptor::bits_get(const MCExpr *Src, uint32_t Shift, in bits_get() argument
|
/llvm-project/llvm/lib/Support/ |
H A D | ScaledNumber.cpp | 48 int Shift = 64 - LeadingZeros; in multiply64() local 64 int Shift = 0; in divide32() local 86 int Shift = 0; in divide64() local 169 int Shift = 63 - (NewE - E); in toStringAPFloat() local 214 if (int Shift = std::min(int16_t(countLeadingZeros64(D)), E)) { in toString() local
|
/llvm-project/llvm/include/llvm/Support/ |
H A D | LEB128.h | 136 unsigned Shift = 0; variable 170 unsigned Shift = 0; variable
|
H A D | ScaledNumber.h | 88 int Shift = llvm::bit_width(Digits) - Width; variable 701 static ScaledNumber adjustToWidth(uint64_t N, int32_t Shift) { in adjustToWidth() 834 template <class DigitsT> void ScaledNumber<DigitsT>::shiftLeft(int32_t Shift) { in shiftLeft() 864 template <class DigitsT> void ScaledNumber<DigitsT>::shiftRight(int32_t Shift) { in shiftRight()
|
/llvm-project/llvm/include/llvm/ADT/ |
H A D | PointerEmbeddedInt.h | 44 Shift = sizeof(uintptr_t) * CHAR_BIT - Bits, enumerator
|
H A D | bit.h | 160 T Shift = std::numeric_limits<T>::digits >> 1; in count() local
|
/llvm-project/llvm/include/llvm/MC/ |
H A D | MCELFExtras.h | 31 const int Shift = llvm::countr_zero(OffsetMask); in encodeCrel() local
|
/llvm-project/clang/include/clang/Basic/ |
H A D | OperatorPrecedence.h | 39 Shift = 12, // <<, >> enumerator
|
/llvm-project/libc/AOR_v20.02/math/ |
H A D | v_sinf.c | 29 #define Shift v_f32 (0x1.8p+23f) macro
|
H A D | v_cosf.c | 29 #define Shift v_f32 (0x1.8p+23f) macro
|
H A D | v_cos.c | 38 #define Shift v_f64 (0x1.8p52) macro
|
H A D | v_sin.c | 37 #define Shift v_f64 (0x1.8p52) macro
|
H A D | v_expf_1u.c | 27 #define Shift v_f32 (0x1.8p23f) macro
|
H A D | v_exp2f_1u.c | 24 #define Shift v_f32 (0x1.8p23f) macro
|
H A D | v_exp2f.c | 27 #define Shift v_f32 (0x1.8p23f) macro
|
H A D | exp2.c | 15 #define Shift __exp_data.exp2_shift macro
|
H A D | v_expf.c | 27 #define Shift v_f32 (0x1.8p23f) macro
|
H A D | v_exp.c | 39 #define Shift v_f64 (0x1.8p+52) macro
|
H A D | exp.c | 18 #define Shift __exp_data.shift macro
|
/llvm-project/llvm/lib/Target/AArch64/ |
H A D | AArch64ExpandImm.cpp | 490 unsigned Shift = 0; // LSL amount for high bits with MOVZ/MOVN in expandMOVImmSimple() local 541 for (unsigned Shift = 0; Shift < BitSize; Shift += 16) { in expandMOVImm() local 584 for (unsigned Shift = 0; Shift < BitSize; Shift += 16) { in expandMOVImm() local
|
/llvm-project/compiler-rt/lib/gwp_asan/ |
H A D | stack_trace_compressor.cpp | 37 uint8_t Shift = 0; in varIntDecode() local
|
/llvm-project/llvm/lib/Target/AMDGPU/ |
H A D | SIProgramInfo.cpp | 151 static const MCExpr *MaskShift(const MCExpr *Val, uint32_t Mask, uint32_t Shift, in MaskShift()
|