Home
last modified time | relevance | path

Searched refs:AMDGPU (Results 1 – 25 of 243) sorted by relevance

12345678910

/openbsd-src/gnu/llvm/llvm/lib/Target/AMDGPU/
H A DSIRegisterInfo.cpp90 Register TmpVGPR = AMDGPU::NoRegister;
96 Register SavedExecReg = AMDGPU::NoRegister;
130 ExecReg = AMDGPU::EXEC_LO; in SGPRSpillBuilder()
131 MovOpc = AMDGPU::S_MOV_B32; in SGPRSpillBuilder()
132 NotOpc = AMDGPU::S_NOT_B32; in SGPRSpillBuilder()
134 ExecReg = AMDGPU::EXEC; in SGPRSpillBuilder()
135 MovOpc = AMDGPU::S_MOV_B64; in SGPRSpillBuilder()
136 NotOpc = AMDGPU::S_NOT_B64; in SGPRSpillBuilder()
139 assert(SuperReg != AMDGPU::M0 && "m0 should never spill"); in SGPRSpillBuilder()
140 assert(SuperReg != AMDGPU::EXEC_LO && SuperReg != AMDGPU::EXEC_HI && in SGPRSpillBuilder()
[all …]
H A DAMDGPUCombinerHelper.cpp22 case AMDGPU::G_FADD: in fnegFoldsIntoMI()
23 case AMDGPU::G_FSUB: in fnegFoldsIntoMI()
24 case AMDGPU::G_FMUL: in fnegFoldsIntoMI()
25 case AMDGPU::G_FMA: in fnegFoldsIntoMI()
26 case AMDGPU::G_FMAD: in fnegFoldsIntoMI()
27 case AMDGPU::G_FMINNUM: in fnegFoldsIntoMI()
28 case AMDGPU::G_FMAXNUM: in fnegFoldsIntoMI()
29 case AMDGPU::G_FMINNUM_IEEE: in fnegFoldsIntoMI()
30 case AMDGPU::G_FMAXNUM_IEEE: in fnegFoldsIntoMI()
31 case AMDGPU::G_FSIN: in fnegFoldsIntoMI()
[all …]
H A DSIInstrInfo.cpp42 namespace AMDGPU { namespace
65 : AMDGPUGenInstrInfo(AMDGPU::ADJCALLSTACKUP, AMDGPU::ADJCALLSTACKDOWN), in SIInstrInfo()
87 int Op0Idx = AMDGPU::getNamedOperandIdx(Opc0, OpName); in nodesHaveSameOperandValue()
88 int Op1Idx = AMDGPU::getNamedOperandIdx(Opc1, OpName); in nodesHaveSameOperandValue()
141 case AMDGPU::S_AND_SAVEEXEC_B32: in resultDependsOnExec()
142 case AMDGPU::S_AND_SAVEEXEC_B64: in resultDependsOnExec()
144 case AMDGPU::S_AND_B32: in resultDependsOnExec()
145 case AMDGPU::S_AND_B64: in resultDependsOnExec()
146 if (!Use.readsRegister(AMDGPU::EXEC)) in resultDependsOnExec()
159 case AMDGPU::V_READFIRSTLANE_B32: in resultDependsOnExec()
[all …]
H A DAMDGPURegisterBankInfo.cpp118 if (Opc == AMDGPU::G_ANYEXT || Opc == AMDGPU::G_ZEXT || in applyBank()
119 Opc == AMDGPU::G_SEXT) { in applyBank()
126 if (SrcBank == &AMDGPU::VCCRegBank) { in applyBank()
130 assert(NewBank == &AMDGPU::VGPRRegBank); in applyBank()
135 auto True = B.buildConstant(S32, Opc == AMDGPU::G_SEXT ? -1 : 1); in applyBank()
149 if (Opc == AMDGPU::G_TRUNC) { in applyBank()
152 assert(DstBank != &AMDGPU::VCCRegBank); in applyBank()
167 assert(NewBank == &AMDGPU::VGPRRegBank && in applyBank()
169 assert((MI.getOpcode() != AMDGPU::G_TRUNC && in applyBank()
170 MI.getOpcode() != AMDGPU::G_ANYEXT) && in applyBank()
[all …]
H A DAMDGPUResourceUsageAnalysis.cpp39 using namespace llvm::AMDGPU;
92 return AMDGPU::getTotalNumVGPRs(ST.hasGFX90AInsts(), ArgNumAGPR, ArgNumVGPR); in getTotalNumVGPRs()
114 if (AMDGPU::getAmdhsaCodeObjectVersion() >= 5) { in runOnModule()
175 Info.UsesFlatScratch = MRI.isPhysRegUsed(AMDGPU::FLAT_SCR_LO) || in analyzeResourceUsage()
176 MRI.isPhysRegUsed(AMDGPU::FLAT_SCR_HI) || in analyzeResourceUsage()
187 (!hasAnyNonFlatUseOfReg(MRI, *TII, AMDGPU::FLAT_SCR) && in analyzeResourceUsage()
188 !hasAnyNonFlatUseOfReg(MRI, *TII, AMDGPU::FLAT_SCR_LO) && in analyzeResourceUsage()
189 !hasAnyNonFlatUseOfReg(MRI, *TII, AMDGPU::FLAT_SCR_HI))) { in analyzeResourceUsage()
204 MRI.isPhysRegUsed(AMDGPU::VCC_LO) || MRI.isPhysRegUsed(AMDGPU::VCC_HI); in analyzeResourceUsage()
210 MCPhysReg HighestVGPRReg = AMDGPU::NoRegister; in analyzeResourceUsage()
[all …]
H A DSILoadStoreOptimizer.cpp321 return AMDGPU::getMUBUFElements(Opc); in getOpcodeWidth()
325 TII.getNamedOperand(MI, AMDGPU::OpName::dmask)->getImm(); in getOpcodeWidth()
329 return AMDGPU::getMTBUFElements(Opc); in getOpcodeWidth()
333 case AMDGPU::S_BUFFER_LOAD_DWORD_IMM: in getOpcodeWidth()
334 case AMDGPU::S_BUFFER_LOAD_DWORD_SGPR: in getOpcodeWidth()
335 case AMDGPU::S_BUFFER_LOAD_DWORD_SGPR_IMM: in getOpcodeWidth()
336 case AMDGPU::S_LOAD_DWORD_IMM: in getOpcodeWidth()
337 case AMDGPU::GLOBAL_LOAD_DWORD: in getOpcodeWidth()
338 case AMDGPU::GLOBAL_LOAD_DWORD_SADDR: in getOpcodeWidth()
339 case AMDGPU::GLOBAL_STORE_DWORD: in getOpcodeWidth()
[all …]
H A DSIOptimizeExecMasking.cpp95 case AMDGPU::COPY: in isCopyFromExec()
96 case AMDGPU::S_MOV_B64: in isCopyFromExec()
97 case AMDGPU::S_MOV_B64_term: in isCopyFromExec()
98 case AMDGPU::S_MOV_B32: in isCopyFromExec()
99 case AMDGPU::S_MOV_B32_term: { in isCopyFromExec()
106 return AMDGPU::NoRegister; in isCopyFromExec()
112 case AMDGPU::COPY: in isCopyToExec()
113 case AMDGPU::S_MOV_B64: in isCopyToExec()
114 case AMDGPU::S_MOV_B32: { in isCopyToExec()
120 case AMDGPU::S_MOV_B64_term: in isCopyToExec()
[all …]
H A DGCNDPPCombine.cpp128 if (AMDGPU::isTrue16Inst(Op)) in isShrinkable()
130 if (const auto *SDst = TII->getNamedOperand(MI, AMDGPU::OpName::sdst)) { in isShrinkable()
140 if (!hasNoImmOrEqual(MI, AMDGPU::OpName::src0_modifiers, 0, Mask) || in isShrinkable()
141 !hasNoImmOrEqual(MI, AMDGPU::OpName::src1_modifiers, 0, Mask) || in isShrinkable()
142 !hasNoImmOrEqual(MI, AMDGPU::OpName::clamp, 0) || in isShrinkable()
143 !hasNoImmOrEqual(MI, AMDGPU::OpName::omod, 0)) { in isShrinkable()
151 int DPP32 = AMDGPU::getDPPOp32(Op); in getDPPOp()
154 int E32 = AMDGPU::getVOPe32(Op); in getDPPOp()
155 DPP32 = (E32 == -1) ? -1 : AMDGPU::getDPPOp32(E32); in getDPPOp()
161 DPP64 = AMDGPU::getDPPOp64(Op); in getDPPOp()
[all …]
H A DSIPeepholeSDWA.cpp113 using namespace AMDGPU::SDWA;
309 if (TII->getNamedOperand(*MI, AMDGPU::OpName::src0) == SrcOp) { in getSrcMods()
310 if (auto *Mod = TII->getNamedOperand(*MI, AMDGPU::OpName::src0_modifiers)) { in getSrcMods()
313 } else if (TII->getNamedOperand(*MI, AMDGPU::OpName::src1) == SrcOp) { in getSrcMods()
314 if (auto *Mod = TII->getNamedOperand(*MI, AMDGPU::OpName::src1_modifiers)) { in getSrcMods()
344 MachineOperand *Src = TII->getNamedOperand(MI, AMDGPU::OpName::src0); in convertToSDWA()
345 MachineOperand *SrcSel = TII->getNamedOperand(MI, AMDGPU::OpName::src0_sel); in convertToSDWA()
347 TII->getNamedOperand(MI, AMDGPU::OpName::src0_modifiers); in convertToSDWA()
351 Src = TII->getNamedOperand(MI, AMDGPU::OpName::src1); in convertToSDWA()
352 SrcSel = TII->getNamedOperand(MI, AMDGPU::OpName::src1_sel); in convertToSDWA()
[all …]
H A DAMDGPUSubtarget.cpp118 ToggleFeature(AMDGPU::FeatureFlatForGlobal); in initializeSubtargetDependencies()
124 ToggleFeature(AMDGPU::FeatureFlatForGlobal); in initializeSubtargetDependencies()
146 if (AMDGPU::isGFX10Plus(*this) && in initializeSubtargetDependencies()
147 !getFeatureBits().test(AMDGPU::FeatureCuMode)) in initializeSubtargetDependencies()
181 MaxWavesPerEU = AMDGPU::IsaInfo::getMaxWavesPerEU(this); in GCNSubtarget()
182 EUsPerCU = AMDGPU::IsaInfo::getEUsPerCU(this); in GCNSubtarget()
196 case AMDGPU::V_LSHLREV_B64_e64: in getConstantBusLimit()
197 case AMDGPU::V_LSHLREV_B64_gfx10: in getConstantBusLimit()
198 case AMDGPU::V_LSHLREV_B64_e64_gfx11: in getConstantBusLimit()
199 case AMDGPU::V_LSHL_B64_e64: in getConstantBusLimit()
[all …]
H A DSIShrinkInstructions.cpp94 int Src0Idx = AMDGPU::getNamedOperandIdx(MI.getOpcode(), AMDGPU::OpName::src0); in foldImmediates()
154 if (AMDGPU::VGPR_32RegClass.contains(Reg) && in shouldShrinkTrue16()
155 !AMDGPU::VGPR_32_Lo128RegClass.contains(Reg)) in shouldShrinkTrue16()
231 int SOPKOpc = AMDGPU::getSOPKOp(MI.getOpcode()); in shrinkScalarCompare()
237 if (SOPKOpc == AMDGPU::S_CMPK_EQ_U32 || SOPKOpc == AMDGPU::S_CMPK_LG_U32) { in shrinkScalarCompare()
241 SOPKOpc = (SOPKOpc == AMDGPU::S_CMPK_EQ_U32) ? in shrinkScalarCompare()
242 AMDGPU::S_CMPK_EQ_I32 : AMDGPU::S_CMPK_LG_I32; in shrinkScalarCompare()
261 const AMDGPU::MIMGInfo *Info = AMDGPU::getMIMGInfo(MI.getOpcode()); in shrinkMIMG()
267 case AMDGPU::MIMGEncGfx10NSA: in shrinkMIMG()
268 NewEncoding = AMDGPU::MIMGEncGfx10Default; in shrinkMIMG()
[all …]
H A DSIFoldOperands.cpp144 case AMDGPU::V_MAC_F32_e64: in macToMad()
145 return AMDGPU::V_MAD_F32_e64; in macToMad()
146 case AMDGPU::V_MAC_F16_e64: in macToMad()
147 return AMDGPU::V_MAD_F16_e64; in macToMad()
148 case AMDGPU::V_FMAC_F32_e64: in macToMad()
149 return AMDGPU::V_FMA_F32_e64; in macToMad()
150 case AMDGPU::V_FMAC_F16_e64: in macToMad()
151 return AMDGPU::V_FMA_F16_gfx9_e64; in macToMad()
152 case AMDGPU::V_FMAC_F16_t16_e64: in macToMad()
153 return AMDGPU::V_FMA_F16_gfx9_e64; in macToMad()
[all …]
H A DAMDGPUInstructionSelector.cpp88 return MRI.getVRegDef(Reg)->getOpcode() != AMDGPU::G_TRUNC && in isVCC()
93 return RB->getID() == AMDGPU::VCCRegBankID; in isVCC()
100 MI.addOperand(*MF, MachineOperand::CreateReg(AMDGPU::EXEC, false, true)); in constrainCopyLikeIntrin()
131 if (SrcReg == AMDGPU::SCC) { in selectCOPY()
151 STI.isWave64() ? AMDGPU::S_MOV_B64 : AMDGPU::S_MOV_B32; in selectCOPY()
162 TRI.isSGPRClass(SrcRC) ? AMDGPU::S_AND_B32 : AMDGPU::V_AND_B32_e32; in selectCOPY()
166 BuildMI(*BB, &I, DL, TII.get(AMDGPU::V_CMP_NE_U32_e64), DstReg) in selectCOPY()
248 BuildMI(*BB, MI, MI->getDebugLoc(), TII.get(AMDGPU::COPY), DstReg) in getSubOperand64()
264 case AMDGPU::sub0: in getSubOperand64()
266 case AMDGPU::sub1: in getSubOperand64()
[all …]
H A DGCNHazardRecognizer.cpp62 MaxLookAhead = MF.getRegInfo().isPhysRegUsed(AMDGPU::AGPR0) ? 19 : 5; in GCNHazardRecognizer()
80 return Opcode == AMDGPU::V_DIV_FMAS_F32_e64 || Opcode == AMDGPU::V_DIV_FMAS_F64_e64; in isDivFMas()
84 return Opcode == AMDGPU::S_GETREG_B32; in isSGetReg()
89 case AMDGPU::S_SETREG_B32: in isSSetReg()
90 case AMDGPU::S_SETREG_B32_mode: in isSSetReg()
91 case AMDGPU::S_SETREG_IMM32_B32: in isSSetReg()
92 case AMDGPU::S_SETREG_IMM32_B32_mode: in isSSetReg()
99 return Opcode == AMDGPU::V_READLANE_B32 || Opcode == AMDGPU::V_WRITELANE_B32; in isRWLane()
103 return Opcode == AMDGPU::S_RFE_B64; in isRFE()
108 case AMDGPU::S_MOVRELS_B32: in isSMovRel()
[all …]
H A DAMDGPUArgumentUsageInfo.cpp95 &AMDGPU::SGPR_128RegClass, LLT::fixed_vector(4, 32)); in getPreloadedValue()
99 &AMDGPU::SGPR_64RegClass, in getPreloadedValue()
103 &AMDGPU::SGPR_32RegClass, LLT::scalar(32)); in getPreloadedValue()
106 &AMDGPU::SGPR_32RegClass, LLT::scalar(32)); in getPreloadedValue()
109 &AMDGPU::SGPR_32RegClass, LLT::scalar(32)); in getPreloadedValue()
112 &AMDGPU::SGPR_32RegClass, LLT::scalar(32)); in getPreloadedValue()
116 &AMDGPU::SGPR_32RegClass, LLT::scalar(32)); in getPreloadedValue()
119 &AMDGPU::SGPR_64RegClass, in getPreloadedValue()
123 &AMDGPU::SGPR_64RegClass, in getPreloadedValue()
127 &AMDGPU::SGPR_64RegClass, LLT::scalar(64)); in getPreloadedValue()
[all …]
H A DSIInsertWaitcnts.cpp153 const AMDGPU::MIMGInfo *Info = AMDGPU::getMIMGInfo(Inst.getOpcode()); in getVmemType()
154 const AMDGPU::MIMGBaseOpcodeInfo *BaseInfo = in getVmemType()
155 AMDGPU::getMIMGBaseOpcodeInfo(Info->BaseOpcode); in getVmemType()
160 void addWait(AMDGPU::Waitcnt &Wait, InstCounterType T, unsigned Count) { in addWait()
247 void simplifyWaitcnt(AMDGPU::Waitcnt &Wait) const;
249 void determineWait(InstCounterType T, int RegNo, AMDGPU::Waitcnt &Wait) const;
250 void applyWaitcnt(const AMDGPU::Waitcnt &Wait);
366 AMDGPU::IsaVersion IV;
421 AMDGPU::Waitcnt allZeroWaitcnt() const { in allZeroWaitcnt()
422 return AMDGPU::Waitcnt::allZero(ST->hasVscnt()); in allZeroWaitcnt()
[all …]
H A DSIPreAllocateWWMRegs.cpp166 if (Opc == AMDGPU::ENTER_STRICT_WWM || Opc == AMDGPU::ENTER_STRICT_WQM || in printWWMInfo()
167 Opc == AMDGPU::ENTER_PSEUDO_WM) { in printWWMInfo()
170 assert(Opc == AMDGPU::EXIT_STRICT_WWM || Opc == AMDGPU::EXIT_STRICT_WQM || in printWWMInfo()
171 Opc == AMDGPU::EXIT_PSEUDO_WM); in printWWMInfo()
175 if (Opc == AMDGPU::ENTER_STRICT_WWM || Opc == AMDGPU::EXIT_STRICT_WWM) { in printWWMInfo()
177 } else if (Opc == AMDGPU::ENTER_PSEUDO_WM || Opc == AMDGPU::EXIT_PSEUDO_WM) { in printWWMInfo()
180 assert(Opc == AMDGPU::ENTER_STRICT_WQM || Opc == AMDGPU::EXIT_STRICT_WQM); in printWWMInfo()
216 if (MI.getOpcode() == AMDGPU::V_SET_INACTIVE_B32 || in runOnMachineFunction()
217 MI.getOpcode() == AMDGPU::V_SET_INACTIVE_B64) in runOnMachineFunction()
220 if (MI.getOpcode() == AMDGPU::ENTER_STRICT_WWM || in runOnMachineFunction()
[all …]
H A DAMDGPUGenRegisterBankInfo.def14 namespace AMDGPU {
195 if (BankID == AMDGPU::VCCRegBankID)
198 Idx = BankID == AMDGPU::SGPRRegBankID ? PM_SGPR1 : PM_VGPR1;
202 case AMDGPU::VGPRRegBankID:
205 case AMDGPU::SGPRRegBankID:
208 case AMDGPU::AGPRRegBankID:
216 case AMDGPU::VGPRRegBankID:
219 case AMDGPU::SGPRRegBankID:
222 case AMDGPU::AGPRRegBankID:
230 case AMDGPU::VGPRRegBankID:
[all …]
/openbsd-src/gnu/llvm/llvm/lib/Target/AMDGPU/MCA/
H A DAMDGPUCustomBehaviour.cpp27 case AMDGPU::S_WAITCNT: in postProcessInstruction()
28 case AMDGPU::S_WAITCNT_EXPCNT: in postProcessInstruction()
29 case AMDGPU::S_WAITCNT_LGKMCNT: in postProcessInstruction()
30 case AMDGPU::S_WAITCNT_VMCNT: in postProcessInstruction()
31 case AMDGPU::S_WAITCNT_VSCNT: in postProcessInstruction()
32 case AMDGPU::S_WAITCNT_EXPCNT_gfx10: in postProcessInstruction()
33 case AMDGPU::S_WAITCNT_LGKMCNT_gfx10: in postProcessInstruction()
34 case AMDGPU::S_WAITCNT_VMCNT_gfx10: in postProcessInstruction()
35 case AMDGPU::S_WAITCNT_VSCNT_gfx10: in postProcessInstruction()
36 case AMDGPU::S_WAITCNT_gfx10: in postProcessInstruction()
[all …]
/openbsd-src/gnu/usr.bin/clang/include/llvm/AMDGPU/
H A DMakefile48 AMDGPUGenAsmMatcher.inc: ${LLVM_SRCS}/lib/Target/AMDGPU/AMDGPU.td
50 -I${LLVM_SRCS}/include -I${LLVM_SRCS}/lib/Target/AMDGPU \
52 AMDGPUGenAsmWriter.inc: ${LLVM_SRCS}/lib/Target/AMDGPU/AMDGPU.td
54 -I${LLVM_SRCS}/include -I${LLVM_SRCS}/lib/Target/AMDGPU \
56 AMDGPUGenCallingConv.inc: ${LLVM_SRCS}/lib/Target/AMDGPU/AMDGPU.td
58 -I${LLVM_SRCS}/include -I${LLVM_SRCS}/lib/Target/AMDGPU \
60 AMDGPUGenDAGISel.inc: ${LLVM_SRCS}/lib/Target/AMDGPU/AMDGPU.td
62 -I${LLVM_SRCS}/include -I${LLVM_SRCS}/lib/Target/AMDGPU \
64 AMDGPUGenDisassemblerTables.inc: ${LLVM_SRCS}/lib/Target/AMDGPU/AMDGPU.td
66 -I${LLVM_SRCS}/include -I${LLVM_SRCS}/lib/Target/AMDGPU \
[all …]
/openbsd-src/gnu/llvm/llvm/lib/Target/AMDGPU/MCTargetDesc/
H A DSIMCCodeEmitter.cpp143 STI.getFeatureBits()[AMDGPU::FeatureInv2PiInlineImm]) in getLit16Encoding()
179 STI.getFeatureBits()[AMDGPU::FeatureInv2PiInlineImm]) in getLit32Encoding()
215 STI.getFeatureBits()[AMDGPU::FeatureInv2PiInlineImm]) in getLit64Encoding()
243 case AMDGPU::OPERAND_REG_IMM_INT32: in getLitEncoding()
244 case AMDGPU::OPERAND_REG_IMM_FP32: in getLitEncoding()
245 case AMDGPU::OPERAND_REG_IMM_FP32_DEFERRED: in getLitEncoding()
246 case AMDGPU::OPERAND_REG_INLINE_C_INT32: in getLitEncoding()
247 case AMDGPU::OPERAND_REG_INLINE_C_FP32: in getLitEncoding()
248 case AMDGPU::OPERAND_REG_INLINE_AC_INT32: in getLitEncoding()
249 case AMDGPU::OPERAND_REG_INLINE_AC_FP32: in getLitEncoding()
[all …]
H A DAMDGPUInstPrinter.cpp25 using namespace llvm::AMDGPU;
147 if (AMDGPU::isGFX10(STI)) { in printFlatOffset()
208 O << ((AMDGPU::isGFX940(STI) && in printCPol()
212 O << (AMDGPU::isGFX940(STI) ? " nt" : " slc"); in printCPol()
213 if ((Imm & CPol::DLC) && AMDGPU::isGFX10Plus(STI)) in printCPol()
215 if ((Imm & CPol::SCC) && AMDGPU::isGFX90A(STI)) in printCPol()
216 O << (AMDGPU::isGFX940(STI) ? " sc1" : " scc"); in printCPol()
243 const AMDGPU::MIMGDimInfo *DimInfo = AMDGPU::getMIMGDimInfoByEncoding(Dim); in printDim()
262 if (STI.hasFeature(AMDGPU::FeatureR128A16)) in printR128A16()
303 using namespace llvm::AMDGPU::MTBUFFormat; in printSymbolicFormat()
[all …]
/openbsd-src/gnu/llvm/llvm/lib/Target/AMDGPU/Disassembler/
H A DAMDGPUDisassembler.cpp42 (isGFX10Plus() ? AMDGPU::EncValues::SGPR_MAX_GFX10 \
43 : AMDGPU::EncValues::SGPR_MAX_SI)
54 if (!STI.getFeatureBits()[AMDGPU::FeatureGCN3Encoding] && !isGFX10Plus()) in AMDGPUDisassembler()
68 int OpIdx = AMDGPU::getNamedOperandIdx(MI.getOpcode(), NameIdx); in insertNamedMCOperand()
310 unsigned Sub = MRI->getSubReg(Op.getReg(), AMDGPU::sub0); in IsAGPROperand()
312 return Reg >= AMDGPU::AGPR0 && Reg <= AMDGPU::AGPR255; in IsAGPROperand()
329 uint16_t DataNameIdx = (TSFlags & SIInstrFlags::DS) ? AMDGPU::OpName::data0 in decodeOperand_AVLdSt_Any()
330 : AMDGPU::OpName::vdata; in decodeOperand_AVLdSt_Any()
332 int DataIdx = AMDGPU::getNamedOperandIdx(Opc, DataNameIdx); in decodeOperand_AVLdSt_Any()
334 int DstIdx = AMDGPU::getNamedOperandIdx(Opc, AMDGPU::OpName::vdst); in decodeOperand_AVLdSt_Any()
[all …]
/openbsd-src/gnu/llvm/llvm/lib/Target/AMDGPU/AsmParser/
H A DAMDGPUAsmParser.cpp44 using namespace llvm::AMDGPU;
273 return isRegOrImmWithInputMods(AMDGPU::VS_32RegClassID, MVT::i16); in isRegOrImmWithInt16InputMods()
277 return isRegOrImmWithInputMods(AMDGPU::VS_32RegClassID, MVT::i32); in isRegOrImmWithInt32InputMods()
281 return isRegOrInline(AMDGPU::VS_32RegClassID, MVT::i16); in isRegOrInlineImmWithInt16InputMods()
285 return isRegOrInline(AMDGPU::VS_32RegClassID, MVT::i32); in isRegOrInlineImmWithInt32InputMods()
289 return isRegOrImmWithInputMods(AMDGPU::VS_64RegClassID, MVT::i64); in isRegOrImmWithInt64InputMods()
293 return isRegOrImmWithInputMods(AMDGPU::VS_32RegClassID, MVT::f16); in isRegOrImmWithFP16InputMods()
297 return isRegOrImmWithInputMods(AMDGPU::VS_32RegClassID, MVT::f32); in isRegOrImmWithFP32InputMods()
301 return isRegOrImmWithInputMods(AMDGPU::VS_64RegClassID, MVT::f64); in isRegOrImmWithFP64InputMods()
305 return isRegOrInline(AMDGPU::VS_32RegClassID, MVT::f16); in isRegOrInlineImmWithFP16InputMods()
[all …]
/openbsd-src/gnu/llvm/llvm/lib/Target/AMDGPU/Utils/
H A DAMDGPUBaseInfo.cpp99 namespace AMDGPU { namespace
160 return AMDGPU::ImplicitArg::MULTIGRID_SYNC_ARG_OFFSET; in getMultigridSyncArgImplicitArgPosition()
177 return AMDGPU::ImplicitArg::HOSTCALL_PTR_OFFSET; in getHostcallImplicitArgPosition()
192 return AMDGPU::ImplicitArg::DEFAULT_QUEUE_OFFSET; in getDefaultQueueImplicitArgPosition()
204 return AMDGPU::ImplicitArg::COMPLETION_ACTION_OFFSET; in getCompletionActionImplicitArgPosition()
458 return AMDGPU::hasNamedOperand(Opc, AMDGPU::OpName::src0X); in isVOPD()
462 return Opc == AMDGPU::V_MAC_F32_e64_gfx6_gfx7 || in isMAC()
463 Opc == AMDGPU::V_MAC_F32_e64_gfx10 || in isMAC()
464 Opc == AMDGPU::V_MAC_F32_e64_vi || in isMAC()
465 Opc == AMDGPU::V_MAC_LEGACY_F32_e64_gfx6_gfx7 || in isMAC()
[all …]

12345678910