| /openbsd-src/gnu/llvm/llvm/lib/CodeGen/ |
| H A D | ComplexDeinterleavingPass.cpp | 435 if (R0->getOpcode() == Instruction::FNeg || in identifyNodeWithImplicitAdd() 436 R1->getOpcode() == Instruction::FNeg) { in identifyNodeWithImplicitAdd() 438 if (R0->getOpcode() == Instruction::FNeg) { in identifyNodeWithImplicitAdd() 448 if (I0->getOpcode() == Instruction::FNeg || in identifyNodeWithImplicitAdd() 449 I1->getOpcode() == Instruction::FNeg) { in identifyNodeWithImplicitAdd() 452 if (I0->getOpcode() == Instruction::FNeg) { in identifyNodeWithImplicitAdd()
|
| /openbsd-src/gnu/llvm/llvm/lib/Transforms/Scalar/ |
| H A D | Float2Int.cpp | 212 case Instruction::FNeg: in walkBackwards() 292 case Instruction::FNeg: { in calcRange() 483 case Instruction::FNeg: in convert()
|
| H A D | SpeculativeExecution.cpp | 245 case Instruction::FNeg: in ComputeSpeculationCost()
|
| H A D | GVNSink.cpp | 454 case Instruction::FNeg: in lookupOrAdd()
|
| H A D | Reassociate.cpp | 839 ? ConstantFoldUnaryOpOperand(Instruction::FNeg, C, DL) in NegateValue() 2269 } else if (I->getOpcode() == Instruction::FNeg || in OptimizeInst()
|
| H A D | LowerMatrixIntrinsics.cpp | 559 case Instruction::FNeg: in isUniformShape() 1915 case Instruction::FNeg: in VisitUnaryOperator()
|
| /openbsd-src/gnu/llvm/llvm/lib/Target/VE/ |
| H A D | VVPInstrPatternsVec.td | 142 multiclass FNeg<ValueType DataVT> { 159 defm: FNeg<v256f32>; 160 defm: FNeg<v256f64>; 162 ///// Packed FNeg /////
|
| /openbsd-src/gnu/llvm/llvm/lib/Target/DirectX/ |
| H A D | DXILPrepare.cpp | 121 if (I.getOpcode() == Instruction::FNeg) { in runOnModule()
|
| /openbsd-src/gnu/llvm/llvm/lib/Transforms/InstCombine/ |
| H A D | InstCombineAddSub.cpp | 2410 if (Constant *NegC = ConstantFoldUnaryOpOperand(Instruction::FNeg, C, DL)) in foldFNegIntoConstant() 2414 if (Constant *NegC = ConstantFoldUnaryOpOperand(Instruction::FNeg, C, DL)) in foldFNegIntoConstant() 2418 if (Constant *NegC = ConstantFoldUnaryOpOperand(Instruction::FNeg, C, DL)) { in foldFNegIntoConstant() 2433 if (Constant *NegC = ConstantFoldUnaryOpOperand(Instruction::FNeg, C, DL)) in foldFNegIntoConstant() 2441 Value *FNeg; in hoistFNegAboveFMulFDiv() local 2442 if (!match(&I, m_FNeg(m_Value(FNeg)))) in hoistFNegAboveFMulFDiv() 2446 if (match(FNeg, m_OneUse(m_FMul(m_Value(X), m_Value(Y))))) in hoistFNegAboveFMulFDiv() 2449 if (match(FNeg, m_OneUse(m_FDiv(m_Value(X), m_Value(Y))))) in hoistFNegAboveFMulFDiv() 2593 if (Constant *NegC = ConstantFoldUnaryOpOperand(Instruction::FNeg, C, DL)) in visitFSub()
|
| H A D | InstCombineCalls.cpp | 1801 Instruction *FNeg = UnaryOperator::CreateFNeg(NewCall); in visitCallInst() local 1802 FNeg->copyIRFlags(II); in visitCallInst() 1803 return FNeg; in visitCallInst() 2076 Instruction *FNeg = UnaryOperator::CreateFNeg(NewSin); in visitCallInst() local 2077 FNeg->copyFastMathFlags(II); in visitCallInst() 2078 return FNeg; in visitCallInst()
|
| H A D | InstCombineMulDivRem.cpp | 561 if (Constant *NegC = ConstantFoldUnaryOpOperand(Instruction::FNeg, C, DL)) in visitFMul() 1464 if (Constant *NegC = ConstantFoldUnaryOpOperand(Instruction::FNeg, C, DL)) in foldFDivConstantDivisor() 1507 if (Constant *NegC = ConstantFoldUnaryOpOperand(Instruction::FNeg, C, DL)) in foldFDivConstantDividend()
|
| /openbsd-src/gnu/llvm/llvm/lib/Transforms/Vectorize/ |
| H A D | VectorCombine.cpp | 623 Instruction *FNeg; in foldInsExtFNeg() local 624 if (!match(&I, m_InsertElt(m_Value(DestVec), m_OneUse(m_Instruction(FNeg)), in foldInsExtFNeg() 631 if (!match(FNeg, m_FNeg(m_CombineAnd( in foldInsExtFNeg() 656 TTI.getArithmeticInstrCost(Instruction::FNeg, ScalarTy) + in foldInsExtFNeg() 666 TTI.getArithmeticInstrCost(Instruction::FNeg, VecTy) + in foldInsExtFNeg() 674 Value *VecFNeg = Builder.CreateFNegFMF(SrcVec, FNeg); in foldInsExtFNeg()
|
| H A D | VPlanRecipes.cpp | 433 Opcode == Instruction::FNeg || Opcode == Instruction::FSub || in setFastMathFlags() 574 case Instruction::FNeg: in execute()
|
| /openbsd-src/gnu/llvm/llvm/include/llvm/IR/ |
| H A D | Operator.h | 307 case Instruction::FNeg: in classof()
|
| H A D | Instruction.def | 142 HANDLE_UNARY_INST(12, FNeg , UnaryOperator)
|
| H A D | VPIntrinsics.def | 294 VP_PROPERTY_FUNCTIONAL_OPC(FNeg)
|
| H A D | IRBuilder.h | 1647 if (Value *Res = Folder.FoldUnOpFMF(Instruction::FNeg, V, FMF)) 1658 if (Value *Res = Folder.FoldUnOpFMF(Instruction::FNeg, V, FMF))
|
| /openbsd-src/gnu/llvm/llvm/include/llvm/Analysis/ |
| H A D | TargetTransformInfoImpl.h | 1091 case Instruction::FNeg: { in getInstructionCost() 1094 if (Opcode != Instruction::FNeg) in getInstructionCost()
|
| /openbsd-src/gnu/llvm/llvm/lib/Target/AMDGPU/ |
| H A D | AMDGPUCodeGenPrepare.cpp | 696 Value *FNeg = Builder.CreateFNeg(Den); in optimizeWithRcp() local 697 return Builder.CreateCall(Decl, { FNeg }); in optimizeWithRcp()
|
| H A D | R600ISelLowering.cpp | 1746 SDValue FNeg = N->getOperand(0); in PerformDAGCombine() local 1747 if (FNeg.getOpcode() != ISD::FNEG) { in PerformDAGCombine() 1750 SDValue SelectCC = FNeg.getOperand(0); in PerformDAGCombine()
|
| /openbsd-src/gnu/llvm/llvm/lib/IR/ |
| H A D | Instruction.cpp | 414 case FNeg: return "fneg"; in getOpcodeName()
|
| H A D | ConstantFold.cpp | 827 case Instruction::FNeg: in ConstantFoldUnaryInstruction() 844 case Instruction::FNeg: in ConstantFoldUnaryInstruction()
|
| /openbsd-src/gnu/llvm/llvm/lib/AsmParser/ |
| H A D | LLLexer.cpp | 831 INSTKEYWORD(fneg, FNeg); in LexIdentifier()
|
| /openbsd-src/gnu/llvm/llvm/lib/ExecutionEngine/Interpreter/ |
| H A D | Execution.cpp | 76 case Instruction::FNeg: in visitUnaryOperator() 93 case Instruction::FNeg: executeFNegInst(R, Src, Ty); break; in visitUnaryOperator()
|
| /openbsd-src/gnu/llvm/llvm/lib/Analysis/ |
| H A D | InstructionSimplify.cpp | 5283 if (Constant *C = foldConstant(Instruction::FNeg, Op, Q)) in simplifyFNegInst() 5490 return foldConstant(Instruction::FNeg, Op1, Q); in simplifyFSubInst() 5694 case Instruction::FNeg: in simplifyUnOp() 5708 case Instruction::FNeg: in simplifyFPUnOp() 6599 case Instruction::FNeg: in simplifyInstructionWithOperands()
|