/llvm-project/libcxx/test/libcxx/algorithms/ |
H A D | ranges_robust_against_copying_projections.pass.cpp | 27 constexpr explicit Proj(int *copies) : copies_(copies) {} in Proj() 75 int copies = 0; in all_the_algorithms() 76 (void)std::ranges::adjacent_find(first, last, Equal(), Proj(&copies)); assert(copies == 0); in all_the_algorithms() 77 (void)std::ranges::adjacent_find(a, Equal(), Proj(&copies)); assert(copies == 0); in all_the_algorithms() 78 (void)std::ranges::all_of(first, last, UnaryTrue(), Proj(&copies)); assert(copies == 0); in all_the_algorithms() 79 (void)std::ranges::all_of(a, UnaryTrue(), Proj(&copies)); asser in all_the_algorithms() 74 int copies = 0; all_the_algorithms() local 240 int copies = 0; test_deque() local [all...] |
H A D | ranges_robust_against_copying_comparators.pass.cpp | 25 constexpr explicit Less(int *copies) : copies_(copies) {} in Less() 33 constexpr explicit Equal(int *copies) : copies_(copies) {} in Equal() 41 constexpr explicit UnaryVoid(int *copies) : copies_(copies) {} in UnaryVoid() 49 constexpr explicit UnaryTrue(int *copies) : copies_(copies) {} in UnaryTrue() 57 constexpr explicit NullaryValue(int *copies) : copies_(copies) {} in NullaryValue() 92 int copies = 0; all_the_algorithms() local [all...] |
H A D | robust_against_copying_comparators.pass.cpp | 22 TEST_CONSTEXPR explicit Less(int *copies) : copies_(copies) {} in Less() 31 TEST_CONSTEXPR explicit Equal(int *copies) : copies_(copies) {} in Equal() 40 TEST_CONSTEXPR explicit UnaryVoid(int *copies) : copies_(copies) {} in UnaryVoid() 49 TEST_CONSTEXPR explicit UnaryTrue(int *copies) : copies_(copies) {} in UnaryTrue() 58 TEST_CONSTEXPR explicit NullaryValue(int *copies) : copies_(copies) {} in NullaryValue() 106 int copies = 0; all_the_algorithms() local [all...] |
/llvm-project/llvm/unittests/CodeGen/GlobalISel/ |
H A D | KnownBitsTest.cpp | 19 unsigned CopyReg = Copies[Copies.size() - 1]; in TEST_F() 39 unsigned CopyReg = Copies[Copies.size() - 1]; in TEST_F() 78 Register CopyReg = Copies[Copies.size() - 1]; in TEST_F() 112 Register CopyReg = Copies[Copies.size() - 1]; in TEST_F() 150 Register CopyReg = Copies[Copies.size() - 1]; in TEST_F() 187 Register CopyReg = Copies[Copies.size() - 1]; in TEST_F() 226 Register CopyReg = Copies[Copies.size() - 1]; in TEST_F() 254 unsigned CopyReg = Copies[Copies.size() - 1]; in TEST_F() 283 Register CopyReg = Copies[Copies.size() - 1]; in TEST_F() 315 Register CopyReg = Copies[Copies.size() - 1]; in TEST_F() [all …]
|
H A D | MachineIRBuilderTest.cpp | 80 SmallVector<Register, 4> Copies; in TEST_F() local 81 collectCopies(Copies, MF); in TEST_F() 84 auto MIBAdd = B.buildAdd(s64, Copies[0], Copies[1]); in TEST_F() 106 SmallVector<Register, 4> Copies; in TEST_F() local 107 collectCopies(Copies, MF); in TEST_F() 108 B.buildUnmerge(LLT::scalar(32), Copies[0]); in TEST_F() 109 B.buildUnmerge(LLT::scalar(16), Copies[1]); in TEST_F() 127 SmallVector<Register, 4> Copies; in TEST_F() local 128 collectCopies(Copies, M in TEST_F() 164 SmallVector<Register, 4> Copies; TEST_F() local 192 SmallVector<Register, 4> Copies; TEST_F() local 220 SmallVector<Register, 4> Copies; TEST_F() local 248 SmallVector<Register, 4> Copies; TEST_F() local 273 SmallVector<Register, 4> Copies; TEST_F() local 302 SmallVector<Register, 4> Copies; TEST_F() local 402 SmallVector<Register, 4> Copies; TEST_F() local 436 SmallVector<Register, 4> Copies; TEST_F() local 459 SmallVector<Register, 4> Copies; TEST_F() local [all...] |
H A D | PatternMatchTest.cpp | 70 B.buildBuildVector(v4s64, {Copies[0], Copies[0], Copies[0], Copies[0]}); in TEST_F() 83 auto MIBAdd = B.buildAdd(LLT::scalar(64), Copies[0], Copies[1]); in TEST_F() 97 auto MIBBrcond = B.buildBrCond(Copies[0], B.getMBB()); in TEST_F() 104 B.buildUAddo(LLT::scalar(64), LLT::scalar(1), Copies[0], Copies[1]); in TEST_F() 117 auto MIBAdd = B.buildAdd(s64, Copies[ in TEST_F() [all...] |
H A D | KnownBitsVectorTest.cpp | 33 Register CopyReg = Copies[Copies.size() - 1]; in TEST_F() 74 Register CopyReg = Copies[Copies.size() - 1]; in TEST_F() 110 Register CopyReg = Copies[Copies.size() - 1]; in TEST_F() 146 Register CopyReg = Copies[Copies.size() - 1]; in TEST_F() 184 Register CopyReg = Copies[Copies in TEST_F() [all...] |
H A D | LegalizerHelperTest.cpp | 38 auto Src = B.buildTrunc(S32, Copies[0]); in TEST_F() 39 auto Amt = B.buildTrunc(S32, Copies[1]); in TEST_F() 91 auto Src = B.buildTrunc(S24, Copies[0]); in TEST_F() 92 auto Amt = B.buildTrunc(S24, Copies[1]); in TEST_F() 149 auto SrcTrunc = B.buildTrunc(S32, Copies[0]); in TEST_F() 151 auto AmtTrunc = B.buildTrunc(S32, Copies[1]); in TEST_F() 196 B.buildInstr(TargetOpcode::G_CTTZ, {LLT::scalar(32)}, {Copies[0]}); in TEST_F() 228 B.buildInstr(TargetOpcode::G_CTTZ, {LLT::scalar(64)}, {Copies[0]}); in TEST_F() 262 B.buildInstr(TargetOpcode::G_CTLZ, {LLT::scalar(32)}, {Copies[0]}); in TEST_F() 297 B.buildInstr(TargetOpcode::G_CTTZ, {LLT::scalar(32)}, {Copies[0]}); in TEST_F() [all …]
|
/llvm-project/libcxx/test/std/algorithms/alg.modifying.operations/alg.move/ |
H A D | contiguous_trivial_optimization.pass.cpp | 27 constexpr TMBNTC(int& copies) : p(&copies) {} in TMBNTC() 36 int copies = 0; in test_trivial_moveassign_but_no_trivial_copyassign() local 37 TMBNTC ia[] = { copies, copies, copies, copies }; in test_trivial_moveassign_but_no_trivial_copyassign() 38 TMBNTC ib[] = { copies, copies, copies, copies }; in test_trivial_moveassign_but_no_trivial_copyassign() 40 assert(copies == 4); in test_trivial_moveassign_but_no_trivial_copyassign() 41 copies = 0; in test_trivial_moveassign_but_no_trivial_copyassign() 43 assert(copies == 4); in test_trivial_moveassign_but_no_trivial_copyassign() 45 copies = 0; in test_trivial_moveassign_but_no_trivial_copyassign() 47 assert(copies == 0); in test_trivial_moveassign_but_no_trivial_copyassign() 49 assert(copies == 0); in test_trivial_moveassign_but_no_trivial_copyassign() [all …]
|
/llvm-project/libcxx/test/std/utilities/memory/util.smartptr/util.smartptr.shared/util.smartptr.shared.create/ |
H A D | make_shared.array.bounded.pass.cpp | 169 // Count the number of copies being made in main() 176 assert(CountCopies::copies() == 0); in main() 182 assert(CountCopies::copies() == 0); in main() 188 assert(CountCopies::copies() == 0); in main() 194 int copies = 0; in main() local 195 CountCopies init(&copies); in main() 197 assert(copies == 8); in main() 201 int copies = 0; in main() local 202 CountCopies init[3] = {&copies, &copies, &copies}; in main() 204 assert(copies == 8 * 3); in main() [all …]
|
H A D | make_shared.array.unbounded.pass.cpp | 210 // Count the number of copies being made in main() 217 assert(CountCopies::copies() == 0); in main() 223 assert(CountCopies::copies() == 0); in main() 229 assert(CountCopies::copies() == 0); in main() 235 int copies = 0; in main() local 236 CountCopies init(&copies); in main() 238 assert(copies == 8); in main() 242 int copies = 0; in main() local 243 CountCopies init[3] = {&copies, &copies, &copies}; in main() 245 assert(copies == 8 * 3); in main() [all …]
|
H A D | allocate_shared.array.bounded.pass.cpp | 170 // Count the number of copies being made in main() 177 assert(CountCopies::copies() == 0); in main() 183 assert(CountCopies::copies() == 0); in main() 189 assert(CountCopies::copies() == 0); in main() 195 int copies = 0; in main() local 196 CountCopies init(&copies); in main() 198 assert(copies == 8); in main() 202 int copies = 0; in main() local 203 CountCopies init[3] = {&copies, &copies, &copies}; in main() 205 assert(copies == 8 * 3); in main() [all …]
|
H A D | allocate_shared.array.unbounded.pass.cpp | 212 // Count the number of copies being made in main() 219 assert(CountCopies::copies() == 0); in main() 225 assert(CountCopies::copies() == 0); in main() 231 assert(CountCopies::copies() == 0); in main() 237 int copies = 0; in main() local 238 CountCopies init(&copies); in main() 240 assert(copies == 8); in main() 244 int copies = 0; in main() local 245 CountCopies init[3] = {&copies, &copies, &copies}; in main() 247 assert(copies == 8 * 3); in main() [all …]
|
/llvm-project/libc/AOR_v20.02/string/aarch64/ |
H A D | memcpy-advsimd.S | 45 Copies are split into 3 main cases: small copies of up to 32 bytes, medium 46 copies of up to 128 bytes, and large copies. The overhead of the overlap 47 check is negligible since it is only required for large copies. 49 Large copies use a software pipelined loop processing 64 bytes per iteration. 63 /* Small copies: 0..32 bytes. */ 105 /* Medium copies: 33..128 bytes. */ 167 /* Large backwards copy for overlapping copies.
|
H A D | memcpy.S | 48 Copies are split into 3 main cases: small copies of up to 32 bytes, medium 49 copies of up to 128 bytes, and large copies. The overhead of the overlap 50 check is negligible since it is only required for large copies. 52 Large copies use a software pipelined loop processing 64 bytes per iteration. 66 /* Small copies: 0..32 bytes. */ 108 /* Medium copies: 33..128 bytes. */ 196 /* Large backwards copy for overlapping copies.
|
/llvm-project/llvm/test/CodeGen/AArch64/ |
H A D | regcoal-physreg.mir | 15 ; We usually should not coalesce copies from allocatable physregs. 21 ; It is fine to coalesce copies from reserved physregs 27 ; It is not fine to coalesce copies from reserved physregs when they are 35 ; Is is fine to coalesce copies from constant physregs even when they are 43 ; Is is fine to coalesce copies from constant physregs even when they are 64 ; It is not fine to coalesce copies from reserved physregs when they are 119 ; We can coalesce copies from physreg to vreg across multiple blocks.
|
/llvm-project/libcxx/test/support/ |
H A D | invocable_with_telemetry.h | 23 int copies; member 41 ++telemetry_->copies; in invocable_with_telemetry() 47 // Not using move-and-swap idiom to ensure that copies and moves remain accurate. 61 // Not using copy-and-swap idiom to ensure that copies and moves remain accurate. 68 ++telemetry_->copies; 86 invocable_with_telemetry(F f, int& invocations, int& moves, int& copies) -> invocable_with_telemetr…
|
/llvm-project/llvm/lib/Target/AMDGPU/ |
H A D | SIFixVGPRCopies.cpp | 1 //===-- SIFixVGPRCopies.cpp - Fix VGPR Copies after regalloc --------------===// 10 /// Add implicit use of exec to vector register copies. 22 #define DEBUG_TYPE "si-fix-vgpr-copies" 41 StringRef getPassName() const override { return "SI Fix VGPR copies"; } in getPassName() 51 INITIALIZE_PASS(SIFixVGPRCopiesLegacy, DEBUG_TYPE, "SI Fix VGPR copies", false,
|
H A D | SILowerWWMCopies.cpp | 1 //===-- SILowerWWMCopies.cpp - Lower Copies after regalloc ---===// 30 #define DEBUG_TYPE "si-lower-wwm-copies" 62 StringRef getPassName() const override { return "SI Lower WWM Copies"; } 75 INITIALIZE_PASS_BEGIN(SILowerWWMCopiesLegacy, DEBUG_TYPE, "SI Lower WWM Copies", in isSCCLiveAtMI() 79 INITIALIZE_PASS_END(SILowerWWMCopiesLegacy, DEBUG_TYPE, "SI Lower WWM Copies", in isSCCLiveAtMI() 157 // For WWM vector copies, manipulate the exec mask around the copy
|
/llvm-project/llvm/test/CodeGen/SystemZ/ |
H A D | fp-copysign-01.ll | 10 ; Test f32 copies in which the sign comes from an f32. 20 ; Test f32 copies in which the sign comes from an f64. 31 ; Test f32 copies in which the sign comes from an f128. 44 ; Test f64 copies in which the sign comes from an f32. 55 ; Test f64 copies in which the sign comes from an f64. 65 ; Test f64 copies in which the sign comes from an f128. 78 ; Test f128 copies in which the sign comes from an f32. We shouldn't
|
/llvm-project/llvm/unittests/ADT/ |
H A D | AnyTest.cpp | 133 : Copies(Other.Copies + 1), Moves(Other.Moves) {} in TEST() 134 TestType(TestType &&Other) : Copies(Other.Copies), Moves(Other.Moves + 1) {} in TEST() 135 int Copies = 0; in TEST() member 141 EXPECT_EQ(0, T1.Copies); in TEST() 147 EXPECT_EQ(1, T2.Copies); in TEST() 152 EXPECT_EQ(0, T3.Copies); in TEST()
|
/llvm-project/libcxx/benchmarks/algorithms/ |
H A D | common.h |
|
/llvm-project/mlir/lib/Dialect/Affine/Transforms/ |
H A D | AffineDataCopyGeneration.cpp | 57 /// inner levels if necessary to determine at what depth copies need to be 60 // TODO: We currently can't generate copies correctly when stores 87 /// Generates copies for memref's living in 'slowMemorySpace' into newly created 103 /// Generate copies for this block. The block is partitioned into separate 163 // We'll recurse and do the copies at an inner level for 'forInst'. in runOnBlock() 167 // We have enough capacity, i.e., copies will be computed for the in runOnBlock() 169 // that for the latter, the copies are placed just before this loop (for in runOnBlock() 170 // incoming copies) and right after (for outgoing ones). in runOnBlock() 172 // Inner loop copies have their own scope - we don't thus update in runOnBlock()
|
/llvm-project/llvm/lib/CodeGen/ |
H A D | MachineCopyPropagation.cpp | 81 STATISTIC(NumDeletes, "Number of dead copies deleted"); 118 DenseMap<MCRegUnit, CopyInfo> Copies; 121 // needed to efficiently remove copies which are invalidated by call in markRegsUnavailable() 153 auto CI = Copies.find(Unit); in invalidateRegister() 154 if (CI != Copies.end()) in invalidateRegister() 180 auto I = Copies.find(Unit); in clobberRegister() 181 if (I != Copies.end()) { in clobberRegister() 189 Copies.erase(Unit); in clobberRegister() 195 auto I = Copies.find(Unit); in clobberRegister() 196 if (I != Copies in clobberRegister() 115 DenseMap<MCRegister, CopyInfo> Copies; global() member in __anond43b9f7b0111::CopyTracker [all...] |
/llvm-project/llvm/test/CodeGen/X86/ |
H A D | misched-copy.ll | 6 ; Argument copies should be hoisted to the top of the block. 7 ; Return copies should be sunk to the end. 8 ; MUL_HiLo PhysReg use copies should be just above the mul. 9 ; MUL_HiLo PhysReg def copies should be just below the mul.
|