Lines Matching full:loc
7 define void @f0(i1 %alwaysFalse, i64 %val, ptr %loc) {
10 ; CHECK-NEXT: store i64 [[VAL:%.*]], ptr [[LOC:%.*]], align 8
13 ; CHECK-NEXT: [[PTR:%.*]] = load ptr addrspace(4), ptr [[LOC]], align 8
20 store i64 %val, ptr %loc
24 %ptr = load ptr addrspace(4), ptr %loc
32 define i64 @f1(i1 %alwaysFalse, ptr addrspace(4) %val, ptr %loc) {
35 ; CHECK-NEXT: store ptr addrspace(4) [[VAL:%.*]], ptr [[LOC:%.*]], align 8
38 ; CHECK-NEXT: [[INT:%.*]] = load i64, ptr [[LOC]], align 8
44 store ptr addrspace(4) %val, ptr %loc
48 %int = load i64, ptr %loc
61 define ptr addrspace(4) @neg_forward_memset(ptr addrspace(4) %loc) {
64 ; CHECK-NEXT: call void @llvm.memset.p4.i64(ptr addrspace(4) align 4 [[LOC:%.*]], i8 7, i64 8, i1 false)
65 ; CHECK-NEXT: [[REF:%.*]] = load ptr addrspace(4), ptr addrspace(4) [[LOC]], align 8
69 call void @llvm.memset.p4.i64(ptr addrspace(4) align 4 %loc, i8 7, i64 8, i1 false)
70 %ref = load ptr addrspace(4), ptr addrspace(4) %loc
74 define <1 x ptr addrspace(4)> @neg_forward_memset_vload(ptr addrspace(4) %loc) {
77 ; CHECK-NEXT: call void @llvm.memset.p4.i64(ptr addrspace(4) align 4 [[LOC:%.*]], i8 7, i64 8, i1 false)
78 ; CHECK-NEXT: [[REF:%.*]] = load <1 x ptr addrspace(4)>, ptr addrspace(4) [[LOC]], align 8
82 call void @llvm.memset.p4.i64(ptr addrspace(4) align 4 %loc, i8 7, i64 8, i1 false)
83 %ref = load <1 x ptr addrspace(4)>, ptr addrspace(4) %loc
89 define ptr addrspace(4) @forward_memset_zero(ptr addrspace(4) %loc) {
92 ; CHECK-NEXT: call void @llvm.memset.p4.i64(ptr addrspace(4) align 4 [[LOC:%.*]], i8 0, i64 8, i1 false)
96 call void @llvm.memset.p4.i64(ptr addrspace(4) align 4 %loc, i8 0, i64 8, i1 false)
97 %ref = load ptr addrspace(4), ptr addrspace(4) %loc
102 define ptr addrspace(4) @neg_forward_store(ptr addrspace(4) %loc) {
105 ; CHECK-NEXT: store i64 5, ptr addrspace(4) [[LOC:%.*]], align 8
106 ; CHECK-NEXT: [[REF:%.*]] = load ptr addrspace(4), ptr addrspace(4) [[LOC]], align 8
110 store i64 5, ptr addrspace(4) %loc
111 %ref = load ptr addrspace(4), ptr addrspace(4) %loc
115 define <1 x ptr addrspace(4)> @neg_forward_store_vload(ptr addrspace(4) %loc) {
118 ; CHECK-NEXT: store i64 5, ptr addrspace(4) [[LOC:%.*]], align 8
119 ; CHECK-NEXT: [[REF:%.*]] = load <1 x ptr addrspace(4)>, ptr addrspace(4) [[LOC]], align 8
123 store i64 5, ptr addrspace(4) %loc
124 %ref = load <1 x ptr addrspace(4)>, ptr addrspace(4) %loc
129 define ptr addrspace(4) @forward_store_zero(ptr addrspace(4) %loc) {
132 ; CHECK-NEXT: store i64 0, ptr addrspace(4) [[LOC:%.*]], align 8
136 store i64 0, ptr addrspace(4) %loc
137 %ref = load ptr addrspace(4), ptr addrspace(4) %loc
142 define ptr addrspace(4) @forward_store_zero2(ptr addrspace(4) %loc) {
145 ; CHECK-NEXT: store <2 x i32> zeroinitializer, ptr addrspace(4) [[LOC:%.*]], align 8
149 store <2 x i32> zeroinitializer, ptr addrspace(4) %loc
150 %ref = load ptr addrspace(4), ptr addrspace(4) %loc
166 define ptr addrspace(4) @neg_forward_memcopy(ptr addrspace(4) %loc) {
169 ; CHECK-NEXT: call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 [[LOC:%.*]], ptr @NonZeroConstant, i64 8, i1 false)
170 ; CHECK-NEXT: [[REF:%.*]] = load ptr addrspace(4), ptr addrspace(4) [[LOC]], align 8
174 call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 %loc, ptr @NonZeroConstant, i64 8, i1 false)
175 %ref = load ptr addrspace(4), ptr addrspace(4) %loc
179 define ptr addrspace(4) @neg_forward_memcopy2(ptr addrspace(4) %loc) {
182 ; CHECK-NEXT: call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 [[LOC:%.*]], ptr @NonZeroConstant, i64 8, i1 false)
183 ; CHECK-NEXT: [[REF:%.*]] = load ptr addrspace(4), ptr addrspace(4) [[LOC]], align 8
187 call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 %loc, ptr @NonZeroConstant, i64 8, i1 false)
188 %ref = load ptr addrspace(4), ptr addrspace(4) %loc
192 define ptr addrspace(4) @forward_memcopy(ptr addrspace(4) %loc) {
195 ; CHECK-NEXT: call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 [[LOC:%.*]], ptr @NonZeroConstant2, i64 8, i1 false)
199 call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 %loc, ptr @NonZeroConstant2, i64 8, i1 false)
200 %ref = load ptr addrspace(4), ptr addrspace(4) %loc
204 define ptr addrspace(4) @forward_memcopy2(ptr addrspace(4) %loc) {
207 ; CHECK-NEXT: call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 [[LOC:%.*]], ptr @NonZeroConstant2, i64 8, i1 false)
211 call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 %loc, ptr @NonZeroConstant2, i64 8, i1 false)
212 %ref = load ptr addrspace(4), ptr addrspace(4) %loc
216 define <1 x ptr addrspace(4)> @neg_forward_memcpy_vload(ptr addrspace(4) %loc) {
219 ; CHECK-NEXT: call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 [[LOC:%.*]], ptr @NonZeroConstant, i64 8, i1 false)
220 ; CHECK-NEXT: [[REF:%.*]] = load <1 x ptr addrspace(4)>, ptr addrspace(4) [[LOC]], align 8
224 call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 %loc, ptr @NonZeroConstant, i64 8, i1 false)
225 %ref = load <1 x ptr addrspace(4)>, ptr addrspace(4) %loc
229 define <4 x ptr addrspace(4)> @neg_forward_memcpy_vload2(ptr addrspace(4) %loc) {
232 ; CHECK-NEXT: call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 [[LOC:%.*]], ptr @NonZeroConstant, i64 32, i1 false)
233 ; CHECK-NEXT: [[REF:%.*]] = load <4 x ptr addrspace(4)>, ptr addrspace(4) [[LOC]], align 32
237 call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 %loc, ptr @NonZeroConstant, i64 32, i1 false)
238 %ref = load <4 x ptr addrspace(4)>, ptr addrspace(4) %loc
242 define <4 x i64> @neg_forward_memcpy_vload3(ptr addrspace(4) %loc) {
245 ; CHECK-NEXT: call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 [[LOC:%.*]], ptr @NonZeroConstant2, i64 32, i1 false)
246 ; CHECK-NEXT: [[REF:%.*]] = load <4 x i64>, ptr addrspace(4) [[LOC]], align 32
250 call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 %loc, ptr @NonZeroConstant2, i64 32, i1 false)
251 %ref = load <4 x i64>, ptr addrspace(4) %loc
255 define <1 x ptr addrspace(4)> @forward_memcpy_vload3(ptr addrspace(4) %loc) {
258 ; CHECK-NEXT: call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 [[LOC:%.*]], ptr @NonZeroConstant2, i64 32, i1 false)
262 call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 %loc, ptr @NonZeroConstant2, i64 32, i1 false)
263 %ref = load <4 x ptr addrspace(4)>, ptr addrspace(4) %loc
270 define ptr addrspace(4) @forward_memcpy_zero(ptr addrspace(4) %loc) {
273 ; CHECK-NEXT: call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 [[LOC:%.*]], ptr @ZeroConstant, i64 8, i1 false)
277 call void @llvm.memcpy.p4.p0.i64(ptr addrspace(4) align 4 %loc, ptr @ZeroConstant, i64 8, i1 false)
278 %ref = load ptr addrspace(4), ptr addrspace(4) %loc
287 define ptr addrspace(4) @neg_store_clobber(ptr addrspace(4) %loc) {
290 ; CHECK-NEXT: store <2 x i64> splat (i64 4), ptr addrspace(4) [[LOC:%.*]], align 16
291 ; CHECK-NEXT: [[LOC_OFF:%.*]] = getelementptr ptr addrspace(4), ptr addrspace(4) [[LOC]], i64 1
296 store <2 x i64> <i64 4, i64 4>, ptr addrspace(4) %loc
297 %loc.off = getelementptr ptr addrspace(4), ptr addrspace(4) %loc, i64 1
298 %ref = load ptr addrspace(4), ptr addrspace(4) %loc.off
306 define ptr addrspace(4) @neg_load_clobber(ptr addrspace(4) %loc) {
309 ; CHECK-NEXT: [[V:%.*]] = load <2 x i64>, ptr addrspace(4) [[LOC:%.*]], align 16
311 ; CHECK-NEXT: [[LOC_OFF:%.*]] = getelementptr ptr addrspace(4), ptr addrspace(4) [[LOC]], i64 1
316 %v = load <2 x i64>, ptr addrspace(4) %loc
318 %loc.off = getelementptr ptr addrspace(4), ptr addrspace(4) %loc, i64 1
319 %ref = load ptr addrspace(4), ptr addrspace(4) %loc.off
323 define ptr addrspace(4) @store_clobber_zero(ptr addrspace(4) %loc) {
326 ; CHECK-NEXT: store <2 x i64> zeroinitializer, ptr addrspace(4) [[LOC:%.*]], align 16
327 ; CHECK-NEXT: [[LOC_OFF:%.*]] = getelementptr ptr addrspace(4), ptr addrspace(4) [[LOC]], i64 1
331 store <2 x i64> zeroinitializer, ptr addrspace(4) %loc
332 %loc.off = getelementptr ptr addrspace(4), ptr addrspace(4) %loc, i64 1
333 %ref = load ptr addrspace(4), ptr addrspace(4) %loc.off
372 define ptr addrspace(5) @multini(i1 %alwaysFalse, ptr addrspace(4) %val, ptr %loc) {
375 ; CHECK-NEXT: store ptr addrspace(4) [[VAL:%.*]], ptr [[LOC:%.*]], align 8
378 ; CHECK-NEXT: [[DIFFERENTAS:%.*]] = load ptr addrspace(5), ptr [[LOC]], align 8
384 store ptr addrspace(4) %val, ptr %loc
388 %differentas = load ptr addrspace(5), ptr %loc